#!/usr/bin/env python # Copyright (c) 2016 Comcast Cable Communications Management, LLC. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at: # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # Filter for .siphon files that are generated by other filters. # The idea is to siphon off certain initializers so that we can better # auto-document the contents of that initializer. import os, sys, argparse, logging import siphon DEFAULT_LOGFILE = None DEFAULT_LOGLEVEL = "info" DEFAULT_SIPHON ="clicmd" DEFAULT_FORMAT = "markdown" DEFAULT_OUTPUT = None DEFAULT_TEMPLATES = os.path.dirname(__file__) + "/siphon_templates" ap = argparse.ArgumentParser() ap.add_argument("--log-file", default=DEFAULT_LOGFILE, help="Log file [%s]" % DEFAULT_LOGFILE) ap.add_argument("--log-level", default=DEFAULT_LOGLEVEL, choices=["debug", "info", "warning", "error", "critical"], help="Logging level [%s]" % DEFAULT_LOGLEVEL) ap.add_argument("--type", '-t', metavar="siphon_type", default=DEFAULT_SIPHON, choices=siphon.process.siphons.keys(), help="Siphon type to process [%s]" % DEFAULT_SIPHON) ap.add_argument("--format", '-f', default=DEFAULT_FORMAT, choices=siphon.process.formats.keys(), help="Output format to generate [%s]" % DEFAULT_FORMAT) ap.add_argument("--output", '-o', metavar="file", default=DEFAULT_OUTPUT, help="Output file (uses stdout if not defined) [%s]" % DEFAULT_OUTPUT) ap.add_argument("--templates", metavar="directory" @media only all and (prefers-color-scheme: dark) { .highlight .hll { background-color: #49483e } .highlight .c { color: #75715e } /* Comment */ .highlight .err { color: #960050; background-color: #1e0010 } /* Error */ .highlight .k { color: #66d9ef } /* Keyword */ .highlight .l { color: #ae81ff } /* Literal */ .highlight .n { color: #f8f8f2 } /* Name */ .highlight .o { color: #f92672 } /* Operator */ .highlight .p { color: #f8f8f2 } /* Punctuation */ .highlight .ch { color: #75715e } /* Comment.Hashbang */ .highlight .cm { color: #75715e } /* Comment.Multiline */ .highlight .cp { color: #75715e } /* Comment.Preproc */ .highlight .cpf { color: #75715e } /* Comment.PreprocFile */ .highlight .c1 { color: #75715e } /* Comment.Single */ .highlight .cs { color: #75715e } /* Comment.Special */ .highlight .gd { color: #f92672 } /* Generic.Deleted */ .highlight .ge { font-style: italic } /* Generic.Emph */ .highlight .gi { color: #a6e22e } /* Generic.Inserted */ .highlight .gs { font-weight: bold } /* Generic.Strong */ .highlight .gu { color: #75715e } /* Generic.Subheading */ .highlight .kc { color: #66d9ef } /* Keyword.Constant */ .highlight .kd { color: #66d9ef } /* Keyword.Declaration */ .highlight .kn { color: #f92672 } /* Keyword.Namespace */ .highlight .kp { color: #66d9ef } /* Keyword.Pseudo */ .highlight .kr { color: #66d9ef } /* Keyword.Reserved */ .highlight .kt { color: #66d9ef } /* Keyword.Type */ .highlight .ld { color: #e6db74 } /* Literal.Date */ .highlight .m { color: #ae81ff } /* Literal.Number */ .highlight .s { color: #e6db74 } /* Literal.String */ .highlight .na { color: #a6e22e } /* Name.Attribute */ .highlight .nb { color: #f8f8f2 } /* Name.Builtin */ .highlight .nc { color: #a6e22e } /* Name.Class */ .highlight .no { color: #66d9ef } /* Name.Constant */ .highlight .nd { color: #a6e22e } /* Name.Decorator */ .highlight .ni { color: #f8f8f2 } /* Name.Entity */ .highlight .ne { color: #a6e22e } /* Name.Exception */ .highlight .nf { color: #a6e22e } /* Name.Function */ .highlight .nl { color: #f8f8f2 } /* Name.Label */ .highlight .nn { color: #f8f8f2 } /* Name.Namespace */ .highlight .nx { color: #a6e22e } /* Name.Other */ .highlight .py { color: #f8f8f2 } /* Name.Property */ .highlight .nt { color: #f92672 } /* Name.Tag */ .highlight .nv { color: #f8f8f2 } /* Name.Variable */ .highlight .ow { color: #f92672 } /* Operator.Word */ .highlight .w { color: #f8f8f2 } /* Text.Whitespace */ .highlight .mb { color: #ae81ff } /* Literal.Number.Bin */ .highlight .mf { color: #ae81ff } /* Literal.Number.Float */ .highlight .mh { color: #ae81ff } /* Literal.Number.Hex */ .highlight .mi { color: #ae81ff } /* Literal.Number.Integer */ .highlight .mo { color: #ae81ff } /* Literal.Number.Oct */ .highlight .sa { color: #e6db74 } /* Literal.String.Affix */ .highlight .sb { color: #e6db74 } /* Literal.String.Backtick */ .highlight .sc { color: #e6db74 } /* Literal.String.Char */ .highlight .dl { color: #e6db74 } /* Literal.String.Delimiter */ .highlight .sd { color: #e6db74 } /* Literal.String.Doc */ .highlight .s2 { color: #e6db74 } /* Literal.String.Double */ .highlight .se { color: #ae81ff } /* Literal.String.Escape */ .highlight .sh { color: #e6db74 } /* Literal.String.Heredoc */ .highlight .si { color: #e6db74 } /* Literal.String.Interpol */ .highlight .sx { color: #e6db74 } /* Literal.String.Other */ .highlight .sr { color: #e6db74 } /* Literal.String.Regex */ .highlight .s1 { color: #e6db74 } /* Literal.String.Single */ .highlight .ss { color: #e6db74 } /* Literal.String.Symbol */ .highlight .bp { color: #f8f8f2 } /* Name.Builtin.Pseudo */ .highlight .fm { color: #a6e22e } /* Name.Function.Magic */ .highlight .vc { color: #f8f8f2 } /* Name.Variable.Class */ .highlight .vg { color: #f8f8f2 } /* Name.Variable.Global */ .highlight .vi { color: #f8f8f2 } /* Name.Variable.Instance */ .highlight .vm { color: #f8f8f2 } /* Name.Variable.Magic */ .highlight .il { color: #ae81ff } /* Literal.Number.Integer.Long */ } @media (prefers-color-scheme: light) { .highlight .hll { background-color: #ffffcc } .highlight .c { color: #888888 } /* Comment */ .highlight .err { color: #a61717; background-color: #e3d2d2 } /* Error */ .highlight .k { color: #008800; font-weight: bold } /* Keyword */ .highlight .ch { color: #888888 } /* Comment.Hashbang */ .highlight .cm { color: #888888 } /* Comment.Multiline */ .highlight .cp { color: #cc0000; font-weight: bold } /* Comment.Preproc */ .highlight .cpf { color: #888888 } /* Comment.PreprocFile */ .highlight .c1 { color: #888888 } /* Comment.Single */ .highlight .cs { color: #cc0000; font-weight: bold; background-color: #fff0f0 } /* Comment.Special */ .highlight .gd { color: #000000; background-color: #ffdddd } /* Generic.Deleted */ .highlight .ge { font-style: italic } /* Generic.Emph */ .highlight .gr { color: #aa0000 } /* Generic.Error */ .highlight .gh { color: #333333 } /* Generic.Heading */ .highlight .gi { color: #000000; background-color: #ddffdd } /* Generic.Inserted */ .highlight .go { color: #888888 } /* Generic.Output */ .highlight .gp { color: #555555 } /* Generic.Prompt */ .highlight .gs { font-weight: bold } /* Generic.Strong */ .highlight .gu { color: #666666 } /* Generic.Subheading */ .highlight .gt { color: #aa0000 } /* Generic.Traceback */ .highlight .kc { color: #008800; font-weight: bold } /* Keyword.Constant */ .highlight .kd { color: #008800; font-weight: bold } /* Keyword.Declaration */ .highlight .kn { color: #008800; font-weight: bold } /* Keyword.Namespace */ .highlight .kp { color: #008800 } /* Keyword.Pseudo */ .highlight .kr { color: #008800; font-weight: bold } /* Keyword.Reserved */ .highlight .kt { color: #888888; font-weight: bold } /* Keyword.Type */ .highlight .m { color: #0000DD; font-weight: bold } /* Literal.Number */ .highlight .s { color: #dd2200; background-color: #fff0f0 } /* Literal.String */ .highlight .na { color: #336699 } /* Name.Attribute */ .highlight .nb { color: #003388 } /* Name.Builtin */ .highlight .nc { color: #bb0066; font-weight: bold } /* Name.Class */ .highlight .no { color: #003366; font-weight: bold } /* Name.Constant */ .highlight .nd { color: #555555 } /* Name.Decorator */ .highlight .ne { color: #bb0066; font-weight: bold } /* Name.Exception */ .highlight .nf { color: #0066bb; font-weight: bold } /* Name.Function */ .highlight .nl { color: #336699; font-style: italic } /* Name.Label */ .highlight .nn { color: #bb0066; font-weight: bold } /* Name.Namespace */ .highlight .py { color: #336699; font-weight: bold } /* Name.Property */ .highlight .nt { color: #bb0066; font-weight: bold } /* Name.Tag */ .highlight .nv { color: #336699 } /* Name.Variable */ .highlight .ow { color: #008800 } /* Operator.Word */ .highlight .w { color: #bbbbbb } /* Text.Whitespace */ .highlight .mb { color: #0000DD; font-weight: bold } /* Literal.Number.Bin */ .highlight .mf { color: #0000DD; font-weight: bold } /* Literal.Number.Float */ .highlight .mh { color: #0000DD; font-weight: bold } /* Literal.Number.Hex */ .highlight .mi { color: #0000DD; font-weight: bold } /* Literal.Number.Integer */ .highlight .mo { color: #0000DD; font-weight: bold } /* Literal.Number.Oct */ .highlight .sa { color: #dd2200; background-color: #fff0f0 } /* Literal.String.Affix */ .highlight .sb { color: #dd2200; background-color: #fff0f0 } /* Literal.String.Backtick */ .highlight .sc { color: #dd2200; background-color: #fff0f0 } /* Literal.String.Char */ .highlight .dl { color: #dd2200; background-color: #fff0f0 } /* Literal.String.Delimiter */ .highlight .sd { color: #dd2200; background-color: #fff0f0 } /* Literal.String.Doc */ .highlight .s2 { color: #dd2200; background-color: #fff0f0 } /* Literal.String.Double */ .highlight .se { color: #0044dd; background-color: #fff0f0 } /* Literal.String.Escape */ .highlight .sh { color: #dd2200; background-color: #fff0f0 } /* Literal.String.Heredoc */ .highlight .si { color: #3333bb; background-color: #fff0f0 } /* Literal.String.Interpol */ .highlight .sx { color: #22bb22; background-color: #f0fff0 } /* Literal.String.Other */ .highlight .sr { color: #008800; background-color: #fff0ff } /* Literal.String.Regex */ .highlight .s1 { color: #dd2200; background-color: #fff0f0 } /* Literal.String.Single */ .highlight .ss { color: #aa6600; background-color: #fff0f0 } /* Literal.String.Symbol */ .highlight .bp { color: #003388 } /* Name.Builtin.Pseudo */ .highlight .fm { color: #0066bb; font-weight: bold } /* Name.Function.Magic */ .highlight .vc { color: #336699 } /* Name.Variable.Class */ .highlight .vg { color: #dd7700 } /* Name.Variable.Global */ .highlight .vi { color: #3333bb } /* Name.Variable.Instance */ .highlight .vm { color: #336699 } /* Name.Variable.Magic */ .highlight .il { color: #0000DD; font-weight: bold } /* Literal.Number.Integer.Long */ } #!/bin/bash # Copyright (c) 2017 Cisco and/or its affiliates. # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at: # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. set -x # Space separated list of available testbeds, described by topology files TOPOLOGIES="topologies/available/lf_testbed1.yaml \ topologies/available/lf_testbed2.yaml \ topologies/available/lf_testbed3.yaml" SCRIPT_DIR="$( cd "$( dirname "${BASH_SOURCE[0]}" )" && pwd )" # Reservation dir RESERVATION_DIR="/tmp/reservation_dir" INSTALLATION_DIR="/tmp/install_dir" PYBOT_ARGS="-W 150 -L TRACE" ARCHIVE_ARTIFACTS=(log.html output.xml report.html output_perf_data.xml output_perf_data.json) # If we run this script from CSIT jobs we want to use stable vpp version if [[ ${JOB_NAME} == csit-* ]] ; then mkdir vpp_download cd vpp_download if [[ ${TEST_TAG} == *NIGHTLY ]] || \ [[ ${TEST_TAG} == *DAILY ]] || \ [[ ${TEST_TAG} == *WEEKLY ]]; then # Download the latest VPP build .deb install packages echo Downloading VPP packages... bash ${SCRIPT_DIR}/resources/tools/download_install_vpp_pkgs.sh --skip-install VPP_DEBS="$( readlink -f *.deb | tr '\n' ' ' )" # Take vpp package and get the vpp version VPP_STABLE_VER="$( expr match $(ls *.deb | head -n 1) 'vpp-\(.*\)-deb.deb' )" else DPDK_STABLE_VER=$(cat ${SCRIPT_DIR}/DPDK_STABLE_VER)_amd64 VPP_REPO_URL=$(cat ${SCRIPT_DIR}/VPP_REPO_URL_UBUNTU) VPP_STABLE_VER=$(cat ${SCRIPT_DIR}/VPP_STABLE_VER_UBUNTU) VPP_CLASSIFIER="-deb" # Download vpp build from nexus and set VPP_DEBS variable wget -q "${VPP_REPO_URL}/vpp/${VPP_STABLE_VER}/vpp-${VPP_STABLE_VER}${VPP_CLASSIFIER}.deb" || exit wget -q "${VPP_REPO_URL}/vpp-dbg/${VPP_STABLE_VER}/vpp-dbg-${VPP_STABLE_VER}${VPP_CLASSIFIER}.deb" || exit wget -q "${VPP_REPO_URL}/vpp-dev/${VPP_STABLE_VER}/vpp-dev-${VPP_STABLE_VER}${VPP_CLASSIFIER}.deb" || exit wget -q "${VPP_REPO_URL}/vpp-dpdk-dkms/${DPDK_STABLE_VER}/vpp-dpdk-dkms-${DPDK_STABLE_VER}${VPP_CLASSIFIER}.deb" || exit wget -q "${VPP_REPO_URL}/vpp-lib/${VPP_STABLE_VER}/vpp-lib-${VPP_STABLE_VER}${VPP_CLASSIFIER}.deb" || exit wget -q "${VPP_REPO_URL}/vpp-plugins/${VPP_STABLE_VER}/vpp-plugins-${VPP_STABLE_VER}${VPP_CLASSIFIER}.deb" || exit VPP_DEBS="$( readlink -f *.deb | tr '\n' ' ' )" fi cd .. # If we run this script from vpp project we want to use local build elif [[ ${JOB_NAME} == vpp-* ]] ; then # Use local packages provided as argument list # Jenkins VPP deb paths (convert to full path) VPP_DEBS="$( readlink -f $@ | tr '\n' ' ' )" # Take vpp package and get the vpp version VPP_STABLE_VER="$( expr match $1 'vpp-\(.*\)-deb.deb' )" # Download DPDK parts not included in dpdk plugin of vpp build DPDK_STABLE_VER=$(cat ${SCRIPT_DIR}/DPDK_STABLE_VER)_amd64 VPP_REPO_URL=$(cat ${SCRIPT_DIR}/VPP_REPO_URL_UBUNTU) VPP_CLASSIFIER="-deb" wget -q "${VPP_REPO_URL}/vpp-dpdk-dkms/${DPDK_STABLE_VER}/vpp-dpdk-dkms-${DPDK_STABLE_VER}${VPP_CLASSIFIER}.deb" || exit VPP_DEBS+=($( readlink -f vpp-dpdk-dkms-${DPDK_STABLE_VER}${VPP_CLASSIFIER}.deb )) else echo "Unable to identify job type based on JOB_NAME variable: ${JOB_NAME}" exit 1 fi WORKING_TOPOLOGY="" export PYTHONPATH=${SCRIPT_DIR} sudo apt-get -y update sudo apt-get -y install libpython2.7-dev python-virtualenv virtualenv --system-site-packages env . env/bin/activate echo pip install pip install -r requirements.txt # We iterate over available topologies and wait until we reserve topology while :; do for TOPOLOGY in ${TOPOLOGIES}; do python ${SCRIPT_DIR}/resources/tools/topo_reservation.py -t ${TOPOLOGY} if [ $? -eq 0 ]; then WORKING_TOPOLOGY=${TOPOLOGY} echo "Reserved: ${WORKING_TOPOLOGY}" break fi done if [ ! -z "${WORKING_TOPOLOGY}" ]; then # Exit the infinite while loop if we made a reservation break fi # Wait ~3minutes before next try SLEEP_TIME=$[ ( $RANDOM % 20 ) + 180 ]s echo "Sleeping ${SLEEP_TIME}" sleep ${SLEEP_TIME} done function cancel_all { python ${SCRIPT_DIR}/resources/tools/topo_installation.py -c -d ${INSTALLATION_DIR} -t $1 python ${SCRIPT_DIR}/resources/tools/topo_reservation.py -c -t $1 } # On script exit we cancel the reservation and installation and delete all vpp # packages trap "cancel_all ${WORKING_TOPOLOGY}" EXIT python ${SCRIPT_DIR}/resources/tools/topo_installation.py -t ${WORKING_TOPOLOGY} \ -d ${INSTALLATION_DIR} \ -p ${VPP_DEBS} if [ $? -eq 0 ]; then echo "VPP Installed on hosts from: ${WORKING_TOPOLOGY}" else echo "Failed to copy vpp deb files to DUTs" exit 1 fi case "$TEST_TAG" in # run specific performance tests based on jenkins job type variable PERFTEST_DAILY ) pybot ${PYBOT_ARGS} \ -v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \ -s "tests.perf" \ --include ndrdiscANDnic_intel-x520-da2AND1t1cORndrdiscANDnic_intel-x520-da2AND2t2c \ tests/ RETURN_STATUS=$(echo $?) ;; PERFTEST_SEMI_WEEKLY ) pybot ${PYBOT_ARGS} \ -v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \ -s "tests.perf" \ --include ndrdiscANDnic_intel-x710AND1t1cORndrdiscANDnic_intel-x710AND2t2cORndrdiscANDnic_intel-xl710AND1t1cORndrdiscANDnic_intel-xl710AND2t2c \ tests/ RETURN_STATUS=$(echo $?) ;; VERIFY-PERF-NDRDISC ) pybot ${PYBOT_ARGS} \ -v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \ -s "tests.perf" \ --include ndrdiscAND1t1cORndrdiscAND2t2c \ tests/ RETURN_STATUS=$(echo $?) ;; VERIFY-PERF-PDRDISC ) pybot ${PYBOT_ARGS} \ -v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \ -s "tests.perf" \ --include pdrdiscAND1t1cORpdrdiscAND2t2c \ tests/ RETURN_STATUS=$(echo $?) ;; VERIFY-PERF-NDRCHK ) pybot ${PYBOT_ARGS} \ -v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \ -s "tests.perf" \ --include ndrchkAND1t1cORndrchkAND2t2c \ tests/ RETURN_STATUS=$(echo $?) ;; PERFTEST_NDRCHK_DAILY ) pybot ${PYBOT_ARGS} \ -v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \ -s "tests.perf" \ --include ndrchkAND1t1cORndrchkAND2t2c \ tests/ RETURN_STATUS=$(echo $?) ;; VERIFY-PERF-IP4 ) pybot ${PYBOT_ARGS} \ -v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \ -s "tests.perf" \ --include ndrdiscANDnic_intel-x520-da2AND1t1cANDip4baseORndrdiscANDnic_intel-x520-da2AND1t1cANDip4fwdANDfib_2m \ tests/ RETURN_STATUS=$(echo $?) ;; VERIFY-PERF-IP6 ) pybot ${PYBOT_ARGS} \ -v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \ -s "tests.perf" \ --include ndrdiscANDnic_intel-x520-da2AND1t1cANDip6baseORndrdiscANDnic_intel-x520-da2AND1t1cANDip6fwdANDfib_2m \ tests/ RETURN_STATUS=$(echo $?) ;; VERIFY-PERF-L2 ) pybot ${PYBOT_ARGS} \ -v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \ -s "tests.perf" \ --include ndrdiscANDnic_intel-x520-da2AND1t1cANDl2xcbaseORndrdiscANDnic_intel-x520-da2AND1t1cANDl2bdbase \ tests/ RETURN_STATUS=$(echo $?) ;; VERIFY-PERF-LISP ) pybot ${PYBOT_ARGS} \ -v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \ -s "tests.perf" \ --include ndrdiscANDnic_intel-x520-da2AND1t1cANDlisp \ tests/ RETURN_STATUS=$(echo $?) ;; VERIFY-PERF-VXLAN ) pybot ${PYBOT_ARGS} \ -v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \ -s "tests.perf" \ --include ndrdiscANDnic_intel-x520-da2AND1t1cANDvxlan \ tests/ RETURN_STATUS=$(echo $?) ;; VERIFY-PERF-VHOST ) pybot ${PYBOT_ARGS} \ -v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \ -s "tests.perf" \ --include ndrdiscANDnic_intel-x520-da2AND1t1cANDvhost \ tests/ RETURN_STATUS=$(echo $?) ;; VPP-VERIFY-PERF-IP4 ) pybot ${PYBOT_ARGS} \ -v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \ -s "tests.perf" \ --include ndrchkANDnic_intel-x520-da2AND1t1cANDip4baseORndrchkANDnic_intel-x520-da2AND1t1cANDip4fwdANDfib_2m \ tests/ RETURN_STATUS=$(echo $?) ;; VPP-VERIFY-PERF-IP6 ) pybot ${PYBOT_ARGS} \ -v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \ -s "tests.perf" \ --include ndrchkANDnic_intel-x520-da2AND1t1cANDip6baseORndrchkANDnic_intel-x520-da2AND1t1cANDip6fwdANDfib_2m \ tests/ RETURN_STATUS=$(echo $?) ;; VPP-VERIFY-PERF-L2 ) pybot ${PYBOT_ARGS} \ -v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \ -s "tests.perf" \ --include ndrchkANDnic_intel-x520-da2AND1t1cANDl2xcbaseORndrchkANDnic_intel-x520-da2AND1t1cANDl2bdbase \ tests/ RETURN_STATUS=$(echo $?) ;; VPP-VERIFY-PERF-LISP ) pybot ${PYBOT_ARGS} \ -v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \ -s "tests.perf" \ --include ndrchkANDnic_intel-x520-da2AND1t1cANDlisp \ tests/ RETURN_STATUS=$(echo $?) ;; VPP-VERIFY-PERF-VXLAN ) pybot ${PYBOT_ARGS} \ -v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \ -s "tests.perf" \ --include ndrchkANDnic_intel-x520-da2AND1t1cANDvxlan \ tests/ RETURN_STATUS=$(echo $?) ;; VPP-VERIFY-PERF-VHOST ) pybot ${PYBOT_ARGS} \ -v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \ -s "tests.perf" \ --include ndrchkANDnic_intel-x520-da2AND1t1cANDvhost \ tests/ RETURN_STATUS=$(echo $?) ;; PERFTEST_LONG ) pybot ${PYBOT_ARGS} \ -v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \ -s "tests.perf" \ --exclude SKIP_PATCH \ -i NDRPDRDISC \ tests/ RETURN_STATUS=$(echo $?) ;; PERFTEST_SHORT ) pybot ${PYBOT_ARGS} \ -v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \ -s "tests.perf" \ -i NDRCHK \ tests/ RETURN_STATUS=$(echo $?) ;; PERFTEST_NIGHTLY ) #run all available tests pybot ${PYBOT_ARGS} \ -v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \ -s "tests.perf" \ tests/ RETURN_STATUS=$(echo $?) ;; * ) # run full performance test suite and exit on fail pybot ${PYBOT_ARGS} \ -v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \ -s "tests.perf" \ tests/ RETURN_STATUS=$(echo $?) esac # Pybot output post-processing echo Post-processing test data... python ${SCRIPT_DIR}/resources/tools/robot_output_parser.py \ -i ${SCRIPT_DIR}/output.xml \ -o ${SCRIPT_DIR}/output_perf_data.xml \ -v ${VPP_STABLE_VER} if [ ! $? -eq 0 ]; then echo "Parsing ${SCRIPT_DIR}/output.xml failed" fi python ${SCRIPT_DIR}/resources/tools/report_gen/run_robot_json_data.py \ --input ${SCRIPT_DIR}/output.xml \ --output ${SCRIPT_DIR}/output_perf_data.json \ --vdevice ${VPP_STABLE_VER} if [ ! $? -eq 0 ]; then echo "Generating JSON data for report from ${SCRIPT_DIR}/output.xml failed" fi # Archive artifacts mkdir archive for i in ${ARCHIVE_ARTIFACTS[@]}; do cp $( readlink -f ${i} | tr '\n' ' ' ) archive/ done echo Post-processing finished. exit ${RETURN_STATUS}
#!/bin/bash # Copyright (c) 2017 Cisco and/or its affiliates. # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at: # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. set -x # Space separated list of available testbeds, described by topology files TOPOLOGIES="topologies/available/lf_testbed1.yaml \ topologies/available/lf_testbed2.yaml \ topologies/available/lf_testbed3.yaml" SCRIPT_DIR="$( cd "$( dirname "${BASH_SOURCE[0]}" )" && pwd )" # Reservation dir RESERVATION_DIR="/tmp/reservation_dir" INSTALLATION_DIR="/tmp/install_dir" PYBOT_ARGS="-W 150 -L TRACE" ARCHIVE_ARTIFACTS=(log.html output.xml report.html output_perf_data.xml output_perf_data.json) # If we run this script from CSIT jobs we want to use stable vpp version if [[ ${JOB_NAME} == csit-* ]] ; then mkdir vpp_download cd vpp_download if [[ ${TEST_TAG} == *NIGHTLY ]] || \ [[ ${TEST_TAG} == *DAILY ]] || \ [[ ${TEST_TAG} == *WEEKLY ]]; then # Download the latest VPP build .deb install packages echo Downloading VPP packages... bash ${SCRIPT_DIR}/resources/tools/download_install_vpp_pkgs.sh --skip-install VPP_DEBS="$( readlink -f *.deb | tr '\n' ' ' )" # Take vpp package and get the vpp version VPP_STABLE_VER="$( expr match $(ls *.deb | head -n 1) 'vpp-\(.*\)-deb.deb' )" else DPDK_STABLE_VER=$(cat ${SCRIPT_DIR}/DPDK_STABLE_VER)_amd64 VPP_REPO_URL=$(cat ${SCRIPT_DIR}/VPP_REPO_URL_UBUNTU) VPP_STABLE_VER=$(cat ${SCRIPT_DIR}/VPP_STABLE_VER_UBUNTU) VPP_CLASSIFIER="-deb" # Download vpp build from nexus and set VPP_DEBS variable wget -q "${VPP_REPO_URL}/vpp/${VPP_STABLE_VER}/vpp-${VPP_STABLE_VER}${VPP_CLASSIFIER}.deb" || exit wget -q "${VPP_REPO_URL}/vpp-dbg/${VPP_STABLE_VER}/vpp-dbg-${VPP_STABLE_VER}${VPP_CLASSIFIER}.deb" || exit wget -q "${VPP_REPO_URL}/vpp-dev/${VPP_STABLE_VER}/vpp-dev-${VPP_STABLE_VER}${VPP_CLASSIFIER}.deb" || exit wget -q "${VPP_REPO_URL}/vpp-dpdk-dkms/${DPDK_STABLE_VER}/vpp-dpdk-dkms-${DPDK_STABLE_VER}${VPP_CLASSIFIER}.deb" || exit wget -q "${VPP_REPO_URL}/vpp-lib/${VPP_STABLE_VER}/vpp-lib-${VPP_STABLE_VER}${VPP_CLASSIFIER}.deb" || exit wget -q "${VPP_REPO_URL}/vpp-plugins/${VPP_STABLE_VER}/vpp-plugins-${VPP_STABLE_VER}${VPP_CLASSIFIER}.deb" || exit VPP_DEBS="$( readlink -f *.deb | tr '\n' ' ' )" fi cd .. # If we run this script from vpp project we want to use local build elif [[ ${JOB_NAME} == vpp-* ]] ; then # Use local packages provided as argument list # Jenkins VPP deb paths (convert to full path) VPP_DEBS="$( readlink -f $@ | tr '\n' ' ' )" # Take vpp package and get the vpp version VPP_STABLE_VER="$( expr match $1 'vpp-\(.*\)-deb.deb' )" # Download DPDK parts not included in dpdk plugin of vpp build DPDK_STABLE_VER=$(cat ${SCRIPT_DIR}/DPDK_STABLE_VER)_amd64 VPP_REPO_URL=$(cat ${SCRIPT_DIR}/VPP_REPO_URL_UBUNTU) VPP_CLASSIFIER="-deb" wget -q "${VPP_REPO_URL}/vpp-dpdk-dkms/${DPDK_STABLE_VER}/vpp-dpdk-dkms-${DPDK_STABLE_VER}${VPP_CLASSIFIER}.deb" || exit VPP_DEBS+=($( readlink -f vpp-dpdk-dkms-${DPDK_STABLE_VER}${VPP_CLASSIFIER}.deb )) else echo "Unable to identify job type based on JOB_NAME variable: ${JOB_NAME}" exit 1 fi WORKING_TOPOLOGY="" export PYTHONPATH=${SCRIPT_DIR} sudo apt-get -y update sudo apt-get -y install libpython2.7-dev python-virtualenv virtualenv --system-site-packages env . env/bin/activate echo pip install pip install -r requirements.txt # We iterate over available topologies and wait until we reserve topology while :; do for TOPOLOGY in ${TOPOLOGIES}; do python ${SCRIPT_DIR}/resources/tools/topo_reservation.py -t ${TOPOLOGY} if [ $? -eq 0 ]; then WORKING_TOPOLOGY=${TOPOLOGY} echo "Reserved: ${WORKING_TOPOLOGY}" break fi done if [ ! -z "${WORKING_TOPOLOGY}" ]; then # Exit the infinite while loop if we made a reservation break fi # Wait ~3minutes before next try SLEEP_TIME=$[ ( $RANDOM % 20 ) + 180 ]s echo "Sleeping ${SLEEP_TIME}" sleep ${SLEEP_TIME} done function cancel_all { python ${SCRIPT_DIR}/resources/tools/topo_installation.py -c -d ${INSTALLATION_DIR} -t $1 python ${SCRIPT_DIR}/resources/tools/topo_reservation.py -c -t $1 } # On script exit we cancel the reservation and installation and delete all vpp # packages trap "cancel_all ${WORKING_TOPOLOGY}" EXIT python ${SCRIPT_DIR}/resources/tools/topo_installation.py -t ${WORKING_TOPOLOGY} \ -d ${INSTALLATION_DIR} \ -p ${VPP_DEBS} if [ $? -eq 0 ]; then echo "VPP Installed on hosts from: ${WORKING_TOPOLOGY}" else echo "Failed to copy vpp deb files to DUTs" exit 1 fi case "$TEST_TAG" in # run specific performance tests based on jenkins job type variable PERFTEST_DAILY ) pybot ${PYBOT_ARGS} \ -v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \ -s "tests.perf" \ --include ndrdiscANDnic_intel-x520-da2AND1t1cORndrdiscANDnic_intel-x520-da2AND2t2c \ tests/ RETURN_STATUS=$(echo $?) ;; PERFTEST_SEMI_WEEKLY ) pybot ${PYBOT_ARGS} \ -v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \ -s "tests.perf" \ --include ndrdiscANDnic_intel-x710AND1t1cORndrdiscANDnic_intel-x710AND2t2cORndrdiscANDnic_intel-xl710AND1t1cORndrdiscANDnic_intel-xl710AND2t2c \ tests/ RETURN_STATUS=$(echo $?) ;; VERIFY-PERF-NDRDISC ) pybot ${PYBOT_ARGS} \ -v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \ -s "tests.perf" \ --include ndrdiscAND1t1cORndrdiscAND2t2c \ tests/ RETURN_STATUS=$(echo $?) ;; VERIFY-PERF-PDRDISC ) pybot ${PYBOT_ARGS} \ -v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \ -s "tests.perf" \ --include pdrdiscAND1t1cORpdrdiscAND2t2c \ tests/ RETURN_STATUS=$(echo $?) ;; VERIFY-PERF-NDRCHK ) pybot ${PYBOT_ARGS} \ -v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \ -s "tests.perf" \ --include ndrchkAND1t1cORndrchkAND2t2c \ tests/ RETURN_STATUS=$(echo $?) ;; PERFTEST_NDRCHK_DAILY ) pybot ${PYBOT_ARGS} \ -v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \ -s "tests.perf" \ --include ndrchkAND1t1cORndrchkAND2t2c \ tests/ RETURN_STATUS=$(echo $?) ;; VERIFY-PERF-IP4 ) pybot ${PYBOT_ARGS} \ -v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \ -s "tests.perf" \ --include ndrdiscANDnic_intel-x520-da2AND1t1cANDip4baseORndrdiscANDnic_intel-x520-da2AND1t1cANDip4fwdANDfib_2m \ tests/ RETURN_STATUS=$(echo $?) ;; VERIFY-PERF-IP6 ) pybot ${PYBOT_ARGS} \ -v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \ -s "tests.perf" \ --include ndrdiscANDnic_intel-x520-da2AND1t1cANDip6baseORndrdiscANDnic_intel-x520-da2AND1t1cANDip6fwdANDfib_2m \ tests/ RETURN_STATUS=$(echo $?) ;; VERIFY-PERF-L2 ) pybot ${PYBOT_ARGS} \ -v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \ -s "tests.perf" \ --include ndrdiscANDnic_intel-x520-da2AND1t1cANDl2xcbaseORndrdiscANDnic_intel-x520-da2AND1t1cANDl2bdbase \ tests/ RETURN_STATUS=$(echo $?) ;; VERIFY-PERF-LISP ) pybot ${PYBOT_ARGS} \ -v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \ -s "tests.perf" \ --include ndrdiscANDnic_intel-x520-da2AND1t1cANDlisp \ tests/ RETURN_STATUS=$(echo $?) ;; VERIFY-PERF-VXLAN ) pybot ${PYBOT_ARGS} \ -v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \ -s "tests.perf" \ --include ndrdiscANDnic_intel-x520-da2AND1t1cANDvxlan \ tests/ RETURN_STATUS=$(echo $?) ;; VERIFY-PERF-VHOST ) pybot ${PYBOT_ARGS} \ -v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \ -s "tests.perf" \ --include ndrdiscANDnic_intel-x520-da2AND1t1cANDvhost \ tests/ RETURN_STATUS=$(echo $?) ;; VPP-VERIFY-PERF-IP4 ) pybot ${PYBOT_ARGS} \ -v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \ -s "tests.perf" \ --include ndrchkANDnic_intel-x520-da2AND1t1cANDip4baseORndrchkANDnic_intel-x520-da2AND1t1cANDip4fwdANDfib_2m \ tests/ RETURN_STATUS=$(echo $?) ;; VPP-VERIFY-PERF-IP6 ) pybot ${PYBOT_ARGS} \ -v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \ -s "tests.perf" \ --include ndrchkANDnic_intel-x520-da2AND1t1cANDip6baseORndrchkANDnic_intel-x520-da2AND1t1cANDip6fwdANDfib_2m \ tests/ RETURN_STATUS=$(echo $?) ;; VPP-VERIFY-PERF-L2 ) pybot ${PYBOT_ARGS} \ -v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \ -s "tests.perf" \ --include ndrchkANDnic_intel-x520-da2AND1t1cANDl2xcbaseORndrchkANDnic_intel-x520-da2AND1t1cANDl2bdbase \ tests/ RETURN_STATUS=$(echo $?) ;; VPP-VERIFY-PERF-LISP ) pybot ${PYBOT_ARGS} \ -v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \ -s "tests.perf" \ --include ndrchkANDnic_intel-x520-da2AND1t1cANDlisp \ tests/ RETURN_STATUS=$(echo $?) ;; VPP-VERIFY-PERF-VXLAN ) pybot ${PYBOT_ARGS} \ -v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \ -s "tests.perf" \ --include ndrchkANDnic_intel-x520-da2AND1t1cANDvxlan \ tests/ RETURN_STATUS=$(echo $?) ;; VPP-VERIFY-PERF-VHOST ) pybot ${PYBOT_ARGS} \ -v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \ -s "tests.perf" \ --include ndrchkANDnic_intel-x520-da2AND1t1cANDvhost \ tests/ RETURN_STATUS=$(echo $?) ;; PERFTEST_LONG ) pybot ${PYBOT_ARGS} \ -v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \ -s "tests.perf" \ --exclude SKIP_PATCH \ -i NDRPDRDISC \ tests/ RETURN_STATUS=$(echo $?) ;; PERFTEST_SHORT ) pybot ${PYBOT_ARGS} \ -v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \ -s "tests.perf" \ -i NDRCHK \ tests/ RETURN_STATUS=$(echo $?) ;; PERFTEST_NIGHTLY ) #run all available tests pybot ${PYBOT_ARGS} \ -v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \ -s "tests.perf" \ tests/ RETURN_STATUS=$(echo $?) ;; * ) # run full performance test suite and exit on fail pybot ${PYBOT_ARGS} \ -v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \ -s "tests.perf" \ tests/ RETURN_STATUS=$(echo $?) esac # Pybot output post-processing echo Post-processing test data... python ${SCRIPT_DIR}/resources/tools/robot_output_parser.py \ -i ${SCRIPT_DIR}/output.xml \ -o ${SCRIPT_DIR}/output_perf_data.xml \ -v ${VPP_STABLE_VER} if [ ! $? -eq 0 ]; then echo "Parsing ${SCRIPT_DIR}/output.xml failed" fi python ${SCRIPT_DIR}/resources/tools/report_gen/run_robot_json_data.py \ --input ${SCRIPT_DIR}/output.xml \ --output ${SCRIPT_DIR}/output_perf_data.json \ --vdevice ${VPP_STABLE_VER} if [ ! $? -eq 0 ]; then echo "Generating JSON data for report from ${SCRIPT_DIR}/output.xml failed" fi # Archive artifacts mkdir archive for i in ${ARCHIVE_ARTIFACTS[@]}; do cp $( readlink -f ${i} | tr '\n' ' ' ) archive/ done echo Post-processing finished. exit ${RETURN_STATUS}