aboutsummaryrefslogtreecommitdiffstats
path: root/bootstrap-verify-perf.sh
diff options
context:
space:
mode:
Diffstat (limited to 'bootstrap-verify-perf.sh')
-rwxr-xr-xbootstrap-verify-perf.sh400
1 files changed, 4 insertions, 396 deletions
diff --git a/bootstrap-verify-perf.sh b/bootstrap-verify-perf.sh
index c72d6ec7dd..afcfd1eca9 100755
--- a/bootstrap-verify-perf.sh
+++ b/bootstrap-verify-perf.sh
@@ -12,400 +12,8 @@
# See the License for the specific language governing permissions and
# limitations under the License.
-set -xo pipefail
+set -exuo pipefail
-# FUNCTIONS
-function warn () {
- # Prints the message to standard error.
- echo "$@" >&2
-}
-
-function die () {
- # Prints the message to standard error end exit with error code specified
- # by first argument.
- status="$1"
- shift
- warn "$@"
- exit "$status"
-}
-
-function help () {
- # Displays help message.
- die 1 "Usage: `basename $0` csit-[dpdk|vpp|ligato]-[2n-skx|3n-skx|3n-hsw]"
-}
-
-function cancel_all () {
- # Trap function to get into consistent state.
- python ${SCRIPT_DIR}/resources/tools/scripts/topo_cleanup.py -t $1 || {
- die 1 "Failure during execution of topology cleanup script!"
- }
- python ${SCRIPT_DIR}/resources/tools/scripts/topo_reservation.py -c -t $1 || {
- die 1 "Failure during execution of topology un-reservation script!"
- }
-}
-
-# VARIABLES
-# Space separated list of available testbeds, described by topology files
-TOPOLOGIES_3N_HSW=(topologies/available/lf_3n_hsw_testbed1.yaml
- topologies/available/lf_3n_hsw_testbed2.yaml
- topologies/available/lf_3n_hsw_testbed3.yaml)
-TOPOLOGIES_2N_SKX=(topologies/available/lf_2n_skx_testbed21.yaml
- topologies/available/lf_2n_skx_testbed24.yaml)
-TOPOLOGIES_3N_SKX=(topologies/available/lf_3n_skx_testbed31.yaml
- topologies/available/lf_3n_skx_testbed32.yaml)
-
-SCRIPT_DIR="$( cd "$( dirname "${BASH_SOURCE[0]}" )" && pwd )"
-export PYTHONPATH=${SCRIPT_DIR}
-
-RESERVATION_DIR="/tmp/reservation_dir"
-DOWNLOAD_DIR="${SCRIPT_DIR}/download_dir"
-ARCHIVE_DIR="${SCRIPT_DIR}/archive"
-
-mkdir -p ${DOWNLOAD_DIR} || {
- die 1 "Failed to create download dir!"
-}
-mkdir -p ${ARCHIVE_DIR} || {
- die 1 "Failed to create archive dir!"
-}
-
-# Get test code.
-TEST_CODE=${JOB_NAME-}
-if [[ -z ${TEST_CODE} ]]; then
- TEST_CODE=${1}
- shift
-fi
-
-# TOPOLOGY SELECTION
-case "$TEST_CODE" in
- *2n-skx*)
- TOPOLOGIES=${TOPOLOGIES_2N_SKX[@]}
- TOPOLOGIES_TAGS="2_node_*_link_topo"
- ;;
- *3n-skx*)
- TOPOLOGIES=${TOPOLOGIES_3N_SKX[@]}
- TOPOLOGIES_TAGS="3_node_*_link_topo"
- ;;
- *)
- # Fallback to 3-node Haswell by default (backward compatibility)
- TOPOLOGIES=${TOPOLOGIES_3N_HSW[@]}
- TOPOLOGIES_TAGS="3_node_*_link_topo"
- ;;
-esac
-
-if [[ -z "${TOPOLOGIES}" ]]; then
- die 1 "No applicable topology found!"
-fi
-
-cd ${DOWNLOAD_DIR}
-case "$TEST_CODE" in
- *hc2vpp*)
- DUT="hc2vpp"
- ;;
- *vpp*)
- DUT="vpp"
-
- case "$TEST_CODE" in
- csit-vpp-*)
- # Use downloaded packages with specific version
- if [[ "$TEST_CODE" == *daily* ]] || \
- [[ "$TEST_CODE" == *weekly* ]] || \
- [[ "$TEST_CODE" == *timed* ]];
- then
- echo Downloading latest VPP packages from NEXUS...
- bash ${SCRIPT_DIR}/resources/tools/scripts/download_install_vpp_pkgs.sh \
- --skip-install || {
- die 1 "Failed to get VPP packages!"
- }
- else
- echo Downloading VPP packages of specific version from NEXUS...
- DPDK_STABLE_VER=$(cat ${SCRIPT_DIR}/DPDK_STABLE_VER)
- VPP_STABLE_VER=$(cat ${SCRIPT_DIR}/VPP_STABLE_VER_UBUNTU)
- bash ${SCRIPT_DIR}/resources/tools/scripts/download_install_vpp_pkgs.sh \
- --skip-install --vpp ${VPP_STABLE_VER} --dkms ${DPDK_STABLE_VER} || {
- die 1 "Failed to get VPP packages!"
- }
- fi
- ;;
- vpp-csit-*)
- # Use local built packages.
- mv ../${DUT}*.deb ${DOWNLOAD_DIR}/
- ;;
- *)
- die 1 "Unable to identify job type from: ${TEST_CODE}!"
- ;;
- esac
- ;;
- *ligato*)
- DUT="kubernetes"
-
- case "$TEST_CODE" in
- csit-*)
- # Use downloaded packages with specific version
- if [[ "$TEST_CODE" == *daily* ]] || \
- [[ "$TEST_CODE" == *weekly* ]] || \
- [[ "$TEST_CODE" == *timed* ]];
- then
- echo Downloading latest VPP packages from NEXUS...
- bash ${SCRIPT_DIR}/resources/tools/scripts/download_install_vpp_pkgs.sh \
- --skip-install || {
- die 1 "Failed to get VPP packages!"
- }
- else
- echo Downloading VPP packages of specific version from NEXUS...
- DPDK_STABLE_VER=$(cat ${SCRIPT_DIR}/DPDK_STABLE_VER)
- VPP_STABLE_VER=$(cat ${SCRIPT_DIR}/VPP_STABLE_VER_UBUNTU)
- bash ${SCRIPT_DIR}/resources/tools/scripts/download_install_vpp_pkgs.sh \
- --skip-install --vpp ${VPP_STABLE_VER} --dkms ${DPDK_STABLE_VER} || {
- die 1 "Failed to get VPP packages!"
- }
- fi
- ;;
- vpp-csit-*)
- # Use local builded packages.
- mv ../${DUT}*.deb ${DOWNLOAD_DIR}/
- ;;
- *)
- die 1 "Unable to identify job type from: ${TEST_CODE}!"
- ;;
- esac
- # Extract VPP API to specific folder
- dpkg -x ${DOWNLOAD_DIR}/vpp_*.deb /tmp/vpp || {
- die 1 "Failed to extract ${DUT} package!"
- }
-
- LIGATO_REPO_URL="https://github.com/ligato/"
- VPP_AGENT_STABLE_VER=$(cat ${SCRIPT_DIR}/VPP_AGENT_STABLE_VER)
- DOCKER_DEB="docker-ce_18.03.0~ce-0~ubuntu_amd64.deb"
-
- # Clone & checkout stable vnf-agent
- cd ../..
- git clone -b ${VPP_AGENT_STABLE_VER} --single-branch \
- ${LIGATO_REPO_URL}/vpp-agent vpp-agent || {
- die 1 "Failed to run: git clone ${LIGATO_REPO_URL}/vpp-agent!"
- }
- cd vpp-agent
-
- # Install Docker
- wget -q https://download.docker.com/linux/ubuntu/dists/xenial/pool/stable/amd64/${DOCKER_DEB} || {
- die 1 "Failed to download Docker package!"
- }
-
- sudo dpkg -i ${DOCKER_DEB} || {
- die 1 "Failed to install Docker!"
- }
-
- # Pull ligato/dev_vpp_agent docker image and re-tag as local
- sudo docker pull ligato/dev-vpp-agent:${VPP_AGENT_STABLE_VER} || {
- die 1 "Failed to pull Docker image!"
- }
-
- sudo docker tag ligato/dev-vpp-agent:${VPP_AGENT_STABLE_VER}\
- dev_vpp_agent:latest || {
- die 1 "Failed to tag Docker image!"
- }
-
- # Start dev_vpp_agent container as daemon
- sudo docker run --rm -itd --name agentcnt dev_vpp_agent bash || {
- die 1 "Failed to run Docker image!"
- }
-
- # Copy latest vpp api into running container
- sudo docker cp /tmp/vpp/usr/share/vpp/api agentcnt:/usr/share/vpp || {
- die 1 "Failed to copy files Docker image!"
- }
-
- for f in ${DOWNLOAD_DIR}/*; do
- sudo docker cp $f agentcnt:/opt/vpp-agent/dev/vpp/build-root/ || {
- die 1 "Failed to copy files Docker image!"
- }
- done
-
- # Recompile vpp-agent
- sudo docker exec -i agentcnt \
- script -qec '. ~/.bashrc; cd /go/src/github.com/ligato/vpp-agent && make generate && make install' || {
- die 1 "Failed to build vpp-agent in Docker image!"
- }
- # Save container state
- sudo docker commit `sudo docker ps -q` dev_vpp_agent:latest || {
- die 1 "Failed to commit state of Docker image!"
- }
-
- # Build prod_vpp_agent docker image
- cd docker/prod/ &&\
- sudo docker build --tag prod_vpp_agent --no-cache . || {
- die 1 "Failed to build Docker image!"
- }
- # Export Docker image
- sudo docker save prod_vpp_agent | gzip > prod_vpp_agent.tar.gz || {
- die 1 "Failed to save Docker image!"
- }
- DOCKER_IMAGE="$( readlink -f prod_vpp_agent.tar.gz | tr '\n' ' ' )"
- rm -r ${DOWNLOAD_DIR}/vpp*
- mv ${DOCKER_IMAGE} ${DOWNLOAD_DIR}/
- ;;
- *dpdk*)
- DUT="dpdk"
-
- DPDK_REPO='https://fast.dpdk.org/rel/'
- # Use downloaded packages with specific version
- if [[ "$TEST_CODE" == *daily* ]] || \
- [[ "$TEST_CODE" == *weekly* ]] || \
- [[ "$TEST_CODE" == *timed* ]];
- then
- echo "Downloading latest DPDK packages from repo..."
- DPDK_STABLE_VER=$(wget --no-check-certificate --quiet -O - ${DPDK_REPO} | \
- grep -v '2015' | grep -Eo 'dpdk-[^\"]+xz' | tail -1)
- else
- echo "Downloading DPDK packages of specific version from repo..."
- DPDK_STABLE_VER='dpdk-18.05.tar.xz'
- fi
- if [[ ! -f ${DPDK_STABLE_VER} ]]; then
- wget --no-check-certificate ${DPDK_REPO}${DPDK_STABLE_VER} || {
- die 1 "Failed to get DPDK package from ${DPDK_REPO}!"
- }
- fi
- ;;
- *)
- die 1 "Unable to identify DUT type from: ${TEST_CODE}!"
- ;;
-esac
-cd ${SCRIPT_DIR}
-
-if [[ ! "$(ls -A ${DOWNLOAD_DIR})" ]]; then
- die 1 "No artifacts downloaded!"
-fi
-
-# ENVIRONMENT PREPARATION
-rm -rf env
-
-pip install virtualenv || {
- die 1 "Failed to install virtual env!"
-}
-virtualenv --system-site-packages env || {
- die 1 "Failed to create virtual env!"
-}
-source env/bin/activate || {
- die 1 "Failed to activate virtual env!"
-}
-pip install -r requirements.txt || {
- die 1 "Failed to install requirements to virtual env!"
-}
-
-# We iterate over available topologies and wait until we reserve topology.
-while :; do
- for TOPOLOGY in ${TOPOLOGIES};
- do
- python ${SCRIPT_DIR}/resources/tools/scripts/topo_reservation.py -t ${TOPOLOGY}
- if [ $? -eq 0 ]; then
- WORKING_TOPOLOGY=${TOPOLOGY}
- echo "Reserved: ${WORKING_TOPOLOGY}"
- # On script exit we clean testbed.
- trap "cancel_all ${WORKING_TOPOLOGY}" EXIT
- break
- fi
- done
-
- if [ -n "${WORKING_TOPOLOGY}" ]; then
- # Exit the infinite while loop if we made a reservation.
- break
- fi
-
- # Wait ~3minutes before next try.
- SLEEP_TIME=$[ ( $RANDOM % 20 ) + 180 ]s
- echo "Sleeping ${SLEEP_TIME}"
- sleep ${SLEEP_TIME}
-done
-
-# Clean testbed before execution.
-python ${SCRIPT_DIR}/resources/tools/scripts/topo_cleanup.py -t ${WORKING_TOPOLOGY} || {
- die 1 "Failed to cleanup topologies!"
-}
-
-# CSIT EXECUTION
-PYBOT_ARGS="--outputdir ${ARCHIVE_DIR} --loglevel TRACE --variable TOPOLOGY_PATH:${WORKING_TOPOLOGY} --suite tests.${DUT}.perf"
-
-# NIC SELECTION
-# All topologies NICs
-TOPOLOGIES_NICS=($(grep -hoPR "model: \K.*" topologies/available/* | sort -u))
-# Selected topology NICs
-TOPOLOGY_NICS=($(grep -hoPR "model: \K.*" ${WORKING_TOPOLOGY} | sort -u))
-# All topologies NICs - Selected topology NICs
-EXCLUDE_NICS=($(comm -13 <(printf '%s\n' "${TOPOLOGY_NICS[@]}") <(printf '%s\n' "${TOPOLOGIES_NICS[@]}")))
-
-case "$TEST_CODE" in
- # Select specific performance tests based on jenkins job type variable.
- *ndrpdr-weekly* )
- TEST_TAG_ARRAY=(ndrpdrAND64bAND1c
- ndrpdrAND78bAND1c)
- ;;
- *mrr-daily* | *mrr-weekly* )
- TEST_TAG_ARRAY=(mrrAND64bAND1c
- mrrAND64bAND2c
- mrrAND64bAND4c
- mrrAND78bAND1c
- mrrAND78bAND2c
- mrrAND78bAND4c
- mrrANDimixAND1cANDvhost
- mrrANDimixAND2cANDvhost
- mrrANDimixAND4cANDvhost
- mrrANDimixAND1cANDmemif
- mrrANDimixAND2cANDmemif
- mrrANDimixAND4cANDmemif)
- ;;
- * )
- if [[ -z "$TEST_TAG_STRING" ]]; then
- # If nothing is specified, we will run pre-selected tests by
- # following tags. Items of array will be concatenated by OR in Robot
- # Framework.
- TEST_TAG_ARRAY=(mrrANDnic_intel-x710AND1cAND64bANDip4base
- mrrANDnic_intel-x710AND1cAND78bANDip6base
- mrrANDnic_intel-x710AND1cAND64bANDl2bdbase)
- else
- # If trigger contains tags, split them into array.
- TEST_TAG_ARRAY=(${TEST_TAG_STRING//:/ })
- fi
- ;;
-esac
-
-# We will add excluded NICs.
-TEST_TAG_ARRAY+=("${EXCLUDE_NICS[@]/#/!NIC_}")
-
-TAGS=()
-
-# We will prefix with perftest to prevent running other tests (e.g. Functional).
-prefix="perftestAND"
-if [[ ${TEST_CODE} == vpp-* ]]; then
- # Automatic prefixing for VPP jobs to limit the NIC used and
- # traffic evaluation to MRR.
- prefix="${prefix}mrrANDnic_intel-x710AND"
-fi
-for TAG in "${TEST_TAG_ARRAY[@]}"; do
- if [[ ${TAG} == "!"* ]]; then
- # Exclude tags are not prefixed.
- TAGS+=("${TAG}")
- else
- TAGS+=("$prefix${TAG}")
- fi
-done
-
-# Catenate TAG selections
-EXPANDED_TAGS=()
-for TAG in "${TAGS[@]}"; do
- if [[ ${TAG} == "!"* ]]; then
- EXPANDED_TAGS+=(" --exclude ${TAG#$"!"} ")
- else
- EXPANDED_TAGS+=(" --include ${TOPOLOGIES_TAGS}AND${TAG} ")
- fi
-done
-
-# Execute the test
-pybot ${PYBOT_ARGS}${EXPANDED_TAGS[@]} tests/
-RETURN_STATUS=$(echo $?)
-
-# We will create additional archive if workspace variable is set. This way if
-# script is running in jenkins all will be automatically archived to logs.fd.io.
-if [[ -n ${WORKSPACE-} ]]; then
- cp -r ${ARCHIVE_DIR}/ $WORKSPACE/archives/
-fi
-
-exit ${RETURN_STATUS}
+# TODO: Delete this file, perhaps replacing it with a symlink.
+here=$(dirname $(readlink -e "${BASH_SOURCE[0]}"))
+source "${here}/resources/libraries/bash/entry/bootstrap_verify_perf.sh"