Whamcloud - gitweb
LU-6142 tests: Add missing /tmp/target2 under cleanup_src_tgt
[fs/lustre-release.git] / lustre / tests / recovery-mds-scale.sh
index ba3b5e6..fdda222 100644 (file)
@@ -1,6 +1,5 @@
 #!/bin/bash
-# vim:shiftwidth=4:softtabstop=4:tabstop=4:
-
+#
 # Was Test 11 in cmd3.
 # For duration of 24 hours repeatedly failover a random MDS at
 # 10 minute intervals and verify that no application errors occur.
@@ -10,243 +9,56 @@ set -e
 
 ONLY=${ONLY:-"$*"}
 
-# bug number for skipped test:
-ALWAYS_EXCEPT="$RECOVERY_MDS_SCALE_EXCEPT"
-# UPDATE THE COMMENT ABOVE WITH BUG NUMBERS WHEN CHANGING ALWAYS_EXCEPT!
-
-LUSTRE=${LUSTRE:-$(cd $(dirname $0)/..; echo $PWD)}
+LUSTRE=${LUSTRE:-$(dirname $0)/..}
 . $LUSTRE/tests/test-framework.sh
-init_test_env $@
-. ${CONFIG:=$LUSTRE/tests/cfg/$NAME.sh}
+. $LUSTRE/tests/recovery-scale-lib.sh
+init_test_env "$@"
 init_logging
 
-remote_mds_nodsh && skip_env "remote MDS with nodsh" && exit 0
-remote_ost_nodsh && skip_env "remote OST with nodsh" && exit 0
-
-[ -z "$CLIENTS" -o $CLIENTCOUNT -lt 3 ] &&
-    skip_env "need three or more clients" && exit 0
+# bug number for skipped test:
+ALWAYS_EXCEPT="$RECOVERY_MDS_SCALE_EXCEPT "
+# UPDATE THE COMMENT ABOVE WITH BUG NUMBERS WHEN CHANGING ALWAYS_EXCEPT!
 
-if [ -z "$SHARED_DIRECTORY" ] || ! check_shared_dir $SHARED_DIRECTORY; then
-    skip_env "SHARED_DIRECTORY should be specified with a shared directory \
-which is accessable on all of the nodes"
-    exit 0
-fi
+build_test_filter
 
-ERRORS_OK=""    # No application failures should occur during this test.
+remote_mds_nodsh && skip_env "remote MDS with nodsh"
 
-if [ "$SLOW" = "no" ]; then
-    DURATION=${DURATION:-$((60 * 30))}
-else
-    DURATION=${DURATION:-$((60 * 60 * 24))}
+if (( CLIENTCOUNT < 3 )); then
+       skip_env "need three or more clients"
 fi
-SERVER_FAILOVER_PERIOD=${SERVER_FAILOVER_PERIOD:-$((60 * 10))} # 10 minutes
-
-MINSLEEP=${MINSLEEP:-120}
-REQFAIL_PERCENT=${REQFAIL_PERCENT:-3}    # bug17839 comment 62
-REQFAIL=${REQFAIL:-$((DURATION / SERVER_FAILOVER_PERIOD *
-                      REQFAIL_PERCENT / 100))}
-
-END_RUN_FILE=${END_RUN_FILE:-$SHARED_DIRECTORY/end_run_file}
-LOAD_PID_FILE=${LOAD_PID_FILE:-$TMP/client-load.pid}
-VMSTAT_PID_FILE=${VMSTAT_PID_FILE:-$TMP/vmstat.pid}
-
-server_numfailovers () {
-    local facet=$1
-    local var=${facet}_numfailovers
-    local val=0
-
-    [[ ${!var} ]] && val=${!var}
-    echo $val
-}
-
-servers_numfailovers () {
-    local facet
-    local var
 
-    for facet in ${MDTS//,/ } ${OSTS//,/ }; do
-        echo "$facet: $(server_numfailovers $facet) times"
-    done
-}
-
-summary_and_cleanup () {
-    local rc=$?
-    trap 0
-
-    # Having not empty END_RUN_FILE means the failed loads only
-    if [ -s $END_RUN_FILE ]; then
-        print_end_run_file $END_RUN_FILE
-        rc=1
-    fi
-
-    echo $(date +'%F %H:%M:%S') Terminating clients loads ...
-    echo "$0" >> $END_RUN_FILE
-    local result=PASS
-    [ $rc -eq 0 ] || result=FAIL
-
-    log "Duration:               $DURATION
-Server failover period: $SERVER_FAILOVER_PERIOD seconds
-Exited after:           $ELAPSED seconds
-Number of failovers before exit:
-$(servers_numfailovers)
-Status: $result: rc=$rc"
-
-    # stop vmstat on OSS nodes
-    [ "$VMSTAT" ] && stop_process $(comma_list $(osts_nodes)) $VMSTAT_PID_FILE
-
-    # stop the client loads
-    stop_client_loads $NODES_TO_USE $LOAD_PID_FILE
-
-    if [ $rc -ne 0 ]; then
-        # we are interested in only on failed clients and servers
-        local failedclients=$(cat $END_RUN_FILE | grep -v $0)
-        # FIXME: need ostfailover-s nodes also for FLAVOR=OST
-        gather_logs $(comma_list $(osts_nodes) $(mdts_nodes) \
-                      $mdsfailover_HOST $failedclients)
-    fi
-
-    exit $rc
-}
-
-failover_target() {
-    local flavor=${1:-"MDS"}
-    local servers
-    local serverfacet
-    local var
-
-    [ "$flavor" = "MDS" ] && servers=$MDTS || servers=$OSTS
-
-    trap summary_and_cleanup EXIT INT
-
-    # start vmstat on OSS nodes
-    [ "$VMSTAT" ] && start_vmstat $(comma_list $(osts_nodes)) $VMSTAT_PID_FILE
-
-    # start client loads
-    rm -f $END_RUN_FILE
-    start_client_loads $NODES_TO_USE
-
-    echo client loads pids:
-    do_nodesv $NODES_TO_USE "cat $LOAD_PID_FILE" || exit 3
-
-    ELAPSED=0
-    local sleep=0
-    local reqfail=0
-    local it_time_start
-    local start_ts=$(date +%s)
-    local current_ts=$start_ts
-
-    while [ $ELAPSED -lt $DURATION -a ! -e $END_RUN_FILE ]; do
-        # In order to perform the
-        # expected number of failovers, we need to account the following:
-        # 1) the time that has elapsed during the client load checking
-        # 2) time takes for failover
-        it_time_start=$(date +%s)
-
-        serverfacet=$(get_random_entry $servers)
-        var=${serverfacet}_numfailovers
-
-        # Check that our client loads are still running. If any have died,
-        # that means they have died outside of recovery, which is unacceptable.
-        log "==== Checking the clients loads BEFORE failover -- failure NOT OK \
-             ELAPSED=$ELAPSED DURATION=$DURATION PERIOD=$SERVER_FAILOVER_PERIOD"
-        check_client_loads $NODES_TO_USE || exit 4
-
-        log "Wait $serverfacet recovery complete before doing next failover..."
-        if ! wait_recovery_complete $serverfacet; then
-            echo "$serverfacet recovery is not completed!"
-            exit 7
-        fi
-
-        log "Checking clients are in FULL state before doing next failover..."
-        if ! wait_clients_import_state $NODES_TO_USE $serverfacet FULL; then
-            echo "Clients import not FULL, please consider to increase \
-SERVER_FAILOVER_PERIOD=$SERVER_FAILOVER_PERIOD!"
-        fi
-
-        log "Starting failover on $serverfacet"
-        facet_failover "$serverfacet" || exit 1
-
-        # Check that our client loads are still running during failover.
-        # No application failures should occur.
-        log "==== Checking the clients loads AFTER failover -- failure NOT OK"
-        if ! check_client_loads $NODES_TO_USE; then
-            log "Client load failed during failover. Exiting..."
-            exit 5
-        fi
-
-        # Increment the number of failovers.
-        val=$((${!var} + 1))
-        eval $var=$val
-
-        current_ts=$(date +%s)
-        ELAPSED=$((current_ts - start_ts))
-
-        sleep=$((SERVER_FAILOVER_PERIOD - (current_ts - it_time_start)))
-
-        # Keep counting the number of iterations when
-        # time spent to failover and two client loads check exceeded
-        # the value ( SERVER_FAILOVER_PERIOD - MINSLEEP ).
-        if [ $sleep -lt $MINSLEEP ]; then
-            reqfail=$((reqfail + 1))
-            log "WARNING: failover and two check_client_loads time exceeded \
-SERVER_FAILOVER_PERIOD - MINSLEEP!
-Failed to load the filesystem with I/O for a minimum period of \
-$MINSLEEP $reqfail times ( REQFAIL=$REQFAIL ).
-This iteration, the load was only applied for sleep=$sleep seconds.
-Estimated max recovery time: $MAX_RECOV_TIME
-Probably the hardware is taking excessively long time to boot.
-Try to increase SERVER_FAILOVER_PERIOD (current is $SERVER_FAILOVER_PERIOD), \
-bug 20918"
-            [ $reqfail -gt $REQFAIL ] && exit 6
-        fi
-
-        log "$serverfacet has failed over ${!var} times, and counting..."
-
-        [ $((ELAPSED + sleep)) -ge $DURATION ] && break
-
-        if [ $sleep -gt 0 ]; then
-            echo "sleeping $sleep seconds... "
-            sleep $sleep
-        fi
-    done
-    exit 0
-}
+# SHARED_DIRECTORY should be specified with a shared directory which is
+# accessable on all of the nodes
+if [[ -z "$SHARED_DIRECTORY" ]] || ! check_shared_dir "$SHARED_DIRECTORY"; then
+       skip_env "SHARED_DIRECTORY not set"
+fi
 
-################################## Main Flow ###################################
-build_test_filter
+ERRORS_OK=""    # No application failures should occur during this test.
 
 check_and_setup_lustre
 rm -rf $DIR/[Rdfs][0-9]*
 
-MAX_RECOV_TIME=$(max_recovery_time)
-
-# The test node needs to be insulated from a lustre failure as much as possible,
-# so not even loading the lustre modules is ideal.
-# -- umount lustre
-# -- remove hostname from clients list
-zconf_umount $HOSTNAME $MOUNT
-NODES_TO_USE=${NODES_TO_USE:-$CLIENTS}
-NODES_TO_USE=$(exclude_items_from_list $NODES_TO_USE $HOSTNAME)
-
-check_progs_installed $NODES_TO_USE ${CLIENT_LOADS[@]}
+insulate_clients
+check_progs_installed $NODES_TO_USE "${CLIENT_LOADS[@]}"
 
+MAX_RECOV_TIME=$(max_recovery_time)
 MDTS=$(get_facets MDS)
 OSTS=$(get_facets OST)
 
+# Print informaiton about settings
+run_info $SERVER_FAILOVER_PERIOD $DURATION $MINSLEEP $SLOW $REQFAIL \
+       $SHARED_DIRECTORY $END_RUN_FILE $LOAD_PID_FILE $VMSTAT_PID_FILE \
+       $CLIENTCOUNT $MDTS $OSTS
+
 test_failover_mds() {
-    # failover a random MDS
-    failover_target MDS
+       # failover a random MDS
+       failover_target MDS
 }
 run_test failover_mds "failover MDS"
 
-test_failover_ost() {
-    # failover a random OST
-    failover_target OST
-}
-run_test failover_ost "failover OST"
-
 zconf_mount $HOSTNAME $MOUNT || error "mount $MOUNT on $HOSTNAME failed"
 client_up || error "start client on $HOSTNAME failed"
 
-complete $(basename $0) $SECONDS
+complete_test $SECONDS
 check_and_cleanup_lustre
 exit_status