X-Git-Url: https://git.whamcloud.com/?a=blobdiff_plain;f=lustre%2Ftests%2Finsanity.sh;h=f0cb4036522c02dbfdf4cb6e9a8c28a156a92687;hb=b031b90a0320401db1ae474c4cbd89226bf5f2c9;hp=9e392c774fdc70539e99328b87ca823f4742d1f0;hpb=50e3ab3c73cd70f6616a5baf4fa6a4c1a84a7e56;p=fs%2Flustre-release.git diff --git a/lustre/tests/insanity.sh b/lustre/tests/insanity.sh index 9e392c7..f0cb403 100755 --- a/lustre/tests/insanity.sh +++ b/lustre/tests/insanity.sh @@ -8,29 +8,40 @@ LUSTRE=${LUSTRE:-`dirname $0`/..} init_test_env $@ -. ${CONFIG:=$LUSTRE/tests/cfg/insanity-local.sh} +. ${CONFIG:=$LUSTRE/tests/cfg/$NAME.sh} +init_logging +# +ALWAYS_EXCEPT="10 $INSANITY_EXCEPT" + +if [ "$FAILURE_MODE" = "HARD" ]; then + skip_env "$0: is not functional with FAILURE_MODE = HARD, please use recovery-double-scale, bz20407" + exit 0 +fi + +# +[ "$SLOW" = "no" ] && EXCEPT_SLOW="" -ALWAYS_EXCEPT="10" +SETUP=${SETUP:-""} +CLEANUP=${CLEANUP:-""} build_test_filter -assert_env MDSCOUNT mds1_HOST ost1_HOST ost2_HOST client_HOST LIVE_CLIENT +SINGLECLIENT=${SINGLECLIENT:-$HOSTNAME} +LIVE_CLIENT=${LIVE_CLIENT:-$SINGLECLIENT} +FAIL_CLIENTS=${FAIL_CLIENTS:-$RCLIENTS} -#### -# Initialize all the ostN_HOST -NUMOST=2 -if [ "$EXTRA_OSTS" ]; then - for host in $EXTRA_OSTS; do - NUMOST=$((NUMOST + 1)) - OST=ost$NUMOST - eval ${OST}_HOST=$host - done -fi +assert_env mds_HOST MDS_MKFS_OPTS +assert_env ost_HOST OST_MKFS_OPTS OSTCOUNT +assert_env LIVE_CLIENT FSNAME -# This can be a regexp, to allow more clients -CLIENTS=${CLIENTS:-"`comma_list $LIVE_CLIENT $FAIL_CLIENTS $EXTRA_CLIENTS`"} +require_dsh_mds || exit 0 +require_dsh_ost || exit 0 + +# FAIL_CLIENTS list should not contain the LIVE_CLIENT +FAIL_CLIENTS=$(echo " $FAIL_CLIENTS " | sed -re "s/\s+$LIVE_CLIENT\s+/ /g") DIR=${DIR:-$MOUNT} +TESTDIR=$DIR/d0.$(basename $0 .sh) ##### # fail clients round robin @@ -49,33 +60,16 @@ set_fail_client() { echo "fail $FAIL_CLIENT, next is $FAIL_NEXT" } -shutdown_client() { - client=$1 - if [ "$FAILURE_MODE" = HARD ]; then - $POWER_DOWN $client - while ping -w 3 -c 1 $client > /dev/null 2>&1; do - echo "waiting for node $client to fail" - sleep 1 - done - elif [ "$FAILURE_MODE" = SOFT ]; then - zconf_umount $client $MOUNT -f - fi -} - -reboot_node() { - NODE=$1 - if [ "$FAILURE_MODE" = HARD ]; then - $POWER_UP $NODE - fi -} - fail_clients() { num=$1 + + log "Request clients to fail: ${num}. Num of clients to fail: ${FAIL_NUM}, already failed: $DOWN_NUM" if [ -z "$num" ] || [ "$num" -gt $((FAIL_NUM - DOWN_NUM)) ]; then num=$((FAIL_NUM - DOWN_NUM)) fi if [ -z "$num" ] || [ "$num" -le 0 ]; then + log "No clients failed!" return fi @@ -91,7 +85,7 @@ fail_clients() { echo "down clients: $DOWN_CLIENTS" for client in $DOWN_CLIENTS; do - reboot_node $client + boot_node $client done DOWN_NUM=`echo $DOWN_CLIENTS | wc -w` client_rmdirs @@ -107,61 +101,8 @@ reintegrate_clients() { DOWN_NUM=0 } -gen_config() { - rm -f $XMLCONFIG - if [ "$MDSCOUNT" -gt 1 ]; then - add_lmv lmv1 - for mds in `mds_list`; do - MDSDEV=$TMP/${mds}-`hostname` - add_mds $mds --dev $MDSDEV --size $MDSSIZE --lmv lmv1 - done - MDS=lmv1 - add_lov_to_lmv lov1 lmv1 --stripe_sz $STRIPE_BYTES \ - --stripe_cnt $STRIPES_PER_OBJ --stripe_pattern 0 - else - add_mds mds1 --dev $MDSDEV --size $MDSSIZE - if [ ! -z "$mds1failover_HOST" ]; then - add_mdsfailover mds1 --dev $MDSDEV --size $MDSSIZE - fi - add_lov lov1 mds1 --stripe_sz $STRIPE_BYTES \ - --stripe_cnt $STRIPES_PER_OBJ --stripe_pattern 0 - MDS=mds1_svc - fi - - for i in `seq $NUMOST`; do - dev=`printf $OSTDEV $i` - add_ost ost$i --lov lov1 --dev $dev --size $OSTSIZE \ - --journal-size $OSTJOURNALSIZE - done - - add_client client --mds $MDS --lov lov1 --path $MOUNT -} - -setup() { - rm -rf logs/* - for i in `seq $NUMOST`; do - wait_for ost$i - start ost$i ${REFORMAT} $OSTLCONFARGS - done - [ "$DAEMONFILE" ] && $LCTL debug_daemon start $DAEMONFILE $DAEMONSIZE - for mds in `mds_list`; do - wait_for $mds - start $mds $MDSLCONFARGS ${REFORMAT} - done - while ! do_node $CLIENTS "ls -d $LUSTRE" > /dev/null; do sleep 5; done - grep " $MOUNT " /proc/mounts || zconf_mount $CLIENTS $MOUNT - -} - -cleanup() { - zconf_umount $CLIENTS $MOUNT - - for mds in `mds_list`; do - stop $mds ${FORCE} $MDSLCONFARGS || : - done - for i in `seq $NUMOST`; do - stop ost$i ${REFORMAT} ${FORCE} $OSTLCONFARGS || : - done +start_ost() { + start ost$1 `ostdevname $1` $OST_MOUNT_OPTS } trap exit INT @@ -170,29 +111,28 @@ client_touch() { file=$1 for c in $LIVE_CLIENT $FAIL_CLIENTS; do if echo $DOWN_CLIENTS | grep -q $c; then continue; fi - $PDSH $c touch $MOUNT/${c}_$file || return 1 + $PDSH $c touch $TESTDIR/${c}_$file || return 1 done } client_rm() { file=$1 for c in $LIVE_CLIENT $FAIL_CLIENTS; do - $PDSH $c rm $MOUNT/${c}_$file + $PDSH $c rm $TESTDIR/${c}_$file done } client_mkdirs() { for c in $LIVE_CLIENT $FAIL_CLIENTS; do - echo "$c mkdir $MOUNT/$c" - $PDSH $c "mkdir $MOUNT/$c" - $PDSH $c "ls -l $MOUNT/$c" + echo "$c mkdir $TESTDIR/$c" + $PDSH $c "mkdir $TESTDIR/$c && ls -l $TESTDIR/$c" done } client_rmdirs() { for c in $LIVE_CLIENT $FAIL_CLIENTS; do - echo "rmdir $MOUNT/$c" - $PDSH $LIVE_CLIENT "rmdir $MOUNT/$c" + echo "rmdir $TESTDIR/$c" + $PDSH $LIVE_CLIENT "rmdir $TESTDIR/$c" done } @@ -201,60 +141,20 @@ clients_recover_osts() { # do_node $CLIENTS "$LCTL "'--device %OSC_`hostname`_'"${facet}_svc_MNT_client_facet recover" } -node_to_ost() { - node=$1 - retvar=$2 - for i in `seq $NUMOST`; do - ostvar="ost${i}_HOST" - if [ "${!ostvar}" == $node ]; then - eval $retvar=ost${i} - return 0 - fi - done - echo "No ost found for node; $node" - return 1 - -} - - - -if [ "$ONLY" == "cleanup" ]; then - cleanup - exit -fi - -if [ -z "$NOSETUP" ]; then - gen_config - setup -fi - -if [ ! -z "$EVAL" ]; then - eval "$EVAL" - exit $? -fi +check_and_setup_lustre -if [ "$ONLY" == "setup" ]; then - exit 0 -fi +rm -rf $TESTDIR +mkdir -p $TESTDIR # 9 Different Failure Modes Combinations echo "Starting Test 17 at `date`" test_0() { - echo "Failover MDS" - facet_failover mds1 - echo "Waiting for df pid: $DFPID" - wait $DFPID || return 1 - - echo "Failing OST1" - facet_failover ost1 - echo "Waiting for df pid: $DFPID" - wait $DFPID || return 2 - - echo "Failing OST2" - facet_failover ost2 - echo "Waiting for df pid: $DFPID" - wait $DFPID || return 3 + fail $SINGLEMDS + + for i in $(seq $OSTCOUNT) ; do + fail ost$i + done return 0 } run_test 0 "Fail all nodes, independently" @@ -270,37 +170,36 @@ run_test 1 "MDS/MDS failure" ############### Second Failure Mode ############### test_2() { echo "Verify Lustre filesystem is up and running" - client_df + [ -z "$(mounted_lustre_filesystems)" ] && error "Lustre is not running" + + clients_up - echo "Failing MDS" - shutdown_facet mds1 - reboot_facet mds1 + shutdown_facet $SINGLEMDS + reboot_facet $SINGLEMDS # prepare for MDS failover - change_active mds1 - reboot_facet mds1 + change_active $SINGLEMDS + reboot_facet $SINGLEMDS - client_df & + clients_up & DFPID=$! sleep 5 - echo "Failing OST" shutdown_facet ost1 echo "Reintegrating OST" reboot_facet ost1 - wait_for ost1 - start ost1 + wait_for_facet ost1 + start_ost 1 || return 2 - echo "Failover MDS" - wait_for mds1 - start mds1 + wait_for_facet $SINGLEMDS + start $SINGLEMDS `mdsdevname 1` $MDS_MOUNT_OPTS || return $? #Check FS wait $DFPID clients_recover_osts ost1 echo "Verify reintegration" - client_df || return 1 + clients_up || return 1 } run_test 2 "Second Failure Mode: MDS/OST `date`" @@ -311,14 +210,14 @@ run_test 2 "Second Failure Mode: MDS/OST `date`" test_3() { #Create files echo "Verify Lustre filesystem is up and running" + [ -z "$(mounted_lustre_filesystems)" ] && error "Lustre is not running" #MDS Portion - facet_failover mds1 - wait $DFPID || echo df failed: $? + fail $SINGLEMDS #Check FS echo "Test Lustre stability after MDS failover" - client_df + clients_up #CLIENT Portion echo "Failing 2 CLIENTS" @@ -326,13 +225,14 @@ test_3() { #Check FS echo "Test Lustre stability after CLIENT failure" - client_df + clients_up #Reintegration echo "Reintegrating CLIENTS" reintegrate_clients || return 1 - client_df || return 3 + clients_up || return 3 + sleep 2 # give it a little time for fully recovered before next test } run_test 3 "Thirdb Failure Mode: MDS/CLIENT `date`" ################################################### @@ -342,83 +242,91 @@ test_4() { echo "Fourth Failure Mode: OST/MDS `date`" #OST Portion - echo "Failing OST ost1" shutdown_facet ost1 #Check FS echo "Test Lustre stability after OST failure" - client_df + clients_up & + DFPIDA=$! + sleep 5 #MDS Portion - echo "Failing MDS" - shutdown_facet mds1 - reboot_facet mds1 + shutdown_facet $SINGLEMDS + reboot_facet $SINGLEMDS # prepare for MDS failover - change_active mds1 - reboot_facet mds1 + change_active $SINGLEMDS + reboot_facet $SINGLEMDS - client_df & - DFPID=$! + clients_up & + DFPIDB=$! sleep 5 #Reintegration echo "Reintegrating OST" reboot_facet ost1 - wait_for ost1 - start ost1 - - echo "Failover MDS" - wait_for mds1 - start mds1 + wait_for_facet ost1 + start_ost 1 + + wait_for_facet $SINGLEMDS + start $SINGLEMDS `mdsdevname 1` $MDS_MOUNT_OPTS #Check FS - wait $DFPID + wait $DFPIDA + wait $DFPIDB clients_recover_osts ost1 echo "Test Lustre stability after MDS failover" - client_df || return 1 + clients_up || return 1 } run_test 4 "Fourth Failure Mode: OST/MDS `date`" ################################################### ############### Fifth Failure Mode ############### test_5() { + [ $OSTCOUNT -lt 2 ] && skip_env "$OSTCOUNT < 2, not enough OSTs" && return 0 + echo "Fifth Failure Mode: OST/OST `date`" #Create files echo "Verify Lustre filesystem is up and running" - client_df + [ -z "$(mounted_lustre_filesystems)" ] && error "Lustre is not running" + + clients_up #OST Portion - echo "Failing OST" shutdown_facet ost1 reboot_facet ost1 #Check FS echo "Test Lustre stability after OST failure" - client_df + clients_up & + DFPIDA=$! + sleep 5 #OST Portion - echo "Failing OST" shutdown_facet ost2 reboot_facet ost2 #Check FS echo "Test Lustre stability after OST failure" - client_df + clients_up & + DFPIDB=$! + sleep 5 #Reintegration echo "Reintegrating OSTs" - wait_for ost1 - start ost1 - wait_for ost2 - start ost2 + wait_for_facet ost1 + start_ost 1 + wait_for_facet ost2 + start_ost 2 clients_recover_osts ost1 clients_recover_osts ost2 sleep $TIMEOUT - client_df || return 2 + wait $DFPIDA + wait $DFPIDB + clients_up || return 2 } run_test 5 "Fifth Failure Mode: OST/OST `date`" ################################################### @@ -429,17 +337,21 @@ test_6() { #Create files echo "Verify Lustre filesystem is up and running" - client_df || return 1 + [ -z "$(mounted_lustre_filesystems)" ] && error "Lustre is not running" + + clients_up client_touch testfile || return 2 #OST Portion - echo "Failing OST" shutdown_facet ost1 reboot_facet ost1 #Check FS echo "Test Lustre stability after OST failure" - client_df + clients_up & + DFPIDA=$! + echo DFPIDA=$DFPIDA + sleep 5 #CLIENT Portion echo "Failing CLIENTs" @@ -447,17 +359,25 @@ test_6() { #Check FS echo "Test Lustre stability after CLIENTs failure" - client_df + clients_up & + DFPIDB=$! + echo DFPIDB=$DFPIDB + sleep 5 #Reintegration echo "Reintegrating OST/CLIENTs" - wait_for ost1 - start ost1 - reintegrate_clients + wait_for_facet ost1 + start_ost 1 + reintegrate_clients || return 1 sleep 5 + wait_remote_prog "stat -f" $((TIMEOUT * 3 + 20)) + wait $DFPIDA + wait $DFPIDB + echo "Verifying mount" - client_df || return 3 + [ -z "$(mounted_lustre_filesystems)" ] && return 3 + clients_up } run_test 6 "Sixth Failure Mode: OST/CLIENT `date`" ################################################### @@ -469,7 +389,9 @@ test_7() { #Create files echo "Verify Lustre filesystem is up and running" - client_df + [ -z "$(mounted_lustre_filesystems)" ] && error "Lustre is not running" + + clients_up client_touch testfile || return 1 #CLIENT Portion @@ -478,9 +400,9 @@ test_7() { #Check FS echo "Test Lustre stability after CLIENTs failure" - client_df - $PDSH $LIVE_CLIENT "ls -l $MOUNT" - $PDSH $LIVE_CLIENT "rm -f $MOUNT/*_testfile" + clients_up + $PDSH $LIVE_CLIENT "ls -l $TESTDIR" + $PDSH $LIVE_CLIENT "rm -f $TESTDIR/*_testfile" #Sleep echo "Wait 1 minutes" @@ -488,23 +410,21 @@ test_7() { #Create files echo "Verify Lustre filesystem is up and running" - client_df + [ -z "$(mounted_lustre_filesystems)" ] && return 2 + + clients_up client_rm testfile #MDS Portion - echo "Failing MDS" - facet_failover mds1 + fail $SINGLEMDS - #Check FS - echo "Test Lustre stability after MDS failover" - wait $DFPID || echo "df on down clients fails " || return 1 - $PDSH $LIVE_CLIENT "ls -l $MOUNT" - $PDSH $LIVE_CLIENT "rm -f $MOUNT/*_testfile" + $PDSH $LIVE_CLIENT "ls -l $TESTDIR" + $PDSH $LIVE_CLIENT "rm -f $TESTDIR/*_testfile" #Reintegration echo "Reintegrating CLIENTs" - reintegrate_clients - client_df || return 2 + reintegrate_clients || return 2 + clients_up #Sleep echo "wait 1 minutes" @@ -520,7 +440,9 @@ test_8() { #Create files echo "Verify Lustre filesystem is up and running" - client_df + [ -z "$(mounted_lustre_filesystems)" ] && error "Lustre is not running" + + clients_up client_touch testfile #CLIENT Portion @@ -529,9 +451,9 @@ test_8() { #Check FS echo "Test Lustre stability after CLIENTs failure" - client_df - $PDSH $LIVE_CLIENT "ls -l $MOUNT" - $PDSH $LIVE_CLIENT "rm -f $MOUNT/*_testfile" + clients_up + $PDSH $LIVE_CLIENT "ls -l $TESTDIR" + $PDSH $LIVE_CLIENT "rm -f $TESTDIR/*_testfile" #Sleep echo "Wait 1 minutes" @@ -539,27 +461,32 @@ test_8() { #Create files echo "Verify Lustre filesystem is up and running" - client_df + [ -z "$(mounted_lustre_filesystems)" ] && error "Lustre is not running" + + clients_up client_touch testfile #OST Portion - echo "Failing OST" shutdown_facet ost1 reboot_facet ost1 #Check FS echo "Test Lustre stability after OST failure" - client_df - $PDSH $LIVE_CLIENT "ls -l $MOUNT" - $PDSH $LIVE_CLIENT "rm -f $MOUNT/*_testfile" + clients_up & + DFPID=$! + sleep 5 + #non-failout hangs forever here + #$PDSH $LIVE_CLIENT "ls -l $TESTDIR" + #$PDSH $LIVE_CLIENT "rm -f $TESTDIR/*_testfile" #Reintegration echo "Reintegrating CLIENTs/OST" - reintegrate_clients - wait_for ost1 - start ost1 - client_df || return 1 + reintegrate_clients || return 3 + wait_for_facet ost1 + start_ost 1 + wait $DFPID + clients_up || return 1 client_touch testfile2 || return 2 #Sleep @@ -576,7 +503,9 @@ test_9() { #Create files echo "Verify Lustre filesystem is up and running" - client_df + [ -z "$(mounted_lustre_filesystems)" ] && error "Lustre is not running" + + clients_up client_touch testfile || return 1 #CLIENT Portion @@ -585,9 +514,9 @@ test_9() { #Check FS echo "Test Lustre stability after CLIENTs failure" - client_df - $PDSH $LIVE_CLIENT "ls -l $MOUNT" || return 1 - $PDSH $LIVE_CLIENT "rm -f $MOUNT/*_testfile" || return 2 + clients_up + $PDSH $LIVE_CLIENT "ls -l $TESTDIR" || return 1 + $PDSH $LIVE_CLIENT "rm -f $TESTDIR/*_testfile" || return 2 #Sleep echo "Wait 1 minutes" @@ -595,7 +524,7 @@ test_9() { #Create files echo "Verify Lustre filesystem is up and running" - $PDSH $LIVE_CLIENT df $MOUNT || return 3 + client_up $LIVE_CLIENT || return 3 client_touch testfile || return 4 #CLIENT Portion @@ -604,14 +533,14 @@ test_9() { #Check FS echo "Test Lustre stability after CLIENTs failure" - client_df - $PDSH $LIVE_CLIENT "ls -l $MOUNT" || return 5 - $PDSH $LIVE_CLIENT "rm -f $MOUNT/*_testfile" || return 6 + clients_up + $PDSH $LIVE_CLIENT "ls -l $TESTDIR" || return 5 + $PDSH $LIVE_CLIENT "rm -f $TESTDIR/*_testfile" || return 6 #Reintegration echo "Reintegrating CLIENTs/CLIENTs" - reintegrate_clients - client_df || return 7 + reintegrate_clients || return 7 + clients_up #Sleep echo "Wait 1 minutes" @@ -628,5 +557,6 @@ test_10() { } run_test 10 "Running Availability for 6 hours..." -equals_msg "Done, cleaning up" -cleanup +complete $(basename $0) $SECONDS +check_and_cleanup_lustre +exit_status