X-Git-Url: https://git.whamcloud.com/?p=fs%2Flustre-release.git;a=blobdiff_plain;f=lustre%2Ftests%2Frecovery-small.sh;h=cc5447940e085dd17762301a15ff73cbe8abb008;hp=ee148e06c293680af62f748c301a4662ec7b46f8;hb=301d76a71176c186129231ddd1323bae21100165;hpb=5f3e926ac9ff8ad134ad920d0e8545e16395ef3b diff --git a/lustre/tests/recovery-small.sh b/lustre/tests/recovery-small.sh index ee148e0..cc54479 100755 --- a/lustre/tests/recovery-small.sh +++ b/lustre/tests/recovery-small.sh @@ -2,29 +2,18 @@ set -e -# bug 5493 LU2034 -ALWAYS_EXCEPT="52 60 $RECOVERY_SMALL_EXCEPT" - -export MULTIOP=${MULTIOP:-multiop} PTLDEBUG=${PTLDEBUG:--1} -LUSTRE=${LUSTRE:-`dirname $0`/..} +LUSTRE=${LUSTRE:-$(dirname $0)/..} . $LUSTRE/tests/test-framework.sh init_test_env $@ -. ${CONFIG:=$LUSTRE/tests/cfg/$NAME.sh} init_logging -require_dsh_mds || exit 0 - -# also long tests: 19, 21a, 21e, 21f, 23, 27 -# 1 2.5 2.5 4 4 (min)" -[ "$SLOW" = "no" ] && EXCEPT_SLOW="17 26a 26b 50 51 57" - -[ $(facet_fstype $SINGLEMDS) = "zfs" ] && -# bug number for skipped test: LU-2194 LU-2547 - ALWAYS_EXCEPT="$ALWAYS_EXCEPT 19b 24a 24b" +ALWAYS_EXCEPT="$RECOVERY_SMALL_EXCEPT " build_test_filter +require_dsh_mds || exit 0 + # Allow us to override the setup if we already have a mounted system by # setting SETUP=" " and CLEANUP=" " SETUP=${SETUP:-""} @@ -149,21 +138,181 @@ test_9() { run_test 9 "pause bulk on OST (bug 1420)" #bug 1521 -test_10() { - do_facet client mcreate $DIR/$tfile || - { error "mcreate failed: $?"; return 1; } - drop_bl_callback "chmod 0777 $DIR/$tfile" || echo "evicted as expected" - # wait for the mds to evict the client - #echo "sleep $(($TIMEOUT*2))" - #sleep $(($TIMEOUT*2)) - do_facet client touch $DIR/$tfile || echo "touch failed, evicted" - do_facet client checkstat -v -p 0777 $DIR/$tfile || - { error "client checkstat failed: $?"; return 3; } - do_facet client "munlink $DIR/$tfile" - # allow recovery to complete - client_up || client_up || sleep $TIMEOUT +test_10a() { + local before=$(date +%s) + local evict + + do_facet client "stat $DIR > /dev/null" || + error "failed to stat $DIR: $?" + drop_bl_callback "chmod 0777 $DIR" || + error "failed to chmod $DIR: $?" + + # let the client reconnect + client_reconnect + evict=$(do_facet client $LCTL get_param mdc.$FSNAME-MDT*.state | + awk -F"[ [,]" '/EVICTED ]$/ { if (mx<$5) {mx=$5;} } END { print mx }') + [ ! -z "$evict" ] && [[ $evict -gt $before ]] || + (do_facet client $LCTL get_param mdc.$FSNAME-MDT*.state; + error "no eviction: $evict before:$before") + + do_facet client checkstat -v -p 0777 $DIR || + error "client checkstat failed: $?" } -run_test 10 "finish request on server after client eviction (bug 1521)" +run_test 10a "finish request on server after client eviction (bug 1521)" + +test_10b() { + local before=$(date +%s) + local evict + + [[ "$MDS1_VERSION" -lt $(version_code 2.6.53) ]] && + skip "Need MDS version at least 2.6.53" + do_facet client "stat $DIR > /dev/null" || + error "failed to stat $DIR: $?" + drop_bl_callback_once "chmod 0777 $DIR" || + error "failed to chmod $DIR: $?" + + # let the client reconnect + client_reconnect + evict=$(do_facet client $LCTL get_param mdc.$FSNAME-MDT*.state | + awk -F"[ [,]" '/EVICTED ]$/ { if (mx<$5) {mx=$5;} } END { print mx }') + + [ -z "$evict" ] || [[ $evict -le $before ]] || + (do_facet client $LCTL get_param mdc.$FSNAME-MDT*.state; + error "eviction happened: $evict before:$before") + + do_facet client checkstat -v -p 0777 $DIR || + error "client checkstat failed: $?" +} +run_test 10b "re-send BL AST" + +test_10c() { + local before=$(date +%s) + local evict + local mdccli + local mdcpath + local conn_uuid + local workdir + local pid + local rc + + workdir="${DIR}/${tdir}" + mkdir -p ${workdir} || error "can't create workdir $?" + stat ${workdir} > /dev/null || + error "failed to stat ${workdir}: $?" + mdtidx=$($LFS getdirstripe -i ${workdir}) + mdtname=$($LFS mdts ${workdir} | grep -e "^$mdtidx:" | + awk '{sub("_UUID", "", $2); print $2;}') + #assume one client + mdccli=$($LCTL dl | grep "${mdtname}-mdc" | awk '{print $4;}') + conn_uuid=$($LCTL get_param -n mdc.${mdccli}.conn_uuid) + mdcpath="mdc.${mdccli}.import=connection=${conn_uuid}" + + drop_bl_callback_once "chmod 0777 ${workdir}" & + pid=$! + + # let chmod blocked + sleep 1 + # force client reconnect + $LCTL set_param "${mdcpath}" + + # wait client reconnect + client_reconnect + wait $pid + rc=$? + evict=$($LCTL get_param mdc.${mdccli}.state | + awk -F"[ [,]" '/EVICTED]$/ { if (t<$4) {t=$4;} } END { print t }') + + [[ $evict -le $before ]] || + ( $LCTL get_param mdc.$FSNAME-MDT*.state; + error "eviction happened: $EVICT before:$BEFORE" ) + + [ $rc -eq 0 ] || error "chmod must finished OK" + checkstat -v -p 0777 "${workdir}" || + error "client checkstat failed: $?" +} +run_test 10c "re-send BL AST vs reconnect race (LU-5569)" + +test_10d() { + local before=$(date +%s) + local evict + + [[ "$MDS1_VERSION" -lt $(version_code 2.6.90) ]] && + skip "Need MDS version at least 2.6.90" + + # sleep 1 is to make sure that BEFORE is not equal to EVICTED below + sleep 1 + rm -f $TMP/$tfile + echo -n ", world" | dd of=$TMP/$tfile bs=1c seek=5 + + remount_client $MOUNT + mount_client $MOUNT2 + + cancel_lru_locks osc + $LFS setstripe -i 0 -c 1 $DIR1/$tfile + echo -n hello | dd of=$DIR1/$tfile bs=5 + + stat $DIR2/$tfile >& /dev/null + $LCTL set_param fail_err=71 + drop_bl_callback "echo -n \\\", world\\\" >> $DIR2/$tfile" + + client_reconnect + + cancel_lru_locks osc + cmp -l $DIR1/$tfile $DIR2/$tfile || error "file contents differ" + cmp -l $DIR1/$tfile $TMP/$tfile || error "wrong content found" + + evict=$(do_facet client $LCTL get_param osc.$FSNAME-OST0000*.state | \ + tr -d '\-\[\] ' | \ + awk -F"[ [,]" '/EVICTED$/ { if (mx<$1) {mx=$1;} } END { print mx }') + + [[ $evict -gt $before ]] || + (do_facet client $LCTL get_param osc.$FSNAME-OST0000*.state; + error "no eviction: $evict before:$before") + + $LCTL set_param fail_err=0 + rm $TMP/$tfile + umount_client $MOUNT2 +} +run_test 10d "test failed blocking ast" + +test_10e() +{ + [[ "$OST1_VERSION" -le $(version_code 2.8.58) ]] && + skip "Need OST version at least 2.8.59" + [ $CLIENTCOUNT -lt 2 ] && skip "need two clients" + [ $(facet_host client) == $(facet_host ost1) ] && + skip "need ost1 and client on different nodes" + local -a clients=(${CLIENTS//,/ }) + local client1=${clients[0]} + local client2=${clients[1]} + + $LFS setstripe -c 1 -i 0 $DIR/$tfile-1 $DIR/$tfile-2 + $MULTIOP $DIR/$tfile-1 Ow1048576c + +#define OBD_FAIL_LDLM_BL_CALLBACK_NET 0x305 + $LCTL set_param fail_loc=0x80000305 + +#define OBD_FAIL_LDLM_ENQUEUE_OLD_EXPORT 0x30e + do_facet ost1 "$LCTL set_param fail_loc=0x1000030e" + # hit OBD_FAIL_LDLM_ENQUEUE_OLD_EXPORT twice: + # 1. to return ENOTCONN from ldlm_handle_enqueue0 + # 2. to pause reconnect handling between resend and setting + # import to LUSTRE_IMP_FULL state + do_facet ost1 "$LCTL set_param fail_val=3" + + # client1 fails ro respond to bl ast + do_node $client2 "$MULTIOP $DIR/$tfile-1 Ow1048576c" & + MULTIPID=$! + + # ost1 returns error on enqueue, which causes client1 to reconnect + do_node $client1 "$MULTIOP $DIR/$tfile-2 Ow1048576c" || + error "multiop failed" + wait $MULTIPID + + do_facet ost1 "$LCTL set_param fail_loc=0" + do_facet ost1 "$LCTL set_param fail_val=0" +} +run_test 10e "re-send BL AST vs reconnect race 2" #bug 2460 # wake up a thread waiting for completion after eviction @@ -177,7 +326,8 @@ test_11(){ do_facet client $MULTIOP $DIR/$tfile or || { error "multiop read failed: $?"; return 3; } - drop_bl_callback $MULTIOP $DIR/$tfile Ow || echo "evicted as expected" + drop_bl_callback_once $MULTIOP $DIR/$tfile Ow || + echo "evicted as expected" do_facet client munlink $DIR/$tfile || { error "munlink failed: $?"; return 4; } @@ -264,10 +414,11 @@ test_16() { sleep $TIMEOUT do_facet client "cmp $TMP/$tfile $DIR/$tfile" || return 2 start_read_ahead + rm -f $TMP/$tfile } run_test 16 "timeout bulk put, don't evict client (2732)" -test_17() { +test_17a() { local at_max_saved=0 remote_ost_nodsh && skip "remote OST with nodsh" && return 0 @@ -301,7 +452,93 @@ test_17() { [ $at_max_saved -ne 0 ] && at_max_set $at_max_saved ost1 return 0 } -run_test 17 "timeout bulk get, don't evict client (2732)" +run_test 17a "timeout bulk get, don't evict client (2732)" + +test_17b() { + [ -z "$RCLIENTS" ] && skip "Needs multiple clients" && return 0 + + # get one of the clients from client list + local rcli=$(echo $RCLIENTS | cut -d ' ' -f 1) + local p="$TMP/$TESTSUITE-$TESTNAME.parameters" + local ldlm_enqueue_min=$(do_facet ost1 find /sys -name ldlm_enqueue_min) + [ -z "$ldlm_enqueue_min" ] && + skip "missing /sys/.../ldlm_enqueue_min" && return 0 + + $LFS setstripe -i 0 -c 1 -S 1048576 $DIR/$tfile || + error "setstripe failed" + $LFS setstripe -i 0 -c 1 -S 1048576 $DIR/${tfile}2 || + error "setstripe 2 failed" + + save_lustre_params ost1 "at_history" > $p + save_lustre_params ost1 "bulk_timeout" >> $p + local dev="${FSNAME}-OST0000" + save_lustre_params ost1 "obdfilter.$dev.brw_size" >> $p + local ldlm_enqueue_min_save=$(do_facet ost1 cat $ldlm_enqueue_min) + + local new_at_history=15 + + do_facet ost1 "$LCTL set_param at_history=$new_at_history" + do_facet ost1 "$LCTL set_param bulk_timeout=30" + do_facet ost1 "echo 30 > /sys/module/ptlrpc/parameters/ldlm_enqueue_min" + + # brw_size is required to be 4m so that bulk transfer timeout + # could be simulated with OBD_FAIL_PTLRPC_CLIENT_BULK_CB3 + local brw_size=$($LCTL get_param -n \ + osc.$FSNAME-OST0000-osc-[^M]*.import | + awk '/max_brw_size/{print $2}') + if [ $brw_size -ne 4194304 ] + then + save_lustre_params ost1 "obdfilter.$dev.brw_size" >> $p + do_facet ost1 "$LCTL set_param obdfilter.$dev.brw_size=4" + remount_client $MOUNT + fi + + # get service estimate expanded + #define OBD_FAIL_OST_BRW_PAUSE_PACK 0x224 + do_facet ost1 "$LCTL set_param fail_loc=0x80000224 fail_val=35" + echo "delay rpc servicing by 35 seconds" + dd if=/dev/zero of=$DIR/$tfile bs=1M count=1 conv=fdatasync + + local estimate + estimate=$($LCTL get_param -n osc.$dev-osc-*.timeouts | + awk '/portal 6/ {print $5}') + echo "service estimates increased to $estimate" + + # let current worst service estimate to get closer to obliteration + sleep $((new_at_history / 3)) + + # start i/o and simulate bulk transfer loss + #define OBD_FAIL_PTLRPC_CLIENT_BULK_CB3 0x520 + do_facet ost1 "$LCTL set_param fail_loc=0xa0000520 fail_val=1" + dd if=/dev/zero of=$DIR/$tfile bs=2M count=1 conv=fdatasync,notrunc & + local writedd=$! + + # start lock conflict handling + sleep $((new_at_history / 3)) + do_node $rcli "dd if=$DIR/$tfile of=/dev/null bs=1M count=1" & + local readdd=$! + + # obliterate the worst service estimate + sleep $((new_at_history / 3 + 1)) + dd if=/dev/zero of=$DIR/${tfile}2 bs=1M count=1 + + estimate=$($LCTL get_param -n osc.$dev-osc-*.timeouts | + awk '/portal 6/ {print $5}') + echo "service estimate dropped to $estimate" + + wait $writedd + [[ $? == 0 ]] || error "write failed" + wait $readdd + [[ $? == 0 ]] || error "read failed" + + restore_lustre_params <$p + if [ $brw_size -ne 4194304 ] + then + remount_client $MOUNT || error "remount_client failed" + fi + do_facet ost1 "echo $ldlm_enqueue_min_save > $ldlm_enqueue_min" +} +run_test 17b "timeout bulk get, dont evict client (3582)" test_18a() { [ -z ${ost2_svc} ] && skip_env "needs 2 osts" && return 0 @@ -332,7 +569,7 @@ test_18a() { rc=0 pgcache_empty || rc=2 $LCTL --device $osc2dev activate - rm -f $f + rm -f $f $TMP/$tfile return $rc } run_test 18a "manual ost invalidate clears page cache immediately" @@ -365,7 +602,7 @@ test_18b() { # cache after the client reconnects? rc=0 pgcache_empty || rc=2 - rm -f $f + rm -f $f $TMP/$tfile return $rc } run_test 18b "eviction and reconnect clears page cache (2766)" @@ -398,13 +635,13 @@ test_18c() { do_facet ost1 lctl set_param fail_loc=0x80000225 # force reconnect sleep 1 - df $MOUNT > /dev/null 2>&1 + $LFS df $MOUNT > /dev/null 2>&1 sleep 2 # my understanding is that there should be nothing in the page - # cache after the client reconnects? + # cache after the client reconnects? rc=0 pgcache_empty || rc=2 - rm -f $f + rm -f $f $TMP/$tfile return $rc } run_test 18c "Dropped connect reply after eviction handing (14755)" @@ -427,8 +664,9 @@ test_19a() { # let the client reconnect client_reconnect - EVICT=$(do_facet client $LCTL get_param mdc.$FSNAME-MDT*.state | \ - awk -F"[ [,]" '/EVICTED]$/ { if (mx<$4) {mx=$4;} } END { print mx }') + EVICT=$(do_facet client $LCTL get_param mdc.$FSNAME-MDT*.state | + awk -F"[ [,]" '/EVICTED ]$/ \ + { if (mx<$5) {mx=$5;} } END { print mx }') [ ! -z "$EVICT" ] && [[ $EVICT -gt $BEFORE ]] || (do_facet client $LCTL get_param mdc.$FSNAME-MDT*.state; @@ -456,8 +694,9 @@ test_19b() { # let the client reconnect client_reconnect - EVICT=$(do_facet client $LCTL get_param osc.$FSNAME-OST*.state | \ - awk -F"[ [,]" '/EVICTED]$/ { if (mx<$4) {mx=$4;} } END { print mx }') + EVICT=$(do_facet client $LCTL get_param osc.$FSNAME-OST*.state | + awk -F"[ [,]" '/EVICTED ]$/ \ + { if (mx < $5) {mx = $5;} } END { print mx }') [ ! -z "$EVICT" ] && [[ $EVICT -gt $BEFORE ]] || (do_facet client $LCTL get_param osc.$FSNAME-OST*.state; @@ -489,7 +728,7 @@ test_19c() { # let the client reconnect sleep 5 EVICT=$(do_facet client $LCTL get_param mdc.$FSNAME-MDT*.state | - awk -F"[ [,]" '/EVICTED]$/ { if (mx<$4) {mx=$4;} } END { print mx }') + awk -F"[ [,]" '/EVICTED ]$/ { if (mx<$5) {mx=$5;} } END { print mx }') [ -z "$EVICT" ] || [[ $EVICT -le $BEFORE ]] || error "eviction happened" } @@ -787,7 +1026,8 @@ test_24b() { dmesg -c > /dev/null mkdir -p $DIR/$tdir - lfs setstripe $DIR/$tdir -s 0 -i 0 -c 1 + lfs setstripe $DIR/$tdir -S 0 -i 0 -c 1 || + error "$LFS setstripe failed" cancel_lru_locks osc multiop_bg_pause $DIR/$tdir/$tfile-1 Ow8192_yc || error "mulitop Ow8192_yc failed" @@ -829,20 +1069,23 @@ test_26a() { # was test_26 bug 5921 - evict dead exports by pinger check_timeout || return 1 - local OST_NEXP=$(do_facet ost1 lctl get_param -n obdfilter.${ost1_svc}.num_exports | cut -d' ' -f2) - - echo starting with $OST_NEXP OST exports + # make sure all imports are connected and not IDLE + do_facet client lfs df > /dev/null # OBD_FAIL_PTLRPC_DROP_RPC 0x505 do_facet client lctl set_param fail_loc=0x505 - # evictor takes PING_EVICT_TIMEOUT + 3 * PING_INTERVAL to evict. - # But if there's a race to start the evictor from various obds, - # the loser might have to wait for the next ping. - + local before=$(date +%s) local rc=0 - wait_client_evicted ost1 $OST_NEXP $((TIMEOUT * 2 + TIMEOUT * 3 / 4)) - rc=$? + + # evictor takes PING_EVICT_TIMEOUT + 3 * PING_INTERVAL to evict. + # But if there's a race to start the evictor from various obds, + # the loser might have to wait for the next ping. + sleep $((TIMEOUT * 2 + TIMEOUT * 3 / 4)) do_facet client lctl set_param fail_loc=0x0 - [ $rc -eq 0 ] || error "client not evicted from OST" + do_facet client lfs df > /dev/null + + local oscs=$(lctl dl | awk '/-osc-/ {print $4}') + check_clients_evicted $before ${oscs[@]} + check_clients_full 10 ${oscs[@]} } run_test 26a "evict dead exports" @@ -876,10 +1119,9 @@ test_26b() { # bug 10140 - evict dead exports by pinger # = 9 * PING_INTERVAL + PING_INTERVAL # = 10 PING_INTERVAL = 10 obd_timeout / 4 = 2.5 obd_timeout # let's wait $((TIMEOUT * 3)) # bug 19887 - local rc=0 - wait_client_evicted ost1 $OST_NEXP $((TIMEOUT * 3)) || \ - error "Client was not evicted by ost" rc=1 - wait_client_evicted $SINGLEMDS $MDS_NEXP $((TIMEOUT * 3)) || \ + wait_client_evicted ost1 $OST_NEXP $((TIMEOUT * 3)) || + error "Client was not evicted by ost" + wait_client_evicted $SINGLEMDS $MDS_NEXP $((TIMEOUT * 3)) || error "Client was not evicted by mds" } run_test 26b "evict dead exports" @@ -908,7 +1150,8 @@ run_test 27 "fail LOV while using OSC's" test_28() { # bug 6086 - error adding new clients do_facet client mcreate $DIR/$tfile || return 1 - drop_bl_callback "chmod 0777 $DIR/$tfile" ||echo "evicted as expected" + drop_bl_callback_once "chmod 0777 $DIR/$tfile" || + echo "evicted as expected" #define OBD_FAIL_MDS_CLIENT_ADD 0x12f do_facet $SINGLEMDS "lctl set_param fail_loc=0x8000012f" # fail once (evicted), reconnect fail (fail_loc), ok @@ -982,7 +1225,7 @@ test_51() { for i in $SEQ do #echo failover in $i sec - log "test_$testnum: failover in $i sec" + log "$TESTNAME: failover in $i sec" sleep $i facet_failover $SINGLEMDS done @@ -1039,21 +1282,22 @@ run_test 52 "failover OST under load" # test of open reconstruct test_53() { touch $DIR/$tfile - drop_ldlm_reply "openfile -f O_RDWR:O_CREAT -m 0755 $DIR/$tfile" ||\ + drop_mdt_ldlm_reply "openfile -f O_RDWR:O_CREAT -m 0755 $DIR/$tfile" ||\ return 2 } run_test 53 "touch: drop rep" test_54() { - zconf_mount `hostname` $MOUNT2 - touch $DIR/$tfile - touch $DIR2/$tfile.1 - sleep 10 - cat $DIR2/$tfile.missing # save transno = 0, rc != 0 into last_rcvd - fail $SINGLEMDS - umount $MOUNT2 - ERROR=`dmesg | egrep "(test 54|went back in time)" | tail -n1 | grep "went back in time"` - [ x"$ERROR" == x ] || error "back in time occured" + zconf_mount $(hostname) $MOUNT2 + touch $DIR/$tfile + touch $DIR2/$tfile.1 + sleep 10 + cat $DIR2/$tfile.missing # save transno = 0, rc != 0 into last_rcvd + fail $SINGLEMDS + umount $MOUNT2 + ERROR=$(dmesg | egrep "(test 54|went back in time)" | tail -n1 | + grep "went back in time") + [ x"$ERROR" == x ] || error "back in time occured" } run_test 54 "back in time" @@ -1063,10 +1307,17 @@ test_55() { mkdir -p $DIR/$tdir + # This test assumes relatively small max_dirty_mb setting + # which we want to walk away from, so just for it we will + # temporarily lower the value + local max_dirty_mb=$(lctl get_param -n osc.*.max_dirty_mb | head -n 1) + lctl set_param -n osc.*.max_dirty_mb=32 + stack_trap "lctl set_param osc.*.max_dirty_mb=$max_dirty_mb" EXIT + # Minimum pass speed is 2MBps local ddtimeout=64 # LU-2887/LU-3089 - set min pass speed to 500KBps - [ "$(facet_fstype ost1)" = "zfs" ] && ddtimeout=256 + [ "$ost1_FSTYPE" = zfs ] && ddtimeout=256 # first dd should be finished quickly $LFS setstripe -c 1 -i 0 $DIR/$tdir/$tfile-1 @@ -1122,11 +1373,11 @@ test_56() { # b=11277 #define OBD_FAIL_MDS_RESEND 0x136 touch $DIR/$tfile do_facet $SINGLEMDS "lctl set_param fail_loc=0x80000136" - stat $DIR/$tfile + stat $DIR/$tfile || error "stat failed" do_facet $SINGLEMDS "lctl set_param fail_loc=0" rm -f $DIR/$tfile } -run_test 56 "do not allow reconnect to busy exports" +run_test 56 "do not fail on getattr resend" test_57_helper() { # no oscs means no client or mdt @@ -1160,10 +1411,10 @@ test_58() { # bug 11546 pid=$! sleep 1 lctl set_param fail_loc=0 - drop_bl_callback rm -f $DIR/$tfile + drop_bl_callback_once rm -f $DIR/$tfile wait $pid # the first 'df' could tigger the eviction caused by - # 'drop_bl_callback', and it's normal case. + # 'drop_bl_callback_once', and it's normal case. # but the next 'df' should return successfully. do_facet client "df $DIR" || do_facet client "df $DIR" } @@ -1187,63 +1438,47 @@ test_59() { # bug 10589 } run_test 59 "Read cancel race on client eviction" -err17935 () { - # we assume that all md changes are in the MDT0 changelog - if [ $MDSCOUNT -gt 1 ]; then - error_ignore bz17935 $* - else - error $* - fi -} - test_60() { - MDT0=$($LCTL get_param -n mdc.*.mds_server_uuid | \ - awk '{gsub(/_UUID/,""); print $1}' | head -1) - - NUM_FILES=15000 - mkdir -p $DIR/$tdir + local num_files=${COUNT:-5000} + test_mkdir $DIR/$tdir # Register (and start) changelog - USER=$(do_facet $SINGLEMDS lctl --device $MDT0 changelog_register -n) - echo "Registered as $MDT0 changelog user $USER" + changelog_register || error "changelog_register failed" # Generate a large number of changelog entries - createmany -o $DIR/$tdir/$tfile $NUM_FILES + createmany -o $DIR/$tdir/$tfile $num_files sync sleep 5 # Unlink files in the background - unlinkmany $DIR/$tdir/$tfile $NUM_FILES & - CLIENT_PID=$! + unlinkmany $DIR/$tdir/$tfile $num_files & + local client_pid=$! sleep 1 # Failover the MDS while unlinks are happening facet_failover $SINGLEMDS # Wait for unlinkmany to finish - wait $CLIENT_PID + wait $client_pid # Check if all the create/unlink events were recorded # in the changelog - $LFS changelog $MDT0 >> $DIR/$tdir/changelog - local cl_count=$(grep UNLNK $DIR/$tdir/changelog | wc -l) - echo "$cl_count unlinks in $MDT0 changelog" - - do_facet $SINGLEMDS lctl --device $MDT0 changelog_deregister $USER - USERS=$(( $(do_facet $SINGLEMDS lctl get_param -n \ - mdd.$MDT0.changelog_users | wc -l) - 2 )) - if [ $USERS -eq 0 ]; then - [ $cl_count -eq $NUM_FILES ] || \ - err17935 "Recorded ${cl_count} unlinks out of $NUM_FILES" - # Also make sure we can clear large changelogs - cl_count=$($LFS changelog $FSNAME | wc -l) - [ $cl_count -le 2 ] || \ - error "Changelog not empty: $cl_count entries" - else - # If there are other users, there may be other unlinks in the log - [ $cl_count -ge $NUM_FILES ] || \ - err17935 "Recorded ${cl_count} unlinks out of $NUM_FILES" - echo "$USERS other changelog users; can't verify clear" + local cl_count=$(changelog_dump | grep -c UNLNK) + echo "$cl_count unlinks in changelog" + + changelog_deregister || error "changelog_deregister failed" + if ! changelog_users $SINGLEMDS | grep -q "^cl"; then + [ $cl_count -eq $num_files ] || + error "Recorded $cl_count of $num_files unlinks" + # Also make sure we can clear large changelogs + cl_count=$(changelog_dump | wc -l) + [ $cl_count -le 2 ] || + error "Changelog not empty: $cl_count entries" + else # If other users, there may be other unlinks in the log + [ $cl_count -ge $num_files ] || + error "Recorded $cl_count of $num_files unlinks" + changelog_users $SINGLEMDS + echo "other changelog users; can't verify clear" fi } run_test 60 "Add Changelog entries during MDS failover" @@ -1262,12 +1497,14 @@ test_61() replay_barrier $SINGLEMDS createmany -o $DIR/$tdir/$tfile-%d 10 - local oid=`do_facet ost1 "lctl get_param -n obdfilter.${ost1_svc}.last_id"` + local oid=$(do_facet ost1 "lctl get_param -n \ + obdfilter.${ost1_svc}.last_id" | sed -e 's/.*://') fail_abort $SINGLEMDS - + touch $DIR/$tdir/$tfile - local id=`$LFS getstripe $DIR/$tdir/$tfile | awk '$1 == 0 { print $2 }'` + local id=$($LFS getstripe $DIR/$tdir/$tfile | + awk '$1 == 0 { print $2 }') [ $id -le $oid ] && error "the orphan objid was reused, failed" # Cleanup @@ -1285,6 +1522,92 @@ run_test 61 "Verify to not reuse orphan objects - bug 17025" #} #run_test 62 "Verify connection flags race - bug LU-1716" +test_65() { + mount_client $DIR2 + + #grant lock1, export2 + $LFS setstripe -i -0 $DIR2/$tfile || error "setstripe failed" + $MULTIOP $DIR2/$tfile Ow || error "multiop failed" + +#define OBD_FAIL_LDLM_BL_EVICT 0x31e + do_facet ost $LCTL set_param fail_loc=0x31e + #get waiting lock2, export1 + $MULTIOP $DIR/$tfile Ow & + PID1=$! + # let enqueue to get asleep + sleep 2 + + #get lock2 blocked + $MULTIOP $DIR2/$tfile Ow & + PID2=$! + sleep 2 + + #evict export1 + ost_evict_client + + sleep 2 + do_facet ost $LCTL set_param fail_loc=0 + + wait $PID1 + wait $PID2 + + umount_client $DIR2 +} +run_test 65 "lock enqueue for destroyed export" + +test_66() +{ + [[ "$MDS1_VERSION" -ge $(version_code 2.7.51) ]] || + skip "Need MDS version at least 2.7.51" + + local list=$(comma_list $(osts_nodes)) + + # modify dir so that next revalidate would not obtain UPDATE lock + touch $DIR + + # drop 1 reply with UPDATE lock + mcreate $DIR/$tfile || error "mcreate failed: $?" + drop_mdt_ldlm_reply_once "stat $DIR/$tfile" & + sleep 2 + + # make the re-sent lock to sleep +#define OBD_FAIL_MDS_RESEND 0x136 + do_nodes $list $LCTL set_param fail_loc=0x80000136 + + #initiate the re-connect & re-send + local mdtname="MDT0000" + local mdccli=$($LCTL dl | grep "${mdtname}-mdc" | awk '{print $4;}') + local conn_uuid=$($LCTL get_param -n mdc.${mdccli}.conn_uuid) + $LCTL set_param "mdc.${mdccli}.import=connection=${conn_uuid}" + sleep 2 + + #initiate the client eviction while enqueue re-send is in progress + mds_evict_client + + client_reconnect + wait +} +run_test 66 "lock enqueue re-send vs client eviction" + +test_67() +{ +#define OBD_FAIL_PTLRPC_CONNECT_RACE 0x531 + $LCTL set_param fail_loc=0x80000531 + + local mdtname="MDT0000" + local mdccli=$($LCTL dl | grep "${mdtname}-mdc" | awk '{print $4;}') + local conn_uuid=$($LCTL get_param -n mdc.${mdccli}.mds_conn_uuid) + $LCTL set_param "mdc.${mdccli}.import=connection=${conn_uuid}" & + sleep 2 + + mds_evict_client + sleep 1 + + client_reconnect + wait +} +run_test 67 "connect vs import invalidate race" + check_cli_ir_state() { local NODE=${1:-$HOSTNAME} @@ -1303,6 +1626,11 @@ check_target_ir_state() local recovery_proc=obdfilter.${!name}.recovery_status local st + while : ; do + st=$(do_facet $target "$LCTL get_param -n $recovery_proc | + awk '/status:/{ print \\\$2}'") + [ x$st = xRECOVERING ] || break + done st=$(do_facet $target "lctl get_param -n $recovery_proc | awk '/IR:/{ print \\\$2}'") [ $st != ON -o $st != OFF -o $st != ENABLED -o $st != DISABLED ] || @@ -1384,8 +1712,8 @@ target_instance_match() local target=${srv}_svc local si=$(do_facet $srv lctl get_param -n $obdname.${!target}.instance) - local ci=$(lctl get_param -n $cliname.${!target}-${cliname}-*.import | \ - awk '/instance/{ print $2 }' |head -1) + local ci=$(lctl get_param -n $cliname.${!target}-${cliname}-*.import | + awk '/instance/{ print $2 }' | head -n1) return $([ $si -eq $ci ]) } @@ -1424,26 +1752,45 @@ test_100() } run_test 100 "IR: Make sure normal recovery still works w/o IR" -test_101() +test_101a() { - do_facet mgs $LCTL list_param mgs.*.ir_timeout || - { skip "MGS without IR support"; return 0; } + do_facet mgs $LCTL list_param mgs.*.ir_timeout || + skip "MGS without IR support" - set_ir_status full + set_ir_status full - local OST1_IMP=$(get_osc_import_name client ost1) + local ost1_imp=$(get_osc_import_name client ost1) - # disable pinger recovery - lctl set_param -n osc.$OST1_IMP.pinger_recov=0 + # disable pinger recovery + lctl set_param -n osc.$ost1_imp.pinger_recov=0 + stack_trap "$LCTL set_param -n osc.$ost1_imp.pinger_recov=1" EXIT - fail ost1 + fail ost1 - target_instance_match ost1 || error "instance mismatch" - nidtbl_versions_match || error "version must match" + target_instance_match ost1 || error "instance mismatch" + nidtbl_versions_match || error "version must match" +} +run_test 101a "IR: Make sure IR works w/o normal recovery" + +test_101b() +{ + do_facet mgs $LCTL list_param mgs.*.ir_timeout || + skip "MGS without IR support" - lctl set_param -n osc.$OST1_IMP.pinger_recov=1 + set_ir_status full + + local ost1_imp=$(get_osc_import_name client ost1) + +#define OBD_FAIL_OST_PREPARE_DELAY 0x247 + do_facet ost1 "$LCTL set_param fail_loc=0x247" + # disable pinger recovery + $LCTL set_param -n osc.$ost1_imp.pinger_recov=0 + stack_trap "$LCTL set_param -n osc.$ost1_imp.pinger_recov=1" EXIT + +#OST may return EAGAIN if it is not configured yet + fail ost1 } -run_test 101 "IR: Make sure IR works w/o normal recovery" +run_test 101b "IR: Make sure IR works w/o normal recovery and proceed EAGAIN" test_102() { @@ -1499,7 +1846,7 @@ test_103() do_facet mgs $LCTL list_param mgs.*.ir_timeout || { skip "MGS without IR support"; return 0; } - combined_mgs_mds && skip "mgs and mds on the same target" && return 0 + combined_mgs_mds && skip "needs separate mgs and mds" && return 0 # workaround solution to generate config log on the mds remount_facet mds1 @@ -1565,10 +1912,9 @@ test_105() # get one of the clients from client list local rcli=$(echo $RCLIENTS |cut -d' ' -f 1) - local old_MOUNTOPT=$MOUNTOPT - MOUNTOPT=${MOUNTOPT},noir + local mount_opts=${MOUNT_OPTS:+$MOUNT_OPTS,}noir zconf_umount $rcli $MOUNT || error "umount failed" - zconf_mount $rcli $MOUNT || error "mount failed" + zconf_mount $rcli $MOUNT $mount_opts || error "mount failed" # make sure lustre mount at $rcli disabling IR local ir_state=$(check_cli_ir_state $rcli) @@ -1578,7 +1924,7 @@ test_105() # Since the client just mounted, its last_rcvd entry is not on disk. # Send an RPC so exp_need_sync forces last_rcvd to commit this export # so the client can reconnect during OST recovery (LU-924, LU-1582) - $SETSTRIPE -i 0 $DIR/$tfile + $LFS setstripe -i 0 $DIR/$tfile dd if=/dev/zero of=$DIR/$tfile bs=1M count=1 conv=sync # make sure MGS's state is Partial @@ -1590,8 +1936,7 @@ test_105() [ $ir_state = "DISABLED" -o $ir_state = "OFF" ] || error "IR status on ost1 should be DISABLED" - # restore it - MOUNTOPT=$old_MOUNTOPT + # remount with the default MOUNT_OPTS zconf_umount $rcli $MOUNT || error "umount failed" zconf_mount $rcli $MOUNT || error "mount failed" @@ -1615,8 +1960,8 @@ cleanup_106() { } test_106() { # LU-1789 - [[ $(lustre_version_code $SINGLEMDS) -ge $(version_code 2.3.50) ]] || - { skip "Need MDS version at least 2.3.50"; return 0; } + [[ "$MDS1_VERSION" -ge $(version_code 2.3.50) ]] || + skip "Need MDS version at least 2.3.50" #define OBD_FAIL_MDC_LIGHTWEIGHT 0x805 $LCTL set_param fail_loc=0x805 @@ -1635,17 +1980,16 @@ test_106() { # LU-1789 touch $DIR2/$tfile || error "failed to create empty file" replay_barrier $SINGLEMDS - $LCTL set_param debug=console + $LCTL set_param debug=ha $LCTL clear facet_failover $SINGLEMDS - # lightweight connection must be evicted + # lightweight goes through LUSTRE_IMP_RECOVER during failover touch -c $DIR2/$tfile || true $LCTL dk $TMP/lustre-log-$TESTNAME.log - evicted=`awk '/This client was evicted by .*MDT0000/ { - print; - }' $TMP/lustre-log-$TESTNAME.log` - [ -z "$evicted" ] && error "lightweight client not evicted by mds" + recovered=$(awk '/MDT0000-mdc-[0-9a-f]*. lwp recover/ { print }' \ + $TMP/lustre-log-$TESTNAME.log) + [ -z "$recovered" ] && error "lightweight client was not recovered" # and all operations performed by lightweight client should be # synchronous, so the file created before mds restart should be there @@ -1680,16 +2024,41 @@ test_107 () { } run_test 107 "drop reint reply, then restart MDT" +test_108() { + mkdir -p $DIR/$tdir + $LFS setstripe -c 1 -i 0 $DIR/$tdir + + dd if=/dev/zero of=$DIR/$tdir/$tfile bs=1M count=256 & + local dd_pid=$! + sleep 0.1 + + ost_evict_client + + wait $dd_pid + + client_up || error "reconnect failed" + rm -f $DIR/$tdir/$tfile +} +run_test 108 "client eviction don't crash" + test_110a () { [ $MDSCOUNT -lt 2 ] && skip "needs >= 2 MDTs" && return 0 local remote_dir=$DIR/$tdir/remote_dir - local MDTIDX=1 + local mdtidx=1 + local num + + #prepare for 110 test, which need set striped dir on remote MDT. + for num in $(seq $MDSCOUNT); do + do_facet mds$num \ + lctl set_param -n mdt.$FSNAME*.enable_remote_dir=1 \ + 2>/dev/null + done mkdir -p $DIR/$tdir - drop_request "$LFS mkdir -i $MDTIDX $remote_dir" || - error "lfs mkdir failed" - local diridx=$($GETSTRIPE -M $remote_dir) - [ $diridx -eq $MDTIDX ] || error "$diridx != $MDTIDX" + drop_request "$LFS mkdir -i $mdtidx -c2 $remote_dir" || + error "lfs mkdir failed" + local diridx=$($LFS getstripe -m $remote_dir) + [ $diridx -eq $mdtidx ] || error "$diridx != $mdtidx" rm -rf $DIR/$tdir || error "rmdir failed" } @@ -1698,14 +2067,14 @@ run_test 110a "create remote directory: drop client req" test_110b () { [ $MDSCOUNT -lt 2 ] && skip "needs >= 2 MDTs" && return 0 local remote_dir=$DIR/$tdir/remote_dir - local MDTIDX=1 + local mdtidx=1 mkdir -p $DIR/$tdir - drop_reint_reply "$LFS mkdir -i $MDTIDX $remote_dir" || - error "lfs mkdir failed" + drop_reint_reply "$LFS mkdir -i $mdtidx -c2 $remote_dir" || + error "lfs mkdir failed" - diridx=$($GETSTRIPE -M $remote_dir) - [ $diridx -eq $MDTIDX ] || error "$diridx != $MDTIDX" + diridx=$($LFS getstripe -m $remote_dir) + [ $diridx -eq $mdtidx ] || error "$diridx != $mdtidx" rm -rf $DIR/$tdir || error "rmdir failed" } @@ -1714,14 +2083,14 @@ run_test 110b "create remote directory: drop Master rep" test_110c () { [ $MDSCOUNT -lt 2 ] && skip "needs >= 2 MDTs" && return 0 local remote_dir=$DIR/$tdir/remote_dir - local MDTIDX=1 + local mdtidx=1 mkdir -p $DIR/$tdir - drop_update_reply $MDTIDX "$LFS mkdir -i $MDTIDX $remote_dir" || - error "lfs mkdir failed" + drop_update_reply $mdtidx "$LFS mkdir -i $mdtidx -c2 $remote_dir" || + error "lfs mkdir failed" - diridx=$($GETSTRIPE -M $remote_dir) - [ $diridx -eq $MDTIDX ] || error "$diridx != $MDTIDX" + diridx=$($LFS getstripe -m $remote_dir) + [ $diridx -eq $mdtidx ] || error "$diridx != $mdtidx" rm -rf $DIR/$tdir || error "rmdir failed" } @@ -1733,7 +2102,7 @@ test_110d () { local MDTIDX=1 mkdir -p $DIR/$tdir - $LFS mkdir -i $MDTIDX $remote_dir || error "lfs mkdir failed" + $LFS mkdir -i $MDTIDX -c2 $remote_dir || error "lfs mkdir failed" drop_request "rm -rf $remote_dir" || error "rm remote dir failed" @@ -1749,7 +2118,7 @@ test_110e () { local MDTIDX=1 mkdir -p $DIR/$tdir - $LFS mkdir -i $MDTIDX $remote_dir || error "lfs mkdir failed" + $LFS mkdir -i $MDTIDX -c2 $remote_dir || error "lfs mkdir failed" drop_reint_reply "rm -rf $remote_dir" || error "rm remote dir failed" rm -rf $DIR/$tdir || error "rmdir failed" @@ -1764,7 +2133,7 @@ test_110f () { local MDTIDX=1 mkdir -p $DIR/$tdir - $LFS mkdir -i $MDTIDX $remote_dir || error "lfs mkdir failed" + $LFS mkdir -i $MDTIDX -c2 $remote_dir || error "lfs mkdir failed" drop_update_reply $MDTIDX "rm -rf $remote_dir" || error "rm remote dir failed" @@ -1772,19 +2141,183 @@ test_110f () { } run_test 110f "remove remote directory: drop slave rep" +test_110g () { + [[ "$MDS1_VERSION" -ge $(version_code 2.11.0) ]] || + skip "Need MDS version at least 2.11.0" + + [ $MDSCOUNT -lt 2 ] && skip "needs >= 2 MDTs" + + mkdir -p $DIR/$tdir + touch $DIR/$tdir/$tfile + + # OBD_FAIL_MDS_REINT_NET_REP 0x119 + do_facet mds1 $LCTL set_param fail_loc=0x119 + $LFS migrate -m 1 $DIR/$tdir & + migrate_pid=$! + sleep 5 + do_facet mds1 $LCTL set_param fail_loc=0 + wait $migrate_pid + + local mdt_index + mdt_index=$($LFS getstripe -m $DIR/$tdir) + [ $mdt_index == 1 ] || error "$tdir is not on MDT1" + mdt_index=$($LFS getstripe -m $DIR/$tdir/$tfile) + [ $mdt_index == 1 ] || error "$tfile is not on MDT1" + + rm -rf $DIR/$tdir || error "rmdir failed" +} +run_test 110g "drop reply during migration" + +test_110h () { + [ $MDSCOUNT -lt 2 ] && skip "needs >= 2 MDTs" + [[ "$MDS1_VERSION" -ge $(version_code 2.7.56) ]] || + skip "Need MDS version at least 2.7.56" + + local src_dir=$DIR/$tdir/source_dir + local tgt_dir=$DIR/$tdir/target_dir + local MDTIDX=1 + + mkdir -p $src_dir + $LFS mkdir -i $MDTIDX $tgt_dir + + dd if=/etc/hosts of=$src_dir/src_file + touch $tgt_dir/tgt_file + drop_update_reply $MDTIDX \ + "mrename $src_dir/src_file $tgt_dir/tgt_file" || + error "mrename failed" + + $CHECKSTAT -t file $src_dir/src_file && + error "src_file present after rename" + + diff /etc/hosts $tgt_dir/tgt_file || + error "file changed after rename" + +} +run_test 110h "drop update reply during cross-MDT file rename" + +test_110i () { + [ $MDSCOUNT -lt 2 ] && skip "needs >= 2 MDTs" + [[ "$MDS1_VERSION" -ge $(version_code 2.7.56) ]] || + skip "Need MDS version at least 2.7.56" + + local src_dir=$DIR/$tdir/source_dir + local tgt_dir=$DIR/$tdir/target_dir + local MDTIDX=1 + + mkdir -p $src_dir + $LFS mkdir -i $MDTIDX $tgt_dir + + mkdir $src_dir/src_dir + touch $src_dir/src_dir/a + mkdir $tgt_dir/tgt_dir + drop_update_reply $MDTIDX \ + "mrename $src_dir/src_dir $tgt_dir/tgt_dir" || + error "mrename failed" + + $CHECKSTAT -t dir $src_dir/src_dir && + error "src_dir present after rename" + + $CHECKSTAT -t dir $tgt_dir/tgt_dir || + error "tgt_dir not present after rename" + + $CHECKSTAT -t file $tgt_dir/tgt_dir/a || + error "a not present after rename" +} +run_test 110i "drop update reply during cross-MDT dir rename" + +test_110j () { + [ $MDSCOUNT -lt 2 ] && skip "needs >= 2 MDTs" + [[ "$MDS1_VERSION" -ge $(version_code 2.7.56) ]] || + skip "Need MDS version at least 2.7.56" + + local remote_dir=$DIR/$tdir/remote_dir + local local_dir=$DIR/$tdir/local_dir + local MDTIDX=1 + + mkdir -p $DIR/$tdir + mkdir $DIR/$tdir/local_dir + $LFS mkdir -i $MDTIDX $remote_dir + + touch $local_dir/local_file + drop_update_reply $MDTIDX \ + "ln $local_dir/local_file $remote_dir/remote_file" || + error "ln failed" + + $CHECKSTAT -t file $remote_dir/remote_file || + error "remote not present after ln" +} +run_test 110j "drop update reply during cross-MDT ln" + +test_110k() { + [ $MDSCOUNT -lt 2 ] && skip "needs >= 2 MDTS" + [[ "$MDS1_VERSION" -ge $(version_code 2.12.55) ]] || + skip "Need MDS version at least 2.12.55" + + umount $MOUNT + stop mds2 || error "stop mds2 failed" + +#define OBD_FAIL_FLD_QUERY_REQ 0x1103 + do_facet mds2 lctl set_param fail_loc=0x1103 + local OPTS="$MDS_MOUNT_OPTS -o abort_recovery" + start mds2 $(mdsdevname 2) $OPTS || + error "start MDS with abort_recovery should succeed" + do_facet mds2 lctl set_param fail_loc=0 + + # cleanup + stop mds2 || error "cleanup: stop mds2 failed" + start mds2 $(mdsdevname 2) $MDS_MOUNT_OPTS || + error "cleanup: start mds2 failed" + zconf_mount $(hostname) $MOUNT || error "cleanup: mount failed" + client_up || error "post-failover df failed" +} +run_test 110k "FID_QUERY failed during recovery" + +test_110m () { + (( $(lustre_version_code $SINGLEMDS) >= $(version_code 2.14.52) )) || + skip "Need MDS version at least 2.14.52" + (( $MDSCOUNT >= 2 )) || skip "needs at least 2 MDTs" + local remote_dir=$DIR/$tdir/remote_dir + local mdccli + local uuid + local diridx + + mkdir_on_mdt0 $DIR/$tdir + +#define OBD_FAIL_PTLRPC_RESEND_RACE 0x0525 + do_facet mds1 $LCTL set_param fail_loc=0x80000525 + $LFS mkdir -i 1 -c2 $remote_dir & + mkdir_pid=$! + sleep 3 + # initiate the re-connect & re-send + mdccli=$(do_facet mds2 $LCTL dl | + awk '/MDT0000-osp-MDT0001/ {print $4;}') + uuid=$(do_facet mds2 $LCTL get_param -n osp.$mdccli.mdt_conn_uuid) + echo "conn_uuid=$uuid" + do_facet mds2 $LCTL set_param "osp.$mdccli.import=connection=$uuid" + + wait $mkdir_pid + (( $? == 0 )) || error "mkdir failed" + + diridx=$($LFS getstripe -m $remote_dir) + (( $diridx == 1 )) || error "$diridx != 1" + rm -rf $DIR/$tdir || error "rmdir failed" +} +run_test 110m "update resent vs original RPC race" + # LU-2844 mdt prepare fail should not cause umount oops test_111 () { - [[ $(lustre_version_code $SINGLEMDS) -ge $(version_code 2.3.62) ]] || - { skip "Need MDS version at least 2.3.62"; return 0; } + [[ "$MDS1_VERSION" -ge $(version_code 2.3.62) ]] || + skip "Need MDS version at least 2.3.62" - local mdsdev=$(mdsdevname ${SINGLEMDS//mds/}) #define OBD_FAIL_MDS_CHANGELOG_INIT 0x151 do_facet $SINGLEMDS lctl set_param fail_loc=0x151 stop $SINGLEMDS || error "stop MDS failed" - start $SINGLEMDS $mdsdev && error "start MDS should fail" + start $SINGLEMDS $(mdsdevname ${SINGLEMDS//mds/}) $MDS_MOUNT_OPTS && + error "start MDS should fail" do_facet $SINGLEMDS lctl set_param fail_loc=0 - start $SINGLEMDS $mdsdev || error "start MDS failed" + start $SINGLEMDS $(mdsdevname ${SINGLEMDS//mds/}) $MDS_MOUNT_OPTS || + error "start MDS failed" } run_test 111 "mdd setup fail should not cause umount oops" @@ -1808,6 +2341,880 @@ test_112a() { } run_test 112a "bulk resend while orignal request is in progress" +test_115_read() { + local fail1=$1 + local fail2=$2 + + df $DIR + dd if=/dev/zero of=$DIR/$tfile bs=4096 count=1 + cancel_lru_locks osc + + # OST_READ = 3, + $LCTL set_param fail_loc=$fail1 fail_val=3 + dd of=/dev/null if=$DIR/$tfile bs=4096 count=1 & + pid=$! + sleep 1 + + set_nodes_failloc "$(osts_nodes)" $fail2 + + wait $pid || error "dd failed" + return 0 +} + +test_115_write() { + local fail1=$1 + local fail2=$2 + local error=$3 + local fail_val2=${4:-0} + + df $DIR + touch $DIR/$tfile + + # OST_WRITE = 4, + $LCTL set_param fail_loc=$fail1 fail_val=4 + dd if=/dev/zero of=$DIR/$tfile bs=4096 count=1 oflag=dsync & + pid=$! + sleep 1 + + df $MOUNT + set_nodes_failloc "$(osts_nodes)" $fail2 $fail_val2 + + wait $pid + rc=$? + [ $error -eq 0 ] && [ $rc -ne 0 ] && error "dd error ($rc)" + [ $error -ne 0 ] && [ $rc -eq 0 ] && error "dd success" + return 0 +} + +test_115a() { + [ "$OST1_VERSION" -lt $(version_code 2.8.50) ] && + skip "need at least 2.8.50 on OST" + + #define OBD_FAIL_PTLRPC_LONG_REQ_UNLINK 0x51b + #define OBD_FAIL_PTLRPC_DROP_BULK 0x51a + test_115_read 0x8000051b 0x8000051a +} +run_test 115a "read: late REQ MDunlink and no bulk" + +test_115b() { + [ "$OST1_VERSION" -lt $(version_code 2.8.50) ] && + skip "need at least 2.8.50 on OST" + + #define OBD_FAIL_PTLRPC_LONG_REQ_UNLINK 0x51b + #define OBD_FAIL_OST_ENOSPC 0x215 + + # pass $OSTCOUNT for the fail_loc to be caught + # appropriately by the IO thread + test_115_write 0x8000051b 0x80000215 1 $OSTCOUNT +} +run_test 115b "write: late REQ MDunlink and no bulk" + +test_115c() { + [ "$OST1_VERSION" -lt $(version_code 2.8.50) ] && + skip "need at least 2.8.50 on OST" + + #define OBD_FAIL_PTLRPC_LONG_REPL_UNLINK 0x50f + #define OBD_FAIL_PTLRPC_DROP_BULK 0x51a + test_115_read 0x8000050f 0x8000051a +} +run_test 115c "read: late Reply MDunlink and no bulk" + +test_115d() { + [ "$OST1_VERSION" -lt $(version_code 2.8.50) ] && + skip "need at least 2.8.50 on OST" + + #define OBD_FAIL_PTLRPC_LONG_REPL_UNLINK 0x50f + #define OBD_FAIL_OST_ENOSPC 0x215 + test_115_write 0x8000050f 0x80000215 0 +} +run_test 115d "write: late Reply MDunlink and no bulk" + +test_115e() { + [ "$OST1_VERSION" -lt $(version_code 2.8.50) ] && + skip "need at least 2.8.50 on OST" + + #define OBD_FAIL_PTLRPC_LONG_BULK_UNLINK 0x510 + #define OBD_FAIL_OST_ALL_REPLY_NET 0x211 + test_115_read 0x80000510 0x80000211 +} +run_test 115e "read: late Bulk MDunlink and no reply" + +test_115f() { + [ "$OST1_VERSION" -lt $(version_code 2.8.50) ] && + skip "need at least 2.8.50 on OST" + + #define OBD_FAIL_PTLRPC_LONG_REQ_UNLINK 0x51b + #define OBD_FAIL_OST_ALL_REPLY_NET 0x211 + test_115_read 0x8000051b 0x80000211 +} +run_test 115f "read: late REQ MDunlink and no reply" + +test_115g() { + [ "$OST1_VERSION" -lt $(version_code 2.8.50) ] && + skip "need at least 2.8.50 on OST" + + #define OBD_FAIL_PTLRPC_LONG_BOTH_UNLINK 0x51c + test_115_read 0x8000051c 0 +} +run_test 115g "read: late REQ MDunlink and Reply MDunlink" + +# parameters: fail_loc CMD RC +test_120_reply() { + local PID + local PID2 + local rc=5 + local fail + + #define OBD_FAIL_LDLM_CP_CB_WAIT2 0x320 + #define OBD_FAIL_LDLM_CP_CB_WAIT3 0x321 + #define OBD_FAIL_LDLM_CP_CB_WAIT4 0x322 + #define OBD_FAIL_LDLM_CP_CB_WAIT5 0x323 + + echo + echo -n "** FLOCK REPLY vs. EVICTION race, lock $2" + [ "$1" = "CLEANUP" ] && + fail=0x80000320 && echo ", $1 cp first" + [ "$1" = "REPLY" ] && + fail=0x80000321 && echo ", $1 cp first" + [ "$1" = "DEADLOCK CLEANUP" ] && + fail=0x80000322 && echo " DEADLOCK, CLEANUP cp first" + [ "$1" = "DEADLOCK REPLY" ] && + fail=0x80000323 && echo " DEADLOCK, REPLY cp first" + + if [ x"$2" = x"get" ]; then + #for TEST lock, take a conflict in advance + # sleep longer than evictor to not confuse fail_loc: 2+2+4 + echo "** Taking conflict **" + flocks_test 5 set read sleep 10 $DIR/$tfile & + PID2=$! + + sleep 2 + fi + + $LCTL set_param fail_loc=$fail + + flocks_test 5 $2 write $DIR/$tfile & + PID=$! + + sleep 2 + echo "** Evicting and re-connecting client **" + mds_evict_client + + client_reconnect + + if [ x"$2" = x"get" ]; then + wait $PID2 + fi + + wait $PID + rc=$? + + # check if the return value is allowed + [ $rc -eq $3 ] && rc=0 + + $LCTL set_param fail_loc=0 + return $rc +} + +# a lock is taken, unlock vs. cleanup_resource() race for destroying +# the ORIGINAL lock. +test_120_destroy() +{ + local PID + + flocks_test 5 set write sleep 4 $DIR/$tfile & + PID=$! + sleep 2 + + # let unlock to sleep in CP CB + $LCTL set_param fail_loc=$1 + sleep 4 + + # let cleanup to cleep in CP CB + mds_evict_client + + client_reconnect + + wait $PID + rc=$? + + $LCTL set_param fail_loc=0 + return $rc +} + +test_120() { + flock_is_enabled || { skip "mount w/o flock enabled" && return; } + touch $DIR/$tfile + + test_120_reply "CLEANUP" set 5 || error "SET race failed" + test_120_reply "CLEANUP" get 5 || error "GET race failed" + test_120_reply "CLEANUP" unlock 5 || error "UNLOCK race failed" + + test_120_reply "REPLY" set 5 || error "SET race failed" + test_120_reply "REPLY" get 5 || error "GET race failed" + test_120_reply "REPLY" unlock 5 || error "UNLOCK race failed" + + # DEADLOCK tests + test_120_reply "DEADLOCK CLEANUP" set 5 || error "DEADLOCK race failed" + test_120_reply "DEADLOCK REPLY" set 35 || error "DEADLOCK race failed" + + test_120_destroy 0x320 || error "unlock-cleanup race failed" +} +run_test 120 "flock race: completion vs. evict" + +test_113() { + local BEFORE=$(date +%s) + local EVICT + + # modify dir so that next revalidate would not obtain UPDATE lock + touch $DIR + + # drop 1 reply with UPDATE lock, + # resend should not create 2nd lock on server + mcreate $DIR/$tfile || error "mcreate failed: $?" + drop_mdt_ldlm_reply_once "stat $DIR/$tfile" || error "stat failed: $?" + + # 2 BL AST will be sent to client, both must find the same lock, + # race them to not get EINVAL for 2nd BL AST + #define OBD_FAIL_LDLM_PAUSE_CANCEL2 0x31f + $LCTL set_param fail_loc=0x8000031f + + $LCTL set_param ldlm.namespaces.*.early_lock_cancel=0 > /dev/null + chmod 0777 $DIR/$tfile || error "chmod failed: $?" + $LCTL set_param ldlm.namespaces.*.early_lock_cancel=1 > /dev/null + + # let the client reconnect + client_reconnect + EVICT=$($LCTL get_param mdc.$FSNAME-MDT*.state | + awk -F"[ [,]" '/EVICTED ]$/ { if (mx<$5) {mx=$5;} } END { print mx }') + + [ -z "$EVICT" ] || [[ $EVICT -le $BEFORE ]] || error "eviction happened" +} +run_test 113 "ldlm enqueue dropped reply should not cause deadlocks" + +T130_PID=0 +test_130_base() { + test_mkdir -p $DIR/$tdir + + # Prevent interference from layout intent RPCs due to + # asynchronous writeback. These will be tested in 130c below. + do_nodes ${CLIENTS:-$HOSTNAME} sync + + # get only LOOKUP lock on $tdir + cancel_lru_locks mdc + ls $DIR/$tdir/$tfile 2>/dev/null + + # get getattr by fid on $tdir + # + # we need to race with unlink, unlink must complete before we will + # take a DLM lock, otherwise unlink will wait until getattr will + # complete; but later than getattr starts so that getattr found + # the object +#define OBD_FAIL_MDS_INTENT_DELAY 0x160 + set_nodes_failloc "$(mdts_nodes)" 0x80000160 + stat $DIR/$tdir & + T130_PID=$! + sleep 2 + + rm -rf $DIR/$tdir + + # drop the reply so that resend happens on an unlinked file. +#define OBD_FAIL_MDS_LDLM_REPLY_NET 0x157 + set_nodes_failloc "$(mdts_nodes)" 0x80000157 +} + +test_130a() { + remote_mds_nodsh && skip "remote MDS with nodsh" + [[ "$MDS1_VERSION" -ge $(version_code 2.7.2) ]] || + skip "Need server version newer than 2.7.1" + + test_130_base + + wait $T130_PID || [ $? -eq 0 ] && error "stat should fail" + return 0 +} +run_test 130a "enqueue resend on not existing file" + +test_130b() { + remote_mds_nodsh && skip "remote MDS with nodsh" + [[ "$MDS1_VERSION" -ge $(version_code 2.7.2) ]] || + skip "Need server version newer than 2.7.1" + + test_130_base + # let the reply to be dropped + sleep 10 + +#define OBD_FAIL_SRV_ENOENT 0x217 + set_nodes_failloc "$(mdts_nodes)" 0x80000217 + + wait $T130_PID || [ $? -eq 0 ] && error "stat should fail" + return 0 +} +run_test 130b "enqueue resend on a stale inode" + +test_130c() { + remote_mds_nodsh && skip "remote MDS with nodsh" && return + + do_nodes ${CLIENTS:-$HOSTNAME} sync + echo XXX > $DIR/$tfile + + cancel_lru_locks mdc + + # Trigger writeback on $tfile. + # + # we need to race with unlink, unlink must complete before we will + # take a DLM lock, otherwise unlink will wait until intent will + # complete; but later than intent starts so that intent found + # the object +#define OBD_FAIL_MDS_INTENT_DELAY 0x160 + set_nodes_failloc "$(mdts_nodes)" 0x80000160 + sync & + T130_PID=$! + sleep 2 + + rm $DIR/$tfile + + # drop the reply so that resend happens on an unlinked file. +#define OBD_FAIL_MDS_LDLM_REPLY_NET 0x157 + set_nodes_failloc "$(mdts_nodes)" 0x80000157 + + # let the reply to be dropped + sleep 10 + +#define OBD_FAIL_SRV_ENOENT 0x217 + set_nodes_failloc "$(mdts_nodes)" 0x80000217 + + wait $T130_PID + + return 0 +} +run_test 130c "layout intent resend on a stale inode" + +test_132() { + local before=$(date +%s) + local evict + + mount_client $MOUNT2 || error "mount filed" + + rm -f $DIR/$tfile + # get a lock on client so that export would reach the stale list + $LFS setstripe -i 0 $DIR/$tfile || error "setstripe failed" + dd if=/dev/zero of=$DIR/$tfile bs=4096 count=1 conv=fsync || + error "dd failed" + + #define OBD_FAIL_OST_PAUSE_PUNCH 0x236 + do_facet ost1 $LCTL set_param fail_val=120 fail_loc=0x80000236 + + $TRUNCATE $DIR/$tfile 100 & + + sleep 1 + dd if=/dev/zero of=$DIR2/$tfile bs=4096 count=1 conv=notrunc || + error "dd failed" + + wait + umount_client $MOUNT2 + + evict=$(do_facet client $LCTL get_param \ + osc.$FSNAME-OST0000-osc-*/state | + awk -F"[ [,]" '/EVICTED ]$/ { if (t<$5) {t=$5;} } END { print t }') + + [ -z "$evict" ] || [[ $evict -le $before ]] || + (do_facet client $LCTL get_param \ + osc.$FSNAME-OST0000-osc-*/state; + error "eviction happened: $evict before:$before") +} +run_test 132 "long punch" + +test_131() { + remote_ost_nodsh && skip "remote OST with nodsh" && return 0 + + rm -f $DIR/$tfile + # get a lock on client so that export would reach the stale list + $LFS setstripe -i 0 $DIR/$tfile || error "setstripe failed" + dd if=/dev/zero of=$DIR/$tfile count=1 || error "dd failed" + + # another IO under the same lock + #define OBD_FAIL_OSC_DELAY_IO 0x414 + $LCTL set_param fail_loc=0x80000414 + $LCTL set_param fail_val=4 fail_loc=0x80000414 + dd if=/dev/zero of=$DIR/$tfile count=1 conv=notrunc oflag=dsync & + local pid=$! + sleep 1 + + #define OBD_FAIL_LDLM_BL_EVICT 0x31e + set_nodes_failloc "$(osts_nodes)" 0x8000031e + ost_evict_client + client_reconnect + + wait $pid && error "dd succeeded" + return 0 +} +run_test 131 "IO vs evict results to IO under staled lock" + +test_133() { + local list=$(comma_list $(mdts_nodes)) + + local t=$((TIMEOUT * 2)) + touch $DIR/$tfile + + flock $DIR/$tfile -c "echo bl lock;sleep $t;echo bl flock unlocked" & + sleep 1 + multiop_bg_pause $DIR/$tfile O_jc || return 1 + PID=$! + + #define OBD_FAIL_MDS_LDLM_REPLY_NET 0x157 + do_nodes $list $LCTL set_param fail_loc=0x80000157 + kill -USR1 $PID + echo "waiting for multiop $PID" + wait $PID || return 2 + + rm -f $DIR/$tfile + + return 0 +} +run_test 133 "don't fail on flock resend" + +test_134() { + [ -z "$CLIENTS" ] && skip "Need two or more clients" && return + [ $CLIENTCOUNT -lt 2 ] && + { skip "Need 2+ clients, have $CLIENTCOUNT" && return; } + + mkdir -p $MOUNT/$tdir/1 $MOUNT/$tdir/2 || error "mkdir failed" + touch $MOUNT/$tdir/1/$tfile $MOUNT/$tdir/2/$tfile || + error "touch failed" + zconf_umount_clients $CLIENTS $MOUNT + zconf_mount_clients $CLIENTS $MOUNT + +#define OBD_FAIL_TGT_REPLY_DATA_RACE 0x722 + # assume commit interval is 5 + do_facet mds1 "$LCTL set_param fail_loc=0x722 fail_val=5" + + local -a clients=(${CLIENTS//,/ }) + local client1=${clients[0]} + local client2=${clients[1]} + + do_node $client1 rm $MOUNT/$tdir/1/$tfile & + rmpid=$! + do_node $client2 mv $MOUNT/$tdir/2/$tfile $MOUNT/$tdir/2/${tfile}_2 & + mvpid=$! + fail mds1 + wait $rmpid || error "rm failed" + wait $mvpid || error "mv failed" + return 0 +} +run_test 134 "race between failover and search for reply data free slot" + +test_135() { + [ "$MDS1_VERSION" -lt $(version_code 2.12.51) ] && + skip "Need MDS version at least 2.12.51" + + mkdir -p $DIR/$tdir + $LFS setstripe -E 1M -L mdt $DIR/$tdir + # to have parent dir write lock before open/resend + touch $DIR/$tdir/$tfile + #define OBD_FAIL_MDS_LDLM_REPLY_NET 0x157 + do_nodes $(comma_list $(mdts_nodes)) $LCTL set_param fail_loc=0x80000157 + openfile -f O_RDWR:O_CREAT -m 0755 $DIR/$tdir/$tfile || + error "Failed to open DOM file" +} +run_test 135 "DOM: open/create resend to return size" + +test_136() { + remote_mds_nodsh && skip "remote MDS with nodsh" + [[ "$MDS1_VERSION" -ge $(version_code 2.12.52) ]] || + skip "Need MDS version at least 2.12.52" + + local mdts=$(comma_list $(mdts_nodes)) + local MDT0=$(facet_svc $SINGLEMDS) + + local clog=$(do_facet mds1 $LCTL --device $MDT0 changelog_register -n) + [ -n "$clog" ] || error "changelog_register failed" + cl_mask=$(do_facet mds1 $LCTL get_param \ + mdd.$MDT0.changelog_mask -n) + changelog_chmask "ALL" + + # generate some changelog records to accumulate + test_mkdir -i 0 -c 0 $DIR/$tdir || error "mkdir $tdir failed" + createmany -m $DIR/$tdir/$tfile 10000 || + error "create $DIR/$tdir/$tfile failed" + + local size1=$(do_facet $SINGLEMDS \ + $LCTL get_param -n mdd.$MDT0.changelog_size) + echo "Changelog size $size1" + + #define OBD_FAIL_LLOG_PURGE_DELAY 0x1318 + do_nodes $mdts $LCTL set_param fail_loc=0x1318 fail_val=30 + + # launch changelog_deregister in background on MDS + do_facet mds1 "nohup $LCTL --device $MDT0 changelog_deregister $clog \ + > foo.out 2> foo.err < /dev/null &" + # give time to reach fail_loc + sleep 15 + + # fail_loc will make MDS sleep in the middle of changelog_deregister + # take this opportunity to abruptly kill MDS + FAILURE_MODE_save=$FAILURE_MODE + FAILURE_MODE=HARD + fail mds1 + FAILURE_MODE=$FAILURE_MODE_save + + do_nodes $mdts $LCTL set_param fail_loc=0x0 fail_val=0 + + local size2=$(do_facet $SINGLEMDS \ + $LCTL get_param -n mdd.$MDT0.changelog_size) + echo "Changelog size $size2" + local clog2=$(do_facet $SINGLEMDS "$LCTL get_param -n \ + mdd.$MDT0.changelog_users | grep $clog") + echo "After crash, changelog user $clog2" + + [ -n "$clog2" -o $size2 -lt $size1 ] || + error "changelog record count unchanged" + + do_facet mds1 $LCTL set_param mdd.$MDT0.changelog_mask=\'$cl_mask\' -n +} +run_test 136 "changelog_deregister leaving pending records" + +test_137() { + df $DIR + mkdir -p $DIR/d1 + mkdir -p $DIR/d2 + dd if=/dev/zero of=$DIR/d1/$tfile bs=4096 count=1 + dd if=/dev/zero of=$DIR/d2/$tfile bs=4096 count=1 + cancel_lru_locks osc + + #define OBD_FAIL_PTLRPC_RESEND_RACE 0x525 + do_facet $SINGLEMDS "lctl set_param fail_loc=0x80000525" + + # RPC1: any reply is to be delayed to disable last_xid logic + ln $DIR/d1/$tfile $DIR/d1/f2 & + sleep 1 + + # RPC2: setattr1 reply is delayed & resent + # original reply comes to client; the resend get asleep + chmod 666 $DIR/d2/$tfile + + # RPC3: setattr2 on the same file; run ahead of RPC2 resend + chmod 777 $DIR/d2/$tfile + + # RPC2 resend wakes up + sleep 5 + [ $(stat -c "%a" $DIR/d2/$tfile) == 777 ] || error "resend got applied" +} +run_test 137 "late resend must be skipped if already applied" + +test_138() { + remote_mds_nodsh && skip "remote MDS with nodsh" + [ $MDSCOUNT -lt 2 ] && skip "needs >= 2 MDTs" && return 0 + [[ "$MDS1_VERSION" -ge $(version_code 2.12.59) ]] || + skip "Need server version newer than 2.12.59" + + zconf_umount_clients $CLIENTS $MOUNT + +#define OBD_FAIL_TGT_RECOVERY_CONNECT 0x724 + #delay a first step of recovey when MDS waiting clients + #and failing to get osp logs + do_facet $SINGLEMDS $LCTL set_param fail_loc=0x724 fail_val=5 + + facet_failover $SINGLEMDS + + #waiting failover and recovery timer + #the valuse is based on target_recovery_overseer() wait_event timeout + sleep 55 + stop $SINGLEMDS || error "stop MDS failed" + do_facet $SINGLEMDS $LCTL set_param fail_loc=0 + start $SINGLEMDS $(mdsdevname ${SINGLEMDS//mds/}) $MDS_MOUNT_OPTS || + error "start MDS failed" + zconf_mount_clients $CLIENTS $MOUNT +} +run_test 138 "Umount MDT during recovery" + +test_139() { + [ $MDSCOUNT -lt 2 ] && skip "needs >= 2 MDTs" && return 0 + [ $MDS1_VERSION -lt $(version_code 2.13.50) ] && + skip "Need MDS version at least 2.13.50" + + mdt_dev=$(mdsdevname 1) + + stop $SINGLEMDS || error "stop $SINGLEMDS failed" + +#define OBD_FAIL_OSP_INVALID_LOGID 0x2106 + do_facet $SINGLEMDS $LCTL set_param fail_val=0x68 fail_loc=0x80002106 + start $SINGLEMDS $mdt_dev $MDS_MOUNT_OPTS || error "Fail to start MDT" +} +run_test 139 "corrupted catid won't cause crash" + +test_140a() { + [ $MDS1_VERSION -lt $(version_code 2.12.58) ] && + skip "Need MDS version at least 2.13.50" + + [ "$SHARED_KEY" = true ] && + skip "server local client incompatible with SSK keys installed" + + slr=$(do_facet mds1 \ + $LCTL get_param -n mdt.$FSNAME-MDT0000.local_recovery) + stack_trap "do_facet mds1 $LCTL set_param \ + mdt.*.local_recovery=$slr" EXIT + + # disable recovery for local clients + # so local clients should be marked with no_recovery flag + do_facet mds1 $LCTL set_param mdt.*.local_recovery=0 + mount_mds_client + + local cnt + cnt=$(do_facet mds1 $LCTL get_param "mdt.*MDT0000.exports.*.export" | + grep export_flags.*no_recovery | wc -l) + echo "$cnt clients with recovery disabled" + umount_mds_client + [ $cnt -eq 0 ] && error "no clients with recovery disabled" + + # enable recovery for local clients + # so no local clients should be marked with no_recovery flag + do_facet mds1 $LCTL set_param mdt.*.local_recovery=1 + mount_mds_client + + cnt=$(do_facet mds1 $LCTL get_param "mdt.*MDT0000.exports.*.export" | + grep export_flags.*no_recovery | wc -l) + echo "$cnt clients with recovery disabled" + umount_mds_client + [ $cnt -eq 0 ] || error "$cnt clients with recovery disabled" +} +run_test 140a "local mount is flagged properly" + +test_140b() { + [ $MDS1_VERSION -lt $(version_code 2.12.58) ] && + skip "Need MDS version at least 2.13.50" + + [ "$SHARED_KEY" = true ] && + skip "server local client incompatible with SSK keys installed" + + slr=$(do_facet mds1 \ + $LCTL get_param -n mdt.$FSNAME-MDT0000.local_recovery) + stack_trap "do_facet mds1 $LCTL set_param \ + mdt.*.local_recovery=$slr" EXIT + + # disable recovery for local clients + do_facet mds1 $LCTL set_param mdt.*.local_recovery=0 + + mount_mds_client + replay_barrier mds1 + umount_mds_client + fail mds1 + # Lustre: tfs-MDT0000: Recovery over after 0:03, of 2 clients 2 rec... + local recovery=$(do_facet mds1 dmesg | + awk '/Recovery over after/ { print $6 }' | tail -1 | + awk -F: '{ print $1 * 60 + $2 }') + (( recovery < TIMEOUT * 2 + 5 )) || + error "recovery took too long $recovery > $((TIMEOUT * 2 + 5))" +} +run_test 140b "local mount is excluded from recovery" + +test_141() { + local oldc + local newc + + [ $PARALLEL == "yes" ] && skip "skip parallel run" + combined_mgs_mds || skip "needs combined MGS/MDT" + ( local_mode || from_build_tree ) && + skip "cannot run in local mode or from build tree" + + # some get_param have a bug to handle dot in param name + do_rpc_nodes $(facet_active_host $SINGLEMDS) cancel_lru_locks MGC + oldc=$(do_facet $SINGLEMDS $LCTL get_param -n \ + 'ldlm.namespaces.MGC*.lock_count') + fail $SINGLEMDS + do_rpc_nodes $(facet_active_host $SINGLEMDS) cancel_lru_locks MGC + newc=$(do_facet $SINGLEMDS $LCTL get_param -n \ + 'ldlm.namespaces.MGC*.lock_count') + + [ $oldc -eq $newc ] || error "mgc lost locks ($oldc != $newc)" + return 0 +} +run_test 141 "do not lose locks on MGS restart" + +test_142() { + [ $MDS1_VERSION -lt $(version_code 2.11.56) ] && + skip "Need MDS version at least 2.11.56" + + #define OBD_FAIL_MDS_ORPHAN_DELETE 0x165 + do_facet mds1 $LCTL set_param fail_loc=0x165 + $MULTIOP $DIR/$tfile Ouc || error "multiop failed" + + stop mds1 + start mds1 $(mdsdevname 1) $MDS_MOUNT_OPTS + + wait_update_facet mds1 "pgrep orph_.*-MDD | wc -l" "0" || + error "MDD orphan cleanup thread not quit" +} +run_test 142 "orphan name stub can be cleaned up in startup" + +test_143() { + [ $(lustre_version_code $SINGLEMDS) -lt $(version_code 2.13.00) ] && + skip "Need MDS version at least 2.13.00" + [ $PARALLEL == "yes" ] && skip "skip parallel run" + + local mntpt=$(facet_mntpt $SINGLEMDS) + stop mds1 + mount_fstype $SINGLEMDS || error "mount as fstype $SINGLEMDS failed" + do_facet $SINGLEMDS touch $mntpt/PENDING/$tfile + unmount_fstype $SINGLEMDS + start mds1 $(mdsdevname 1) $MDS_MOUNT_OPTS || error "mds1 start fail" + + wait_recovery_complete $SINGLEMDS || error "MDS recovery not done" + wait_update_facet mds1 "pgrep orph_.*-MDD | wc -l" "0" || + error "MDD orphan cleanup thread not quit" +} +run_test 143 "orphan cleanup thread shouldn't be blocked even delete failed" + +test_145() { + [ $MDSCOUNT -lt 3 ] && skip "needs >= 3 MDTs" + [ $(facet_active_host mds2) = $(facet_active_host mds3) ] && + skip "needs mds2 and mds3 on separate nodes" + + replay_barrier mds1 + + touch $DIR/$tfile + +#define OBD_FAIL_PTLRPC_DELAY_RECOV 0x507 + echo block mds_connect from mds2 + do_facet mds2 "$LCTL set_param fail_loc=0x507" + +#define OBD_FAIL_OUT_UPDATE_DROP 0x1707 + echo block recovery updates from mds3 + do_facet mds3 "$LCTL set_param fail_loc=0x1707" + + local hard_timeout=\ +$(do_facet mds1 $LCTL get_param -n mdt.$FSNAME-MDT0000.recovery_time_hard) + + fail mds1 & + + local get_soft_timeout_cmd=\ +"$LCTL get_param -n mdt.$FSNAME-MDT0000.recovery_time_soft 2>/dev/null" + + echo wait until mds1 recovery_time_soft is $hard_timeout + wait_update $(facet_host mds1) "$get_soft_timeout_cmd" \ +"$hard_timeout" $hard_timeout + + echo unblock mds_connect from mds2 + do_facet mds2 "$LCTL set_param fail_loc=0" + + echo upblock recovery updates from mds3 + do_facet mds3 "$LCTL set_param fail_loc=0" + + wait + [ -f $DIR/$tfile ] || error "$DIR/$tfile does not exist" +} +run_test 145 "connect mdtlovs and process update logs after recovery expire" + +test_147() { + local obd_timeout=200 + local old=$($LCTL get_param -n timeout) + local f=$DIR/$tfile + local connection_count + + $LFS setstripe -i 0 -c 1 $f + stripe_index=$($LFS getstripe -i $f) + if [ $stripe_index -ne 0 ]; then + $LFS getstripe $f + error "$f: stripe_index $stripe_index != 0" && return + fi + + $LCTL set_param timeout=$obd_timeout + stack_trap "$LCTL set_param timeout=$old && client_reconnect" EXIT + + # OBD_FAIL_OST_CONNECT_NET2 + # lost reply to connect request + do_facet ost1 lctl set_param fail_loc=0x00000225 timeout=$obd_timeout + stack_trap "do_facet ost1 $LCTL set_param fail_loc=0 timeout=$old" EXIT + + + ost_evict_client + # force reconnect + $LFS df $MOUNT > /dev/null 2>&1 & + sleep $((obd_timeout * 3 / 4)) + + $LCTL get_param osc.$FSNAME-OST0000-osc-*.state + connection_count=$($LCTL get_param osc.$FSNAME-OST0000-osc-*.state | + tac | sed "/FULL/,$ d" | grep CONNECTING | wc -l) + + echo $connection_count + (($connection_count >= 6)) || error "Client reconnected too slow" +} +run_test 147 "Check client reconnect" + +test_148() { + local wce_param="obdfilter.$FSNAME-OST0000.writethrough_cache_enable" + local p="$TMP/$TESTSUITE-$TESTNAME.parameters" + local amc=$(at_max_get client) + local amo=$(at_max_get ost1) + local timeout + + at_max_set 0 client + at_max_set 0 ost1 + timeout=$(request_timeout client) + + [ "$(facet_fstype ost1)" = "ldiskfs" ] && { + # save old r/o cache settings + save_lustre_params ost1 $wce_param > $p + + # disable r/o cache + do_facet ost1 "$LCTL set_param -n $wce_param=0" + } + + $LFS setstripe -i 0 -c 1 $DIR/$tfile + dd if=/dev/zero of=$DIR/$tfile bs=4096 count=1 oflag=direct + cp $DIR/$tfile $TMP/$tfile + #define OBD_FAIL_OST_BRW_PAUSE_BULK2 0x227 + do_facet ost1 $LCTL set_param fail_loc=0x80000227 + do_facet ost1 $LCTL set_param fail_val=$((timeout+2)) + dd if=/dev/urandom of=$DIR/$tfile bs=4096 count=1 conv=notrunc,fdatasync + dd if=/dev/zero of=$DIR/$tfile bs=4096 count=1 conv=notrunc,fdatasync + sleep 2 + cancel_lru_locks osc + cmp -b $DIR/$tfile $TMP/$tfile || error "wrong data" + + rm -f $DIR/$tfile $TMP/$tfile + + at_max_set $amc client + at_max_set $amo ost1 + + [ "$(facet_fstype ost1)" = "ldiskfs" ] && { + # restore initial r/o cache settings + restore_lustre_params < $p + } + + return 0 +} +run_test 148 "data corruption through resend" + +test_149() { + [ $MDSCOUNT -lt 2 ] && skip "needs >= 2 MDTs" + + test_mkdir -i 0 -c $MDSCOUNT $DIR/$tdir || error "mkdir $tdir failed" + + # make an orphan striped dir + $MULTIOP $DIR/$tdir D_c & + local PID=$! + sleep 0.3 + rmdir $DIR/$tdir || error "can't rmdir" + + # stop a slave MDT where one ons stripe is located + stop mds2 -f + + # stopping should not cause orphan as another MDT can + # be stopped yet + stop mds1 -f + + start mds1 $(mdsdevname 1) $MDS_MOUNT_OPTS || error "mds1 start fail" + start mds2 $(mdsdevname 2) $MDS_MOUNT_OPTS || error "mds1 start fail" + + kill -USR1 $PID + wait $PID + + clients_up + return 0 +} +run_test 149 "skip orphan removal at umount" + complete $SECONDS check_and_cleanup_lustre exit_status