X-Git-Url: https://git.whamcloud.com/?a=blobdiff_plain;f=lustre%2Ftests%2Fconf-sanity.sh;h=6702d6c7a8bd7d820c61756ee7832dff0faca64c;hb=faa958d6c5dda97f0c014d3ba36d40b324c2941a;hp=1c385382d292a557967270e4fbfc48fe23df7655;hpb=927effd2c2374508d904d2022c74574e508c07f1;p=fs%2Flustre-release.git diff --git a/lustre/tests/conf-sanity.sh b/lustre/tests/conf-sanity.sh old mode 100644 new mode 100755 index 1c38538..6702d6c --- a/lustre/tests/conf-sanity.sh +++ b/lustre/tests/conf-sanity.sh @@ -52,7 +52,7 @@ init_test_env $@ . ${CONFIG:=$LUSTRE/tests/cfg/$NAME.sh} # use small MDS + OST size to speed formatting time -# do not use too small MDSSIZE/OSTSIZE, which affect the default jouranl size +# do not use too small MDSSIZE/OSTSIZE, which affect the default journal size # STORED_MDSSIZE is used in test_18 STORED_MDSSIZE=$MDSSIZE STORED_OSTSIZE=$OSTSIZE @@ -86,8 +86,9 @@ init_logging # require_dsh_mds || exit 0 require_dsh_ost || exit 0 -# -[ "$SLOW" = "no" ] && EXCEPT_SLOW="30a 31 45 69" + +# 8 22 (min)" +[ "$SLOW" = "no" ] && EXCEPT_SLOW="45 69" assert_DIR @@ -1332,8 +1333,8 @@ run_test 31 "Connect to non-existent node (shouldn't crash)" T32_QID=60000 -T32_BLIMIT=20480 # Kbytes -T32_ILIMIT=2 +T32_BLIMIT=40960 # Kbytes +T32_ILIMIT=4 # # This is not really a test but a tool to create new disk @@ -1349,6 +1350,10 @@ test_32newtarball() { local dst=. local src=/etc/rc.d local tmp=$TMP/t32_image_create + local server_version=$(lustre_version_code $SINGLEMDS) + local remote_dir + local striped_dir + local pushd_dir if [ $FSNAME != t32fs -o \( -z "$MDSDEV" -a -z "$MDSDEV1" \) -o \ $OSTCOUNT -ne 1 -o -z "$OSTDEV1" ]; then @@ -1367,7 +1372,7 @@ test_32newtarball() { mkdir $tmp/src || return 1 tar cf - -C $src . | tar xf - -C $tmp/src dd if=/dev/zero of=$tmp/src/t32_qf_old bs=1M \ - count=$(($T32_BLIMIT / 1024 / 2)) + count=$(($T32_BLIMIT / 1024 / 4)) chown $T32_QID.$T32_QID $tmp/src/t32_qf_old # format ost with comma-separated NIDs to verify LU-4460 @@ -1376,18 +1381,42 @@ test_32newtarball() { setupall - [ $(lustre_version_code $SINGLEMDS) -lt $(version_code 2.3.50) ] && + [[ $server_version -ge $(version_code 2.3.50) ]] || $LFS quotacheck -ug /mnt/$FSNAME $LFS setquota -u $T32_QID -b 0 -B $T32_BLIMIT -i 0 -I $T32_ILIMIT \ /mnt/$FSNAME tar cf - -C $tmp/src . | tar xf - -C /mnt/$FSNAME + + if [[ $MDSCOUNT -ge 2 ]]; then + remote_dir=/mnt/$FSNAME/remote_dir + $LFS mkdir -i 1 $remote_dir + tar cf - -C $tmp/src . | tar xf - -C $remote_dir + + if [[ $server_version -ge $(version_code 2.7.0) ]]; then + striped_dir=/mnt/$FSNAME/striped_dir_old + $LFS mkdir -i 1 -c 2 $striped_dir + tar cf - -C $tmp/src . | tar xf - -C $striped_dir + fi + fi + stopall mkdir $tmp/img || return 1 setupall - pushd /mnt/$FSNAME + + pushd_dir=/mnt/$FSNAME + if [[ $MDSCOUNT -ge 2 ]]; then + pushd_dir=$remote_dir + if [[ $server_version -ge $(version_code 2.7.0) ]]; then + pushd $striped_dir + ls -Rni --time-style=+%s >$tmp/img/list2 + popd + fi + fi + + pushd $pushd_dir ls -Rni --time-style=+%s >$tmp/img/list find ! -name .lustre -type f -exec sha1sum {} \; | sort -k 2 >$tmp/img/sha1sums @@ -1395,7 +1424,7 @@ test_32newtarball() { $LCTL get_param -n version | head -n 1 | sed -e 's/^lustre: *//' >$tmp/img/commit - [ $(lustre_version_code $SINGLEMDS) -lt $(version_code 2.3.50) ] && + [[ $server_version -ge $(version_code 2.3.50) ]] || $LFS quotaon -ug /mnt/$FSNAME $LFS quota -u $T32_QID -v /mnt/$FSNAME $LFS quota -v -u $T32_QID /mnt/$FSNAME | @@ -1406,11 +1435,8 @@ test_32newtarball() { awk 'BEGIN { num='5' } { if ($1 == "'/mnt/$FSNAME'") \ { if (NF == 1) { getline } else { num++ } ; print $num;} }' \ | tr -d "*" > $tmp/img/ispace - - if [ $MDSCOUNT -ge 2 ]; then - $LFS mkdir -i 1 /mnt/$FSNAME/remote_dir - tar cf - -C $tmp/src . | tar xf - -C /mnt/$FSNAME/remote_dir - fi + echo $T32_BLIMIT > $tmp/img/blimit + echo $T32_ILIMIT > $tmp/img/ilimit stopall @@ -1428,6 +1454,9 @@ test_32newtarball() { mv ${MDSDEV1:-$MDSDEV} $tmp/img for num in $(seq 2 $MDSCOUNT); do local devname=$(mdsdevname $num) + local facet=mds$num + [[ $(facet_fstype $facet) != zfs ]] || + devname=$(mdsvdevname $num) mv $devname $tmp/img done mv $OSTDEV1 $tmp/img @@ -1473,7 +1502,8 @@ t32_check() { t32_test_cleanup() { local tmp=$TMP/t32 - local fstype=$(facet_fstype $SINGLEMDS) + local facet=$SINGLEMDS + local fstype=$(facet_fstype $facet) local rc=$? if $shall_cleanup_lustre; then @@ -1491,9 +1521,15 @@ t32_test_cleanup() { $r rm -rf $tmp rm -rf $tmp - if [ $fstype == "zfs" ]; then - $r $ZPOOL destroy t32fs-mdt1 || rc=$? - $r $ZPOOL destroy t32fs-ost1 || rc=$? + if [[ $fstype == zfs ]]; then + local poolname + local poolname_list="t32fs-mdt1 t32fs-ost1" + + ! $mdt2_is_available || poolname_list+=" t32fs-mdt2" + + for poolname in $poolname_list; do + destroy_zpool $facet $poolname + done fi return $rc } @@ -1583,8 +1619,8 @@ t32_verify_quota() { awk 'BEGIN { num='3' } { if ($1 == "'$mnt'") \ { if (NF == 1) { getline } else { num++ } ; print $num;} }' \ | tr -d "*") - [ $qval -eq $T32_BLIMIT ] || { - echo "blimit, act:$qval, exp:$T32_BLIMIT" + [ $qval -eq $img_blimit ] || { + echo "blimit, act:$qval, exp:$img_blimit" return 1 } @@ -1592,8 +1628,8 @@ t32_verify_quota() { awk 'BEGIN { num='7' } { if ($1 == "'$mnt'") \ { if (NF == 1) { getline } else { num++ } ; print $num;} }' \ | tr -d "*") - [ $qval -eq $T32_ILIMIT ] || { - echo "ilimit, act:$qval, exp:$T32_ILIMIT" + [ $qval -eq $img_ilimit ] || { + echo "ilimit, act:$qval, exp:$img_ilimit" return 1 } @@ -1615,18 +1651,18 @@ t32_verify_quota() { chmod 0777 $mnt runas -u $T32_QID -g $T32_QID dd if=/dev/zero of=$mnt/t32_qf_new \ - bs=1M count=$(($T32_BLIMIT / 1024)) oflag=sync && { + bs=1M count=$((img_blimit / 1024)) oflag=sync && { echo "Write succeed, but expect -EDQUOT" return 1 } rm -f $mnt/t32_qf_new runas -u $T32_QID -g $T32_QID createmany -m $mnt/t32_qf_ \ - $T32_ILIMIT && { + $img_ilimit && { echo "Create succeed, but expect -EDQUOT" return 1 } - unlinkmany $mnt/t32_qf_ $T32_ILIMIT + unlinkmany $mnt/t32_qf_ $img_ilimit return 0 } @@ -1640,6 +1676,7 @@ t32_test() { local shall_cleanup_mdt1=false local shall_cleanup_ost=false local shall_cleanup_lustre=false + local mdt2_is_available=false local node=$(facet_active_host $SINGLEMDS) local r="do_node $node" local node2=$(facet_active_host mds2) @@ -1649,6 +1686,8 @@ t32_test() { local img_arch local img_bspace local img_ispace + local img_blimit + local img_ilimit local fsname=t32fs local nid=$($r $LCTL list_nids | head -1) local mopts @@ -1661,6 +1700,7 @@ t32_test() { local mdt2_dev=$tmp/mdt2 local ost_dev=$tmp/ost local stripe_index + local stripe_count local dir trap 'trap - RETURN; t32_test_cleanup' RETURN @@ -1677,6 +1717,14 @@ t32_test() { img_arch=$($r cat $tmp/arch) img_bspace=$($r cat $tmp/bspace) img_ispace=$($r cat $tmp/ispace) + + # older images did not have "blimit" and "ilimit" files + # use old values for T32_BLIMIT and T32_ILIMIT + $r test -f $tmp/blimit && img_blimit=$($r cat $tmp/blimit) || + img_blimit=20480 + $r test -f $tmp/ilimit && img_ilimit=$($r cat $tmp/ilimit) || + img_ilimit=2 + echo "Upgrading from $(basename $tarball), created with:" echo " Commit: $img_commit" echo " Kernel: $img_kernel" @@ -1691,12 +1739,23 @@ t32_test() { $(lustre_version_code ost1) -lt $(version_code 2.5.0) ] && ff_convert="no" - if [ $fstype == "zfs" ]; then + ! $r test -f $mdt2_dev || mdt2_is_available=true + + if [[ $fstype == zfs ]]; then # import pool first - $r $ZPOOL import -f -d $tmp t32fs-mdt1 - $r $ZPOOL import -f -d $tmp t32fs-ost1 + local poolname + local poolname_list="t32fs-mdt1 t32fs-ost1" + + ! $mdt2_is_available || poolname_list+=" t32fs-mdt2" + + for poolname in $poolname_list; do + $r "$ZPOOL list -H $poolname >/dev/null 2>&1 || + $ZPOOL import -f -d $tmp $poolname" + done + mdt_dev=t32fs-mdt1/mdt1 ost_dev=t32fs-ost1/ost1 + ! $mdt2_is_available || mdt2_dev=t32fs-mdt2/mdt2 wait_update_facet $SINGLEMDS "$ZPOOL list | awk '/^t32fs-mdt1/ { print \\\$1 }'" "t32fs-mdt1" || { error_noexit "import zfs pool failed" @@ -1712,7 +1771,7 @@ t32_test() { return 1 } - if $r test -f $mdt2_dev; then + if $mdt2_is_available; then $r $TUNEFS --dryrun $mdt2_dev || { $r losetup -a error_noexit "tunefs.lustre before mounting the MDT" @@ -1729,7 +1788,7 @@ t32_test() { error_noexit "Enable mdt quota feature" return 1 } - if $r test -f $mdt2_dev; then + if $mdt2_is_available; then $r $TUNEFS --quota $mdt2_dev || { $r losetup -a error_noexit "Enable mdt quota feature" @@ -1771,7 +1830,7 @@ t32_test() { } shall_cleanup_mdt=true - if $r test -f $mdt2_dev; then + if $mdt2_is_available; then mopts=mgsnode=$nid,$mopts $r $MOUNT_CMD -o $mopts $mdt2_dev $tmp/mnt/mdt1 || { $r losetup -a @@ -1799,6 +1858,8 @@ t32_test() { return 1 } + [[ $(facet_fstype mds1) != zfs ]] || import_zpool fs2mds + $r $TUNEFS --dryrun $fs2mdsdev || { error_noexit "tunefs.lustre before mounting the MDT" return 1 @@ -1943,9 +2004,60 @@ t32_test() { return 1 } + if $r test -f $tmp/list; then + # + # There is not a Test Framework API to copy files to or + # from a remote node. + # + # LU-2393 - do both sorts on same node to ensure locale + # is identical + local list_file=$tmp/list + + if $mdt2_is_available; then + if [[ -d $tmp/mnt/lustre/striped_dir_old ]] && + $r test -f $tmp/list2; then + list_file=$tmp/list2 + pushd $tmp/mnt/lustre/striped_dir_old + else + pushd $tmp/mnt/lustre/remote_dir + fi + else + pushd $tmp/mnt/lustre + fi + $r cat $list_file | sort -k 6 >$tmp/list.orig + ls -Rni --time-style=+%s | sort -k 6 | + sed 's/\. / /' >$tmp/list || { + error_noexit "ls" + return 1 + } + popd + # + # 32-bit and 64-bit clients use different algorithms to + # convert FIDs into inode numbers. Hence, remove the + # inode numbers from the lists, if the original list was + # created on an architecture with different number of + # bits per "long". + # + if [ $(t32_bits_per_long $(uname -m)) != \ + $(t32_bits_per_long $img_arch) ]; then + echo "Different number of bits per \"long\"" \ + "from the disk image" + for list in list.orig list; do + sed -i -e 's/^[0-9]\+[ \t]\+//' \ + $tmp/$list + done + fi + if ! diff -ub $tmp/list.orig $tmp/list; then + error_noexit "list verification failed" + return 1 + fi + else + echo "list verification skipped" + fi + if [ "$dne_upgrade" != "no" ]; then $LFS mkdir -i 1 -c2 $tmp/mnt/lustre/striped_dir || { - error_noexit "set remote dir failed" + error_noexit "set striped dir failed" return 1 } @@ -1953,9 +2065,10 @@ t32_test() { pushd $tmp/mnt/lustre tar -cf - . --exclude=./striped_dir \ + --exclude=./striped_dir_old \ --exclude=./remote_dir | tar -xvf - -C striped_dir 1>/dev/null || { - error_noexit "cp to remote dir failed" + error_noexit "cp to striped dir failed" return 1 } popd @@ -1963,20 +2076,42 @@ t32_test() { # If it is upgrade from DNE (2.5), then rename the remote dir, # which is created in 2.5 to striped dir. - if $r test -f $mdt2_dev; then - stripe_index=$(LFS getdirstripe -i \ + if $mdt2_is_available && [[ "$dne_upgrade" != "no" ]]; then + stripe_index=$($LFS getdirstripe -i \ $tmp/mnt/lustre/remote_dir) - [ $stripe_index -eq 1 ] || { - error_noexit "get index $striped_index failed" + + [[ $stripe_index -eq 1 ]] || { + error_noexit "get index \"$stripe_index\"" \ + "from remote dir failed" return 1 } mv $tmp/mnt/lustre/remote_dir \ $tmp/mnt/lustre/striped_dir/ || { - error_noexit "mv failed" + error_noexit "mv remote dir failed" return 1 } fi + # If it is upgraded from DNE (2.7), then move the striped dir + # which was created in 2.7 to the new striped dir. + if $mdt2_is_available && [[ "$dne_upgrade" != "no" ]] && + [[ -d $tmp/mnt/lustre/striped_dir_old ]]; then + stripe_count=$($LFS getdirstripe -c \ + $tmp/mnt/lustre/striped_dir_old) + [[ $stripe_count -eq 2 ]] || { + error_noexit "get count $stripe_count" \ + "from striped dir failed" + return 1 + } + mv $tmp/mnt/lustre/striped_dir_old \ + $tmp/mnt/lustre/striped_dir/ || { + error_noexit "mv striped dir failed" + return 1 + } + fi + + sync; sleep 5; sync + $r $LCTL set_param -n osd*.*.force_sync=1 dd if=/dev/zero of=$tmp/mnt/lustre/tmp_file bs=10k count=10 || { error_noexit "dd failed" return 1 @@ -1996,8 +2131,8 @@ t32_test() { pushd $tmp/mnt/lustre fi - find ! -path "*remote_dir*" ! -name .lustre -type f \ - -exec sha1sum {} \; | + find ! -path "*remote_dir*" ! -path "*striped_dir*" \ + ! -name .lustre -type f -exec sha1sum {} \; | sort -k 2 >$tmp/sha1sums || { popd error_noexit "sha1sum" @@ -2009,22 +2144,34 @@ t32_test() { return 1 fi - # if upgrade from DNE (2.5), then check remote directory - if $r test -f $mdt2_dev; then - pushd $tmp/mnt/lustre/striped_dir/remote_dir - find ! -name .lustre -type f \ - -exec sha1sum {} \; | - sort -k 2 >$tmp/sha1sums || { + # if upgrade from DNE(2.5), then check remote directory + # if upgrade from DNE(2.7), then check striped directory + if $mdt2_is_available && + [[ "$dne_upgrade" != "no" ]]; then + local new_dir="$tmp/mnt/lustre/striped_dir" + local striped_dir_old="$new_dir/striped_dir_old" + + local dir_list="$new_dir/remote_dir" + [[ ! -d $triped_dir_old ]] || + dir_list+=" $striped_dir_old" + + for dir in $dir_list; do + pushd $dir + find ! -name .lustre -type f \ + -exec sha1sum {} \; | + sort -k 2 >$tmp/sha1sums || { + popd + error_noexit "sha1sum" + return 1 + } popd - error_noexit "sha1sum" - return 1 - } - popd - if ! diff -ub $tmp/sha1sums.orig \ - $tmp/sha1sums; then - error_noexit "sha1sum dne failed" - return 1 - fi + if ! diff -ub $tmp/sha1sums.orig \ + $tmp/sha1sums; then + error_noexit "sha1sum $dir" \ + "failed" + return 1 + fi + done fi else echo "sha1sum verification skipped" @@ -2037,42 +2184,6 @@ t32_test() { } fi - if $r test -f $tmp/list; then - # - # There is not a Test Framework API to copy files to or - # from a remote node. - # - # LU-2393 - do both sorts on same node to ensure locale - # is identical - $r cat $tmp/list | sort -k 6 >$tmp/list.orig - pushd $tmp/mnt/lustre - ls -Rni --time-style=+%s | sort -k 6 >$tmp/list || { - error_noexit "ls" - return 1 - } - popd - # - # 32-bit and 64-bit clients use different algorithms to - # convert FIDs into inode numbers. Hence, remove the inode - # numbers from the lists, if the original list was created - # on an architecture with different number of bits per - # "long". - # - if [ $(t32_bits_per_long $(uname -m)) != \ - $(t32_bits_per_long $img_arch) ]; then - echo "Different number of bits per \"long\" from the disk image" - for list in list.orig list; do - sed -i -e 's/^[0-9]\+[ \t]\+//' $tmp/$list - done - fi - if ! diff -ub $tmp/list.orig $tmp/list; then - error_noexit "list verification failed" - return 1 - fi - else - echo "list verification skipped" - fi - # migrate files/dirs to remote MDT, then move them back if [ $(lustre_version_code mds1) -ge $(version_code 2.7.50) -a \ $dne_upgrade != "no" ]; then @@ -2136,7 +2247,7 @@ t32_test() { } shall_cleanup_lustre=false else - if [ "$dne_upgrade" != "no" ]; then + if [[ "$dne_upgrade" != "no" ]] || $mdt2_is_available; then $r $UMOUNT $tmp/mnt/mdt1 || { error_noexit "Unmounting the MDT2" return 1 @@ -2161,6 +2272,12 @@ t32_test() { return 1 } + if [[ $fstype == zfs ]]; then + local poolname=t32fs-mdt1 + $r "$ZPOOL list -H $poolname >/dev/null 2>&1 || + $ZPOOL import -f -d $tmp $poolname" + fi + # mount a second time to make sure we didnt leave upgrade flag on $r $TUNEFS --dryrun $mdt_dev || { $r losetup -a @@ -2713,7 +2830,6 @@ test_41a() { #bug 14134 stop ost1 -f || error "unable to stop OST1" stop_mds || error "Unable to stop MDS" stop_mds || error "Unable to stop MDS on second try" - unload_modules_conf || error "unload_modules_conf failed" } run_test 41a "mount mds with --nosvc and --nomgs" @@ -2865,7 +2981,6 @@ test_42() { #bug 14693 do_facet mgs $LCTL conf_param $FSNAME.sys.some_wrong_param=20 cleanup || error "stopping $FSNAME failed with invalid sys param" - load_modules setup check_mount || error "client was not mounted with invalid sys param" cleanup || error "stopping $FSNAME failed with invalid sys param" @@ -2878,7 +2993,7 @@ test_43a() { [ $UID -ne 0 -o $RUNAS_ID -eq 0 ] && skip_env "run as root" ID1=${ID1:-501} - USER1=$(cat /etc/passwd | grep :$ID1:$ID1: | cut -d: -f1) + USER1=$(getent passwd | grep :$ID1:$ID1: | cut -d: -f1) [ -z "$USER1" ] && skip_env "missing user with uid=$ID1 gid=$ID1" && return @@ -3084,8 +3199,8 @@ test_45() { #17310 df -h $MOUNT & log "sleep 60 sec" sleep 60 - #define OBD_FAIL_PTLRPC_LONG_UNLINK 0x50f - do_facet client "$LCTL set_param fail_loc=0x50f" +#define OBD_FAIL_PTLRPC_LONG_REPL_UNLINK 0x50f + do_facet client "$LCTL set_param fail_loc=0x50f fail_val=0" log "sleep 10 sec" sleep 10 manual_umount_client --force || error "manual_umount_client failed" @@ -3545,7 +3660,6 @@ test_50i() { [ "$MDSCOUNT" -lt "2" ] && skip_env "$MDSCOUNT < 2, skipping" && return [ $(facet_fstype ost1) == zfs ] && import_zpool ost1 - load_modules do_facet mds2 "$TUNEFS --param mdc.active=0 $(mdsdevname 2)" || error "tunefs MDT2 failed" start_mds || error "Unable to start MDT" @@ -3698,9 +3812,9 @@ test_52() { echo # backup files - echo backup files to $TMP/files + echo backup files to $TMP/$tdir local files=$(find $DIR/$tdir -type f -newer $TMP/modified_first) - copy_files_xattrs $(hostname) $TMP/files $TMP/file_xattrs $files || + copy_files_xattrs $(hostname) $TMP/$tdir $TMP/file_xattrs $files || error "Unable to copy files" umount_client $MOUNT || error "Unable to umount client" @@ -3740,10 +3854,10 @@ test_52() { error "Some entry under /lost+found should be repaired" # compare files - diff_files_xattrs $(hostname) $TMP/files $TMP/file_xattrs $files || + diff_files_xattrs $(hostname) $TMP/$tdir $TMP/file_xattrs $files || error "Unable to diff files" - rm -rf $TMP/files $TMP/file_xattrs || + rm -rf $TMP/$tdir $TMP/file_xattrs || error "Unable to delete temporary files" do_node $ost1node "rm -rf $ost1tmp" || error "Unable to delete temporary files" @@ -3844,7 +3958,6 @@ thread_sanity() { local newvalue="${opts}=$(expr $basethr \* $ncpts)" setmodopts -a $modname "$newvalue" oldvalue - load_modules setup check_mount || return 41 @@ -3860,7 +3973,6 @@ thread_sanity() { return $? cleanup - load_modules setup } @@ -3957,7 +4069,7 @@ test_56() { MDSJOURNALSIZE=16 for num in $(seq 1 $MDSCOUNT); do - reformat_mdt $num + format_mdt $num done add ost1 $(mkfs_opts ost1 $(ostdevname 1)) --index=10000 --reformat \ $(ostdevname 1) $(ostvdevname 1) @@ -4642,6 +4754,59 @@ test_70d() { } run_test 70d "stop MDT1, mkdir succeed, create remote dir fail" +test_70e() { + [ $MDSCOUNT -lt 2 ] && skip "needs >= 2 MDTs" && return + + [ $(lustre_version_code $SINGLEMDS) -ge $(version_code 2.7.62) ] || + { skip "Need MDS version at least 2.7.62"; return 0; } + + cleanup || error "cleanup failed with $?" + + local mdsdev=$(mdsdevname 1) + local ostdev=$(ostdevname 1) + local mdsvdev=$(mdsvdevname 1) + local ostvdev=$(ostvdevname 1) + local opts_mds="$(mkfs_opts mds1 $mdsdev) --reformat $mdsdev $mdsvdev" + local opts_ost="$(mkfs_opts ost1 $ostdev) --reformat $ostdev $ostvdev" + + add mds1 $opts_mds || error "add mds1 failed" + start_mdt 1 || error "start mdt1 failed" + add ost1 $opts_ost || error "add ost1 failed" + start_ost || error "start ost failed" + mount_client $MOUNT > /dev/null || error "mount client $MOUNT failed" + + local soc=$(do_facet mds1 "$LCTL get_param -n \ + mdt.*MDT0000.sync_lock_cancel") + [ $soc == "never" ] || error "SoC enabled on single MDS" + + for i in $(seq 2 $MDSCOUNT); do + mdsdev=$(mdsdevname $i) + mdsvdev=$(mdsvdevname $i) + opts_mds="$(mkfs_opts mds$i $mdsdev) --reformat $mdsdev \ + $mdsvdev" + add mds$i $opts_mds || error "add mds$i failed" + start_mdt $i || error "start mdt$i fail" + done + + wait_dne_interconnect + + for i in $(seq $MDSCOUNT); do + soc=$(do_facet mds$i "$LCTL get_param -n \ + mdt.*MDT000$((i - 1)).sync_lock_cancel") + [ $soc == "blocking" ] || error "SoC not enabled on DNE" + done + + for i in $(seq 2 $MDSCOUNT); do + stop_mdt $i || error "stop mdt$i fail" + done + soc=$(do_facet mds1 "$LCTL get_param -n \ + mdt.*MDT0000.sync_lock_cancel") + [ $soc == "never" ] || error "SoC enabled on single MDS" + + cleanup || error "cleanup failed with $?" +} +run_test 70e "Sync-on-Cancel will be enabled by default on DNE" + test_71a() { [ $MDSCOUNT -lt 2 ] && skip "needs >= 2 MDTs" && return if combined_mgs_mds; then @@ -4809,8 +4974,8 @@ test_72() { #LU-2634 for num in $(seq $MDSCOUNT); do add mds${num} $(mkfs_opts mds$num $(mdsdevname $num)) \ - --reformat $(mdsdevname $num) $(mdsvdevname $num) || - error "add mds $num failed" + --reformat $(mdsdevname $num) $(mdsvdevname $num) || + error "add mds $num failed" do_facet mds${num} "$TUNE2FS -O extents $(mdsdevname $num)" || error "$TUNE2FS failed on mds${num}" done @@ -4841,7 +5006,6 @@ test_72() { #LU-2634 run_test 72 "test fast symlink with extents flag enabled" test_73() { #LU-3006 - load_modules [ $(facet_fstype ost1) == zfs ] && import_zpool ost1 do_facet ost1 "$TUNEFS --failnode=1.2.3.4@$NETTYPE $(ostdevname 1)" || error "1st tunefs failed" @@ -4953,6 +5117,28 @@ test_76b() { # LU-4783 } run_test 76b "verify params log setup correctly" +test_76c() { + [[ $(lustre_version_code mgs) -ge $(version_code 2.8.54) ]] || + { skip "Need MDS version at least 2.4.52" && return 0; } + setupall + local MASK_PARAM="mdd.*.changelog_mask" + echo "Change changelog_mask" + do_facet mgs $LCTL set_param -P $MASK_PARAM=-CLOSE || + error "Can't change changlog_mask" + wait_update $(facet_host mds) "$LCTL get_param -n $MASK_PARAM | + grep 'CLOSE'" "" + + echo "Check the value is stored after mds remount" + stop_mds || error "Failed to stop MDS" + start_mds || error "Failed to start MDS" + local CHANGELOG_MASK=$(do_facet mgs $LCTL get_param -n $MASK_PARAM) + echo $CHANGELOG_MASK | grep CLOSE > /dev/null && + error "changelog_mask is not changed" + + stopall +} +run_test 76c "verify changelog_mask is applied with set_param -P" + test_77() { # LU-3445 local server_version=$(lustre_version_code $SINGLEMDS) @@ -5000,8 +5186,12 @@ test_78() { skip "only applicable to ldiskfs-based MDTs and OSTs" && return # reformat the Lustre filesystem with a smaller size + local saved_MDSCOUNT=$MDSCOUNT local saved_MDSSIZE=$MDSSIZE + local saved_OSTCOUNT=$OSTCOUNT local saved_OSTSIZE=$OSTSIZE + MDSCOUNT=1 + OSTCOUNT=1 MDSSIZE=$((MDSSIZE - 20000)) OSTSIZE=$((OSTSIZE - 20000)) reformat || error "(1) reformat Lustre filesystem failed" @@ -5016,11 +5206,26 @@ test_78() { local i local file local num_files=100 + mkdir $MOUNT/$tdir || error "(3) mkdir $MOUNT/$tdir failed" + $LFS df; $LFS df -i for i in $(seq $num_files); do file=$MOUNT/$tdir/$tfile-$i - dd if=/dev/urandom of=$file count=1 bs=1M || + dd if=/dev/urandom of=$file count=1 bs=1M || { + $LCTL get_param osc.*.cur*grant* + $LFS df; $LFS df -i; + # stop creating files if there is no more space + if [ ! -e $file ]; then + num_files=$((i - 1)) + break + fi + + $LFS getstripe -v $file + local ost_idx=$(LFS getstripe -i $file) + do_facet ost$((ost_idx + 1)) \ + $LCTL get_param obdfilter.*.*grant* error "(4) create $file failed" + } done # unmount the Lustre filesystem @@ -5132,6 +5337,9 @@ test_78() { # unmount and reformat the Lustre filesystem cleanup || error "(12) cleanup Lustre filesystem failed" combined_mgs_mds || stop_mgs || error "(13) stop mgs failed" + + MDSCOUNT=$saved_MDSCOUNT + OSTCOUNT=$saved_OSTCOUNT reformat || error "(14) reformat Lustre filesystem failed" } run_test 78 "run resize2fs on MDT and OST filesystems" @@ -5319,8 +5527,9 @@ test_82a() { # LU-4665 local i local index local ost_indices + local LOV_V1_INSANE_STRIPE_COUNT=65532 for i in $(seq $OSTCOUNT); do - index=$((RANDOM * 2)) + index=$(((RANDOM * 2) % LOV_V1_INSANE_STRIPE_COUNT)) ost_indices+=" $index" done ost_indices=$(comma_list $ost_indices) @@ -5426,8 +5635,9 @@ test_82b() { # LU-4665 local i local index local ost_indices + local LOV_V1_INSANE_STRIPE_COUNT=65532 for i in $(seq $OSTCOUNT); do - index=$((RANDOM * 2)) + index=$(((RANDOM * 2) % LOV_V1_INSANE_STRIPE_COUNT)) ost_indices+=" $index" done ost_indices=$(comma_list $ost_indices) @@ -5555,32 +5765,6 @@ test_83() { run_test 83 "ENOSPACE on OST doesn't cause message VFS: \ Busy inodes after unmount ..." -recovery_time_min() { - local CONNECTION_SWITCH_MIN=5 - local CONNECTION_SWITCH_INC=5 - local CONNECTION_SWITCH_MAX - local RECONNECT_DELAY_MAX - local INITIAL_CONNECT_TIMEOUT - local max - local TO_20 - - #CONNECTION_SWITCH_MAX=min(50, max($CONNECTION_SWITCH_MIN,$TIMEOUT) - (($CONNECTION_SWITCH_MIN>$TIMEOUT)) && \ - max=$CONNECTION_SWITCH_MIN || max=$TIMEOUT - (($max<50)) && CONNECTION_SWITCH_MAX=$max || CONNECTION_SWITCH_MAX=50 - - #INITIAL_CONNECT_TIMEOUT = max(CONNECTION_SWITCH_MIN, \ - #obd_timeout/20) - TO_20=$(($TIMEOUT/20)) - (($CONNECTION_SWITCH_MIN>$TO_20)) && \ - INITIAL_CONNECT_TIMEOUT=$CONNECTION_SWITCH_MIN || \ - INITIAL_CONNECT_TIMEOUT=$TO_20 - - RECONNECT_DELAY_MAX=$(($CONNECTION_SWITCH_MAX+$CONNECTION_SWITCH_INC+ \ - $INITIAL_CONNECT_TIMEOUT)) - echo $((2*$RECONNECT_DELAY_MAX)) -} - test_84() { local facet=$SINGLEMDS local num=$(echo $facet | tr -d "mds") @@ -5591,7 +5775,6 @@ test_84() { local correct_clients local wrap_up=5 - load_modules echo "start mds service on $(facet_active_host $facet)" start_mds \ "-o recovery_time_hard=$time_min,recovery_time_soft=$time_min" $@ || @@ -5656,6 +5839,15 @@ test_85() { } run_test 85 "osd_ost init: fail ea_fid_set" +cleanup_86() { + trap 0 + + # ost1 has already registered to the MGS before the reformat. + # So after reformatting it with option "-G", it could not be + # mounted to the MGS. Cleanup the system for subsequent tests. + reformat_and_config +} + test_86() { [ "$(facet_fstype ost1)" = "zfs" ] && skip "LU-6442: no such mkfs params for ZFS OSTs" && return @@ -5679,6 +5871,9 @@ test_86() { echo "params: $opts" + trap cleanup_86 EXIT ERR + + stopall add ost1 $opts || error "add ost1 failed with new params" local FOUNDSIZE=$(do_facet ost1 "$DEBUGFS -c -R stats $(ostdevname 1)" | @@ -5686,7 +5881,8 @@ test_86() { [[ $FOUNDSIZE == $NEWSIZE ]] || error "Flex block group size: $FOUNDSIZE, expected: $NEWSIZE" - return 0 + + cleanup_86 } run_test 86 "Replacing mkfs.lustre -G option" @@ -6102,8 +6298,6 @@ test_91() { local nid local found - load_modules - [[ $(lustre_version_code ost1) -ge $(version_code 2.7.63) ]] || { skip "Need OST version at least 2.7.63" && return 0; } [[ $(lustre_version_code $SINGLEMDS) -ge $(version_code 2.7.63) ]] || @@ -6157,6 +6351,434 @@ test_91() { } run_test 91 "evict-by-nid support" +generate_ldev_conf() { + # generate an ldev.conf file + local ldevconfpath=$1 + local fstype= + local fsldevformat="" + touch $ldevconfpath + + fstype=$(facet_fstype mgs) + if [ "$fstype" == "zfs" ]; then + fsldevformat="$fstype:" + else + fsldevformat="" + fi + + printf "%s\t-\t%s-MGS0000\t%s%s\n" \ + $mgs_HOST \ + $FSNAME \ + $fsldevformat \ + $(mgsdevname) > $ldevconfpath + + local mdsfo_host=$mdsfailover_HOST; + if [ -z "$mdsfo_host" ]; then + mdsfo_host="-" + fi + + for num in $(seq $MDSCOUNT); do + fstype=$(facet_fstype mds$num) + if [ "$fstype" == "zfs" ]; then + fsldevformat="$fstype:" + else + fsldevformat="" + fi + + printf "%s\t%s\t%s-MDT%04d\t%s%s\n" \ + $mds_HOST \ + $mdsfo_host \ + $FSNAME \ + $num \ + $fsldevformat \ + $(mdsdevname $num) >> $ldevconfpath + done + + local ostfo_host=$ostfailover_HOST; + if [ -z "$ostfo_host" ]; then + ostfo_host="-" + fi + + for num in $(seq $OSTCOUNT); do + fstype=$(facet_fstype ost$num) + if [ "$fstype" == "zfs" ]; then + fsldevformat="$fstype:" + else + fsldevformat="" + fi + + printf "%s\t%s\t%s-OST%04d\t%s%s\n" \ + $ost_HOST \ + $ostfo_host \ + $FSNAME \ + $num \ + $fsldevformat \ + $(ostdevname $num) >> $ldevconfpath + done + + echo "----- $ldevconfpath -----" + cat $ldevconfpath + echo "--- END $ldevconfpath ---" + +} + +generate_nids() { + # generate a nids file (mapping between hostname to nid) + # looks like we only have the MGS nid available to us + # so just echo that to a file + local nidspath=$1 + echo -e "${mgs_HOST}\t${MGSNID}" > $nidspath + + echo "----- $nidspath -----" + cat $nidspath + echo "--- END $nidspath ---" +} + +compare_ldev_output() { + ldev_output=$1 + expected_output=$2 + + sort $expected_output -o $expected_output + sort $ldev_output -o $ldev_output + + echo "-- START OF LDEV OUTPUT --" + cat $ldev_output + echo "--- END OF LDEV OUTPUT ---" + + echo "-- START OF EXPECTED OUTPUT --" + cat $expected_output + echo "--- END OF EXPECTED OUTPUT ---" + + diff $expected_output $ldev_output + return $? +} + +test_92() { + if [ -z "$LDEV" ]; then + error "ldev is missing!" + fi + + local LDEVCONFPATH=$TMP/ldev.conf + local NIDSPATH=$TMP/nids + + echo "Host is $(hostname)" + + generate_ldev_conf $LDEVCONFPATH + generate_nids $NIDSPATH + + # echo the mgs nid and compare it to environment variable MGSNID + # also, ldev.conf and nids is a server side thing, use the OSS + # hostname + local output + output=$($LDEV -c $LDEVCONFPATH -H $ost_HOST -n $NIDSPATH echo %m) + + echo "-- START OF LDEV OUTPUT --" + echo -e "$output" + echo "--- END OF LDEV OUTPUT ---" + + # ldev failed, error + if [ $? -ne 0 ]; then + rm $LDEVCONFPATH $NIDSPATH + error "ldev failed to execute!" + fi + + # need to process multiple lines because of combined MGS and MDS + echo -e $output | awk '{ print $2 }' | while read -r line ; do + if [ "$line" != "$MGSNID" ]; then + rm $LDEVCONFPATH $NIDSPATH + error "ldev failed mgs nid '$line', expected '$MGSNID'" + fi + done + + rm $LDEVCONFPATH $NIDSPATH +} +run_test 92 "ldev returns MGS NID correctly in command substitution" + +test_93() { + [ $MDSCOUNT -lt 3 ] && skip "needs >= 3 MDTs" && return + + reformat + #start mgs or mgs/mdt0 + if ! combined_mgs_mds ; then + start_mgs + start_mdt 1 + else + start_mdt 1 + fi + + start_ost || error "OST0 start fail" + + #define OBD_FAIL_MGS_WRITE_TARGET_DELAY 0x90e + do_facet mgs "$LCTL set_param fail_val = 10 fail_loc=0x8000090e" + for num in $(seq 2 $MDSCOUNT); do + start_mdt $num & + done + + mount_client $MOUNT || error "mount client fails" + wait_osc_import_state mds ost FULL + wait_osc_import_state client ost FULL + check_mount || error "check_mount failed" + + cleanup || error "cleanup failed with $?" +} +run_test 93 "register mulitple MDT at the same time" + +test_94() { + if [ -z "$LDEV" ]; then + error "ldev is missing!" + fi + + local LDEVCONFPATH=$TMP/ldev.conf + local NIDSPATH=$TMP/nids + + generate_ldev_conf $LDEVCONFPATH + generate_nids $NIDSPATH + + local LDEV_OUTPUT=$TMP/ldev-output.txt + $LDEV -c $LDEVCONFPATH -n $NIDSPATH -F $FSNAME > $LDEV_OUTPUT + + # ldev failed, error + if [ $? -ne 0 ]; then + rm $LDEVCONFPATH $NIDSPATH $LDEV_OUTPUT + error "ldev failed to execute!" + fi + + # expected output + local EXPECTED_OUTPUT=$TMP/ldev-expected.txt + + printf "%s-MGS0000\n" $FSNAME > $EXPECTED_OUTPUT + + for num in $(seq $MDSCOUNT); do + printf "%s-MDT%04d\n" $FSNAME $num >> $EXPECTED_OUTPUT + done + + for num in $(seq $OSTCOUNT); do + printf "%s-OST%04d\n" $FSNAME $num >> $EXPECTED_OUTPUT + done + + compare_ldev_output $LDEV_OUTPUT $EXPECTED_OUTPUT + + if [ $? -ne 0 ]; then + rm $LDEVCONFPATH $NIDSPATH $EXPECTED_OUTPUT $LDEV_OUTPUT + error "ldev failed to produce the correct hostlist!" + fi + + rm $LDEVCONFPATH $NIDSPATH $EXPECTED_OUTPUT $LDEV_OUTPUT +} +run_test 94 "ldev outputs correct labels for file system name query" + +test_95() { + if [ -z "$LDEV" ]; then + error "ldev is missing!" + fi + + local LDEVCONFPATH=$TMP/ldev.conf + local NIDSPATH=$TMP/nids + + generate_ldev_conf $LDEVCONFPATH + generate_nids $NIDSPATH + + # SUCCESS CASES + # file sys filter + $LDEV -c $LDEVCONFPATH -n $NIDSPATH -F $FSNAME &>/dev/null + if [ $? -ne 0 ]; then + rm $LDEVCONFPATH $NIDSPATH + error "ldev label filtering w/ -F failed!" + fi + + # local filter + $LDEV -c $LDEVCONFPATH -n $NIDSPATH -l &>/dev/null + if [ $? -ne 0 ]; then + rm $LDEVCONFPATH $NIDSPATH + error "ldev label filtering w/ -l failed!" + fi + + # foreign filter + $LDEV -c $LDEVCONFPATH -n $NIDSPATH -f &>/dev/null + if [ $? -ne 0 ]; then + rm $LDEVCONFPATH $NIDSPATH + error "ldev label filtering w/ -f failed!" + fi + + # all filter + $LDEV -c $LDEVCONFPATH -n $NIDSPATH -a &>/dev/null + if [ $? -ne 0 ]; then + rm $LDEVCONFPATH $NIDSPATH + error "ldev label filtering w/ -a failed!" + fi + + # FAILURE CASES + # all & file sys + $LDEV -c $LDEVCONFPATH -n $NIDSPATH -a -F $FSNAME &>/dev/null + if [ $? -eq 0 ]; then + rm $LDEVCONFPATH $NIDSPATH + error "ldev label filtering w/ -a and -F incorrectly succeeded" + fi + + # all & foreign + $LDEV -c $LDEVCONFPATH -n $NIDSPATH -a -f &>/dev/null + if [ $? -eq 0 ]; then + rm $LDEVCONFPATH $NIDSPATH + error "ldev label filtering w/ -a and -f incorrectly succeeded" + fi + + # all & local + $LDEV -c $LDEVCONFPATH -n $NIDSPATH -a -l &>/dev/null + if [ $? -eq 0 ]; then + rm $LDEVCONFPATH $NIDSPATH + error "ldev label filtering w/ -a and -l incorrectly succeeded" + fi + + # foreign & local + $LDEV -c $LDEVCONFPATH -n $NIDSPATH -f -l &>/dev/null + if [ $? -eq 0 ]; then + rm $LDEVCONFPATH $NIDSPATH + error "ldev label filtering w/ -f and -l incorrectly succeeded" + fi + + # file sys & local + $LDEV -c $LDEVCONFPATH -n $NIDSPATH -F $FSNAME -l &>/dev/null + if [ $? -eq 0 ]; then + rm $LDEVCONFPATH $NIDSPATH + error "ldev label filtering w/ -F and -l incorrectly succeeded" + fi + + # file sys & foreign + $LDEV -c $LDEVCONFPATH -n $NIDSPATH -F $FSNAME -f &>/dev/null + if [ $? -eq 0 ]; then + rm $LDEVCONFPATH $NIDSPATH + error "ldev label filtering w/ -F and -f incorrectly succeeded" + fi + + rm $LDEVCONFPATH $NIDSPATH +} +run_test 95 "ldev should only allow one label filter" + +test_96() { + if [ -z "$LDEV" ]; then + error "ldev is missing!" + fi + + local LDEVCONFPATH=$TMP/ldev.conf + local NIDSPATH=$TMP/nids + + generate_ldev_conf $LDEVCONFPATH + generate_nids $NIDSPATH + + local LDEV_OUTPUT=$TMP/ldev-output.txt + $LDEV -c $LDEVCONFPATH -n $NIDSPATH -H $mgs_HOST \ + echo %H-%b | \ + awk '{print $2}' > $LDEV_OUTPUT + + # ldev failed, error + if [ $? -ne 0 ]; then + rm $LDEVCONFPATH $NIDSPATH $LDEV_OUTPUT + error "ldev failed to execute!" + fi + + # expected output + local EXPECTED_OUTPUT=$TMP/ldev-expected-output.txt + + echo "$mgs_HOST-$(facet_fstype mgs)" > $EXPECTED_OUTPUT + + if [ "$mgs_HOST" == "$mds_HOST" ]; then + for num in $(seq $MDSCOUNT); do + echo "$mds_HOST-$(facet_fstype mds$num)" \ + >> $EXPECTED_OUTPUT + done + fi + + if [ "$mgs_HOST" == "$ost_HOST" ]; then + for num in $(seq $OSTCOUNT); do + echo "$ost_HOST-$(facet_fstype ost$num)" \ + >> $EXPECTED_OUTPUT + done + fi + + compare_ldev_output $LDEV_OUTPUT $EXPECTED_OUTPUT + + if [ $? -ne 0 ]; then + rm $LDEVCONFPATH $NIDSPATH $EXPECTED_OUTPUT $LDEV_OUTPUT + error "ldev failed to produce the correct output!" + fi + + rm $LDEVCONFPATH $NIDSPATH $EXPECTED_OUTPUT $LDEV_OUTPUT +} +run_test 96 "ldev returns hostname and backend fs correctly in command sub" + +test_97() { + if [ -z "$LDEV" ]; then + error "ldev is missing!" + fi + + local LDEVCONFPATH=$TMP/ldev.conf + local NIDSPATH=$TMP/nids + + generate_ldev_conf $LDEVCONFPATH + generate_nids $NIDSPATH + + local LDEV_OUTPUT=$TMP/ldev-output.txt + local EXPECTED_OUTPUT=$TMP/ldev-expected-output.txt + + echo -e "\nMDT role" + $LDEV -c $LDEVCONFPATH -n $NIDSPATH -F $FSNAME -R mdt > $LDEV_OUTPUT + + if [ $? -ne 0 ]; then + rm $LDEVCONFPATH $NIDSPATH $LDEV_OUTPUT + error "ldev failed to execute for mdt role!" + fi + + for num in $(seq $MDSCOUNT); do + printf "%s-MDT%04d\n" $FSNAME $num >> $EXPECTED_OUTPUT + done + + compare_ldev_output $LDEV_OUTPUT $EXPECTED_OUTPUT + + if [ $? -ne 0 ]; then + rm $LDEVCONFPATH $NIDSPATH $EXPECTED_OUTPUT $LDEV_OUTPUT + error "ldev failed to produce the correct output for mdt role!" + fi + + echo -e "\nOST role" + $LDEV -c $LDEVCONFPATH -n $NIDSPATH -F $FSNAME -R ost > $LDEV_OUTPUT + + if [ $? -ne 0 ]; then + rm $LDEVCONFPATH $NIDSPATH $LDEV_OUTPUT $EXPECTED_OUTPUT + error "ldev failed to execute for ost role!" + fi + + rm $EXPECTED_OUTPUT + for num in $(seq $OSTCOUNT); do + printf "%s-OST%04d\n" $FSNAME $num >> $EXPECTED_OUTPUT + done + + compare_ldev_output $LDEV_OUTPUT $EXPECTED_OUTPUT + + if [ $? -ne 0 ]; then + rm $LDEVCONFPATH $NIDSPATH $EXPECTED_OUTPUT $LDEV_OUTPUT + error "ldev failed to produce the correct output for ost role!" + fi + + echo -e "\nMGS role" + $LDEV -c $LDEVCONFPATH -n $NIDSPATH -F $FSNAME -R mgs > $LDEV_OUTPUT + + if [ $? -ne 0 ]; then + rm $LDEVCONFPATH $NIDSPATH $LDEV_OUTPUT $EXPECTED_OUTPUT + error "ldev failed to execute for mgs role!" + fi + + printf "%s-MGS0000\n" $FSNAME > $EXPECTED_OUTPUT + + compare_ldev_output $LDEV_OUTPUT $EXPECTED_OUTPUT + + if [ $? -ne 0 ]; then + rm $LDEVCONFPATH $NIDSPATH $EXPECTED_OUTPUT $LDEV_OUTPUT + error "ldev failed to produce the correct output for mgs role!" + fi + + rm $LDEVCONFPATH $NIDSPATH $EXPECTED_OUTPUT $LDEV_OUTPUT +} +run_test 97 "ldev returns correct ouput when querying based on role" + if ! combined_mgs_mds ; then stop mgs fi