X-Git-Url: https://git.whamcloud.com/?p=fs%2Flustre-release.git;a=blobdiff_plain;f=lustre%2Ftests%2Fconf-sanity.sh;h=f540da286be2e82eca9f8d6d44b3671da0277248;hp=27fc2b1849b92918776ce2d71f9b410667118ccf;hb=f38f09e02a05c82718344ad86f80a4a0f399af9d;hpb=38c8fdfde3953f239bd3d86a91a3213737231ce5 diff --git a/lustre/tests/conf-sanity.sh b/lustre/tests/conf-sanity.sh index 27fc2b1..f540da2 100644 --- a/lustre/tests/conf-sanity.sh +++ b/lustre/tests/conf-sanity.sh @@ -15,14 +15,12 @@ init_logging ALWAYS_EXCEPT="$CONF_SANITY_EXCEPT 32newtarball" # bug number for skipped test: LU-11915 -ALWAYS_EXCEPT="$ALWAYS_EXCEPT 110" +ALWAYS_EXCEPT="$ALWAYS_EXCEPT 110 115" # UPDATE THE COMMENT ABOVE WITH BUG NUMBERS WHEN CHANGING ALWAYS_EXCEPT! if $SHARED_KEY; then # bug number for skipped tests: LU-9795 (all below) - ALWAYS_EXCEPT="$ALWAYS_EXCEPT 0 31 32a 32d 35a" - ALWAYS_EXCEPT="$ALWAYS_EXCEPT 53a 53b 54b 76a 76b" - ALWAYS_EXCEPT="$ALWAYS_EXCEPT 76c 76d 78 103" + ALWAYS_EXCEPT="$ALWAYS_EXCEPT 84 86 103" fi if ! combined_mgs_mds; then @@ -71,7 +69,7 @@ gen_config() { # and stop to generate the startup logs. start_mds start_ost - wait_osc_import_state mds ost FULL + wait_osc_import_state mds ost FULL stop_ost stop_mds } @@ -95,7 +93,7 @@ writeconf_or_reformat() { } reformat() { - formatall + formatall } start_mgs () { @@ -113,6 +111,16 @@ start_mdt() { start $facet ${dev} $MDS_MOUNT_OPTS $@ || return 94 } +stop_mdt_no_force() { + local num=$1 + local facet=mds$num + local dev=$(mdsdevname $num) + shift 1 + + echo "stop mds service on `facet_active_host $facet`" + stop $facet || return 97 +} + stop_mdt() { local num=$1 local facet=mds$num @@ -130,6 +138,9 @@ start_mds() { for num in $(seq $MDSCOUNT); do start_mdt $num $@ || return 94 done + for num in $(seq $MDSCOUNT); do + wait_clients_import_state ${CLIENTS:-$HOSTNAME} mds${num} FULL + done } start_mgsmds() { @@ -147,14 +158,15 @@ stop_mds() { } stop_mgs() { - echo "stop mgs service on `facet_active_host mgs`" - # These tests all use non-failover stop - stop mgs -f || return 97 + echo "stop mgs service on `facet_active_host mgs`" + # These tests all use non-failover stop + stop mgs -f || return 97 } start_ost() { echo "start ost1 service on `facet_active_host ost1`" start ost1 $(ostdevname 1) $OST_MOUNT_OPTS $@ || return 95 + wait_clients_import_state ${CLIENTS:-$HOSTNAME} ost1 FULL } stop_ost() { @@ -166,6 +178,7 @@ stop_ost() { start_ost2() { echo "start ost2 service on `facet_active_host ost2`" start ost2 $(ostdevname 2) $OST_MOUNT_OPTS $@ || return 92 + wait_clients_import_state ${CLIENTS:-$HOSTNAME} ost2 FULL } stop_ost2() { @@ -175,16 +188,11 @@ stop_ost2() { } mount_client() { - local MOUNTPATH=$1 - echo "mount $FSNAME on ${MOUNTPATH}....." - zconf_mount $(hostname) $MOUNTPATH || return 96 -} + local mountpath=$1 + local mountopt="$2" -remount_client() { - local mountopt="remount,$1" - local MOUNTPATH=$2 - echo "remount '$1' lustre on ${MOUNTPATH}....." - zconf_mount $(hostname) $MOUNTPATH "$mountopt" || return 96 + echo "mount $FSNAME ${mountopt:+with opts $mountopt} on $mountpath....." + zconf_mount $HOSTNAME $mountpath $mountopt || return 96 } umount_client() { @@ -249,8 +257,8 @@ check_mount() { do_facet client "cp /etc/passwd $DIR/a" || return 71 do_facet client "rm $DIR/a" || return 72 # make sure lustre is actually mounted (touch will block, - # but grep won't, so do it after) - do_facet client "grep $MOUNT' ' /proc/mounts > /dev/null" || return 73 + # but grep won't, so do it after) + do_facet client "grep $MOUNT' ' /proc/mounts > /dev/null" || return 73 echo "setup single mount lustre success" } @@ -706,10 +714,10 @@ test_20() { mount_client $MOUNT || error "mount_client $MOUNT failed" check_mount || error "check_mount failed" rm -f $DIR/$tfile || error "remove $DIR/$tfile failed." - remount_client ro $MOUNT || error "remount_client with ro failed" + mount_client $MOUNT remount,ro || error "remount client with ro failed" touch $DIR/$tfile && error "$DIR/$tfile created incorrectly" [ -e $DIR/$tfile ] && error "$DIR/$tfile exists incorrectly" - remount_client rw $MOUNT || error "remount_client with rw failed" + mount_client $MOUNT remount,rw || error "remount client with rw failed" touch $DIR/$tfile || error "touch $DIR/$tfile failed" MCNT=$(grep -c $MOUNT' ' /etc/mtab) [ "$MCNT" -ne 1 ] && error "$MOUNT in /etc/mtab $MCNT times" @@ -762,12 +770,18 @@ test_21d() { start_mds || error "MDS start failed" wait_osc_import_state mds ost2 FULL + local zkeeper=${KEEP_ZPOOL} + stack_trap "KEEP_ZPOOL=$zkeeper" EXIT + KEEP_ZPOOL="true" + stop_ost || error "Unable to stop OST1" stop_ost2 || error "Unable to stop OST2" stop_mds || error "Unable to stop MDS" stop_mgs #writeconf to remove all ost2 traces for subsequent tests writeconf_or_reformat + KEEP_ZPOOL="${zkeeper}" + start_mgs || error "unable to start MGS" } run_test 21d "start mgs then ost and then mds" @@ -1426,6 +1440,16 @@ test_32newtarball() { echo $T32_BLIMIT > $tmp/img/blimit echo $T32_ILIMIT > $tmp/img/ilimit + $MULTIOP /mnt/$FSNAME/orph_file Ouw_c& + pid=$! + sync + stop_mdt_no_force 1 + debugfs -R "ls /PENDING" ${MDSDEV1:-$MDSDEV} + cp ${MDSDEV1:-$MDSDEV} $tmp/img + start_mdt 1 + kill -s USR1 $pid + wait $pid + stopall pushd $tmp/src @@ -1439,7 +1463,6 @@ test_32newtarball() { uname -r >$tmp/img/kernel uname -m >$tmp/img/arch - mv ${MDSDEV1:-$MDSDEV} $tmp/img for num in $(seq 2 $MDSCOUNT); do local devname=$(mdsdevname $num) local facet=mds$num @@ -1661,6 +1684,7 @@ t32_test() { local writeconf=$2 local dne_upgrade=${dne_upgrade:-"no"} local dom_upgrade=${dom_upgrade:-"no"} + local project_upgrade=${project_upgrade:-"no"} local ff_convert=${ff_convert:-"no"} local shall_cleanup_mdt=false local shall_cleanup_mdt1=false @@ -1757,6 +1781,16 @@ t32_test() { error_noexit "import zfs pool failed" return 1 } + elif [ "$project_upgrade" != "no" ]; then + ! $r tune2fs -O project $mdt_dev && + error_noexit "enable project on mdt0 failed" && + return 1 + $mdt2_is_available && ! $r tune2fs "-O project" $mdt2_dev && + error_noexit "enable project on mdt failed" && + return 1 + ! $r tune2fs -O project $ost_dev && + error_noexit "enable project on mdt failed" && + return 1 fi $r $LCTL set_param debug="$PTLDEBUG" @@ -2136,9 +2170,9 @@ t32_test() { $LFS setdirstripe -D -c2 $tmp/mnt/lustre/striped_dir pushd $tmp/mnt/lustre - tar -cf - . --exclude=./striped_dir \ - --exclude=./striped_dir_old \ - --exclude=./remote_dir | + tar -c --exclude=./striped_dir \ + --exclude=./striped_dir_old \ + --exclude=./remote_dir -f - .| tar -xvf - -C striped_dir 1>/dev/null || { error_noexit "cp to striped dir failed" return 1 @@ -2317,6 +2351,26 @@ t32_test() { } shall_cleanup_lustre=false else + $MOUNT_CMD $nid:/$fsname $tmp/mnt/lustre || { + error_noexit "Mounting the client" + return 1 + } + + if [ "$mds1_FSTYPE" == ldiskfs -a \ + "$project_upgrade" != "no" ]; then + ! $LFS project -d -p 1 $tmp/mnt/lustre/* && + error_noexit "set project failed" && + return 1 + fi + + [[ $(do_facet mds1 pgrep orph_.*-MDD | wc -l) == 0 ]] || + error "MDD orphan cleanup thread not quit" + + umount $tmp/mnt/lustre || { + error_noexit "Unmounting the client" + return 1 + } + if [[ "$dne_upgrade" != "no" ]] || $mdt2_is_available; then $r $UMOUNT $tmp/mnt/mdt1 || { error_noexit "Unmounting the MDT2" @@ -2388,6 +2442,7 @@ test_32a() { t32_check for tarball in $tarballs; do + banner "testing $tarball upgrade" t32_test $tarball || let "rc += $?" done return $rc @@ -2401,6 +2456,7 @@ test_32b() { t32_check for tarball in $tarballs; do + banner "testing $tarball upgrade with writeconf" t32_test $tarball writeconf || let "rc += $?" done return $rc @@ -2416,8 +2472,9 @@ test_32c() { t32_check for tarball in $tarballs; do # Do not support 1_8 and 2_1 direct upgrade to DNE2 anymore */ - echo $tarball | grep "1_8" && continue - echo $tarball | grep "2_1" && continue + [[ "$tarball" =~ "1_8" ]] && echo "skip $tarball" && continue + [[ "$tarball" =~ "2_1" ]] && echo "skip $tarball" && continue + banner "testing $tarball upgrade with DNE" load_modules dne_upgrade=yes t32_test $tarball writeconf || rc=$? done @@ -2432,11 +2489,17 @@ test_32d() { t32_check for tarball in $tarballs; do - ff_convert=yes t32_test $tarball || rc=$? + banner "testing $tarball upgrade with ff convert and project upgrade" + project_upgrade="no" + [[ "$MDS1_VERSION" -ge $(version_code 2.13.54) ]] && + [[ "$tarball" =~ "disk2_4-ldiskfs" ]] && + project_upgrade="yes" + project_upgrade=$project_upgrade ff_convert=yes t32_test \ + $tarball || rc=$? done return $rc } -run_test 32d "convert ff test" +run_test 32d "convert ff and project quota upgrade test" test_32e() { [[ "$MDS1_VERSION" -ge $(version_code 2.10.56) ]] || @@ -2448,8 +2511,9 @@ test_32e() { t32_check for tarball in $tarballs; do - echo $tarball | grep "2_9" || continue + [[ "$tarball" =~ "2_9" ]] || continue #load_modules + banner "testing $tarball upgrade with DoM" dom_upgrade=yes t32_test $tarball writeconf || let "rc += $?" done return $rc @@ -3511,16 +3575,13 @@ test_48() { # bz-17636 LU-7473 # LOV EA, and so on. These EA will use some EA space that is shared by # ACL entries. So here we only check some reasonable ACL entries count, # instead of the max number that is calculated from the max_ea_size. - if [ "$MDS1_VERSION" -lt $(version_code 2.8.57) ]; - then + if [ "$MDS1_VERSION" -lt $(version_code 2.8.57) ]; then count=28 # hard coded of RPC protocol - elif [ "$mds1_FSTYPE" != ldiskfs ]; then - count=4000 # max_num 4091 max_ea_size = ~65536 - elif ! large_xattr_enabled; then - count=450 # max_num 497 max_ea_size = 4012 - else + elif large_xattr_enabled; then count=4500 # max_num 8187 max_ea_size = 65452 - # not create too much (>5000) to save test time + # not create too many (4500) to save test time + else + count=450 # max_num 497 max_ea_size = 4012 fi echo "It is expected to hold at least $count ACL entries" @@ -4645,6 +4706,10 @@ test_63() { skip "ldiskfs module has not been loaded" fi + if grep -q "CONFIG_DEBUG_LOCK_ALLOC=y" /boot/config-$(uname -r); then + skip "test is not compatible with CONFIG_DEBUG_LOCK_ALLOC=y" + fi + echo "$inode_slab ldiskfs inodes per page" [ "${inode_slab%.*}" -ge "3" ] && return 0 @@ -5864,22 +5929,15 @@ test_81() { # LU-4665 # Check max_easize. local max_easize=$($LCTL get_param -n llite.*.max_easize) - if [ $MDS1_VERSION -lt $(version_code 2.12.51) ] - then - [[ $max_easize -eq 128 ]] || - error "max_easize is $max_easize, should be 128 bytes" + # 65452 is XATTR_SIZE_MAX less ldiskfs ea overhead + if large_xattr_enabled; then + [[ $max_easize -ge 65452 ]] || + error "max_easize is $max_easize, should be at least 65452 bytes" else # LU-11868 - # 4012 is 4096 - ldiskfs ea overhead + # 4012 is 4096 less ldiskfs ea overhead [[ $max_easize -ge 4012 ]] || - error "max_easize is $max_easize, should be at least 4012 bytes" - - # 65452 is XATTR_SIZE_MAX - ldiskfs ea overhead - if large_xattr_enabled; - then - [[ $max_easize -ge 65452 ]] || - error "max_easize is $max_easize, should be at least 65452 bytes" - fi + error "max_easize is $max_easize, should be at least 4012 bytes" fi restore_ostindex @@ -6132,7 +6190,7 @@ test_83() { error "format ost1 error" if ! test -b $dev; then - mnt_opts=$(csa_add "$OST_MOUNT_OPTS" -o loop) + mnt_opts=$(csa_add "$OST_MOUNT_FS_OPTS" -o loop) fi echo "mnt_opts $mnt_opts" do_facet ost1 mount -t "$ost1_FSTYPE" $dev \ @@ -6251,23 +6309,15 @@ test_86() { [[ "$MDS1_VERSION" -ge $(version_code 2.7.56) ]] || skip "Need server version newer than 2.7.55" - local OST_OPTS="$(mkfs_opts ost1 $(ostdevname 1)) \ - --reformat $(ostdevname 1) $(ostvdevname 1)" - local NEWSIZE=1024 local OLDSIZE=$(do_facet ost1 "$DEBUGFS -c -R stats $(ostdevname 1)" | awk '/Flex block group size: / { print $NF; exit; }') - local opts=OST_OPTS - if [[ ${!opts} != *mkfsoptions* ]]; then - eval opts=\"${!opts} \ - --mkfsoptions='\\\"-O flex_bg -G $NEWSIZE\\\"'\" - else - val=${!opts//--mkfsoptions=\\\"/ \ - --mkfsoptions=\\\"-O flex_bg -G $NEWSIZE } - eval opts='${val}' - fi + [ "$OLDSIZE" == "$NEWSIZE" ] && skip "$NEWSIZE groups already" + local opts=" -O flex_bg -G $NEWSIZE" + opts=$(OST_FS_MKFS_OPTS+="$opts" mkfs_opts ost1 $(ostdevname 1)) + opts+=" --reformat $(ostdevname 1) $(ostvdevname 1)" echo "params: $opts" trap cleanup_86 EXIT ERR @@ -6292,6 +6342,7 @@ test_87() { #LU-6544 skip "ldiskfs only test" [[ $OSTCOUNT -gt 59 ]] && skip "Ignore wide striping situation" + [ -n "$FILESET" ] && skip "Not functional for FILESET set" local mdsdev=$(mdsdevname 1) local mdsvdev=$(mdsvdevname 1) @@ -6370,7 +6421,7 @@ test_87() { #LU-6544 more than $left_size-byte space left in inode." echo "Verified: at most $left_size-byte space left in inode." - umount_ldiskfs $SINGLEMDS + unmount_ldiskfs $SINGLEMDS for i in $(seq $OSTCOUNT); do stop ost$i -f || error "stop ost$i failed" @@ -6631,10 +6682,41 @@ test_90b() { } run_test 90b "check max_mod_rpcs_in_flight is enforced after update" +save_params_90c() { + # get max_rpcs_in_flight value + mrif_90c=$($LCTL get_param -n \ + mdc.$FSNAME-MDT0000-mdc-*.max_rpcs_in_flight) + echo "max_rpcs_in_flight is $mrif_90c" + + # get max_mod_rpcs_in_flight value + mmrif_90c=$($LCTL get_param -n \ + mdc.$FSNAME-MDT0000-mdc-*.max_mod_rpcs_in_flight) + echo "max_mod_rpcs_in_flight is $mmrif_90c" + + # get MDT max_mod_rpcs_per_client value + mmrpc_90c=$(do_facet mds1 \ + cat /sys/module/mdt/parameters/max_mod_rpcs_per_client) + echo "max_mod_rpcs_per_client is $mmrpc_90c" +} + +restore_params_90c() { + trap 0 + + # restore max_rpcs_in_flight value + do_facet mgs $LCTL set_param -P \ + mdc.$FSNAME-MDT0000-mdc-*.max_rpcs_in_flight=$mrif_90c + + # restore max_mod_rpcs_in_flight value + do_facet mgs $LCTL set_param -P \ + mdc.$FSNAME-MDT0000-mdc-*.max_mod_rpcs_in_flight=$mmrif_90c + + # restore MDT max_mod_rpcs_per_client value + do_facet mds1 "echo $mmrpc_90c > \ + /sys/module/mdt/parameters/max_mod_rpcs_per_client" +} + test_90c() { local tmp - local mrif - local mmrpc setup @@ -6650,42 +6732,64 @@ test_90c() { skip "Client not able to send multiple modify RPCs in parallel" fi - # get max_rpcs_in_flight value - mrif=$($LCTL get_param -n mdc.$FSNAME-MDT0000-mdc-*.max_rpcs_in_flight) - echo "max_rpcs_in_flight is $mrif" - - # get MDT max_mod_rpcs_per_client - mmrpc=$(do_facet mds1 \ - cat /sys/module/mdt/parameters/max_mod_rpcs_per_client) - echo "max_mod_rpcs_per_client is $mmrpc" + save_params_90c + stack_trap restore_params_90c # testcase 1 # attempt to set max_mod_rpcs_in_flight to max_rpcs_in_flight value # prerequisite: set max_mod_rpcs_per_client to max_rpcs_in_flight value - umount_client $MOUNT - do_facet mds1 \ - "echo $mrif > /sys/module/mdt/parameters/max_mod_rpcs_per_client" - mount_client $MOUNT + do_facet mds1 "echo $mrif_90c > \ + /sys/module/mdt/parameters/max_mod_rpcs_per_client" - $LCTL set_param \ - mdc.$FSNAME-MDT0000-mdc-*.max_mod_rpcs_in_flight=$mrif && - error "set max_mod_rpcs_in_flight to $mrif should fail" + # if max_mod_rpcs_in_flight is set to be equal to or larger than + # max_rpcs_in_flight, then max_rpcs_in_flight will be increased + if [[ "$CLIENT_VERSION" -ge $(version_code 2.13.53) ]]; then + $LCTL set_param \ + mdc.$FSNAME-MDT0000-mdc-*.max_mod_rpcs_in_flight=$mrif_90c || + error "set max_mod_rpcs_in_flight to $mrif_90c failed" + + local new_mrif=$($LCTL get_param -n \ + mdc.$FSNAME-MDT0000-mdc-*.max_rpcs_in_flight) + ((new_mrif == mrif_90c + 1)) || + error "max_rpcs_in_flight was not increased" + fi umount_client $MOUNT - do_facet mds1 \ - "echo $mmrpc > /sys/module/mdt/parameters/max_mod_rpcs_per_client" + do_facet mds1 "echo $mmrpc_90c > \ + /sys/module/mdt/parameters/max_mod_rpcs_per_client" mount_client $MOUNT # testcase 2 # attempt to set max_mod_rpcs_in_flight to max_mod_rpcs_per_client+1 # prerequisite: set max_rpcs_in_flight to max_mod_rpcs_per_client+2 $LCTL set_param \ - mdc.$FSNAME-MDT0000-mdc-*.max_rpcs_in_flight=$((mmrpc + 2)) + mdc.$FSNAME-MDT0000-mdc-*.max_rpcs_in_flight=$((mmrpc_90c + 2)) $LCTL set_param \ - mdc.$FSNAME-MDT0000-mdc-*.max_mod_rpcs_in_flight=$((mmrpc + 1)) && - error "set max_mod_rpcs_in_flight to $((mmrpc + 1)) should fail" + mdc.$FSNAME-MDT0000-mdc-*.max_mod_rpcs_in_flight=$((mmrpc_90c + 1)) && + error "set max_mod_rpcs_in_flight to $((mmrpc_90c + 1)) should fail" + + # testcase 3 + # attempt to set max_mod_rpcs_in_flight permanently + do_facet mgs $LCTL set_param -P \ + mdc.$FSNAME-MDT0000-mdc-*.max_rpcs_in_flight=$mrif_90c + + do_facet mgs $LCTL set_param -P \ + mdc.$FSNAME-MDT0000-mdc-*.max_mod_rpcs_in_flight=$mrif_90c + + remount_client $MOUNT + + wait_update_facet --verbose client "$LCTL get_param -n \ + mdc.$FSNAME-MDT0000-mdc-*.max_rpcs_in_flight" \ + "$((mrif_90c + 1))" || + error "expected '$((mrif_90c + 1))' for max_rpcs_in_flight" + wait_update_facet --verbose client "$LCTL get_param -n \ + mdc.$FSNAME-MDT0000-mdc-*.max_mod_rpcs_in_flight" \ + "$mrif_90c" || + error "expected '$mrif_90c' for max_mod_rpcs_in_flight" + + restore_params_90c cleanup } run_test 90c "check max_mod_rpcs_in_flight update limits" @@ -6765,7 +6869,7 @@ check_uuid_on_ost() { check_uuid_on_mdt() { local nid=$1 - do_facet $SINGLEMDS "$LCTL get_param mdt.${FSNAME}*.exports.'$nid'.uuid" + do_facet $SINGLEMDS "$LCTL get_param mdt.${mds1_svc}*.exports.'$nid'.uuid" } test_91() { @@ -6809,6 +6913,7 @@ test_91() { echo "evict $nid" do_facet $SINGLEMDS \ "$LCTL set_param -n mdt.${mds1_svc}.evict_client nid:$nid" + sleep 1 # eviction above is async, give it some time to proceed found=$(check_uuid_on_mdt $nid | grep $uuid) [ -n "$found" ] && error "found $uuid $nid on MDT" @@ -7265,7 +7370,7 @@ test_98() for ((x = 1; x <= 400; x++)); do mountopt="$mountopt,user_xattr" done - remount_client $mountopt $MOUNT 2>&1 | grep "too long" || + mount_client $MOUNT remount,$mountopt 2>&1 | grep "too long" || error "Buffer overflow check failed" cleanup || error "cleanup failed" } @@ -7314,9 +7419,9 @@ test_100() { # Desired output # MGS: # 0@lo - # lustre-MDT0000: + # $FSNAME-MDT0000: # 0@lo - # lustre-OST0000: + # $FSNAME-OST0000: # 0@lo do_facet mgs 'lshowmount -v' | awk 'BEGIN {NR == 0; rc=1} /MGS:/ {rc=0} END {exit rc}' || error "lshowmount have no output MGS" @@ -7331,7 +7436,7 @@ test_100() { } run_test 100 "check lshowmount lists MGS, MDT, OST and 0@lo" -test_101() { +test_101a() { local createmany_pid local dev=$FSNAME-OST0000-osc-MDT0000 setup @@ -7358,7 +7463,29 @@ test_101() { unlinkmany $DIR1/$tdir/$tfile-%d 50000 cleanup } -run_test 101 "Race MDT->OST reconnection with create" +run_test 101a "Race MDT->OST reconnection with create" + +test_101b () { + local dev=$FSNAME-OST0000-osc-MDT0000 + local dir=$DIR1/$tdir + setup + + mkdir $dir + $LFS setstripe -c 1 -i 0 $dir + do_facet $SINGLEMDS "$LCTL --device $dev deactivate;" +#define OBD_FAIL_OSP_CON_EVENT_DELAY 0x2107 + do_facet mds1 "$LCTL set_param fail_loc=0x80002107 fail_val=20" + do_facet $SINGLEMDS "$LCTL --device $dev activate;" + stop_ost + sleep 25 + start_ost + + wait_osc_import_state client ost1 FULL + touch $dir/$tfile || error "Can't create file" + + cleanup +} +run_test 101b "Race events DISCONNECT and ACTIVE in osp" test_102() { [[ "$MDS1_VERSION" -gt $(version_code 2.9.53) ]] || @@ -7527,7 +7654,7 @@ test_103() { } run_test 103 "rename filesystem name" -test_104() { # LU-6952 +test_104a() { # LU-6952 local mds_mountopts=$MDS_MOUNT_OPTS local ost_mountopts=$OST_MOUNT_OPTS local mds_mountfsopts=$MDS_MOUNT_FS_OPTS @@ -7577,7 +7704,22 @@ test_104() { # LU-6952 OST_MOUNT_OPTS=$ost_mountopts MDS_MOUNT_FS_OPTS=$mds_mountfsopts } -run_test 104 "Make sure user defined options are reflected in mount" +run_test 104a "Make sure user defined options are reflected in mount" + +test_104b() { # LU-12859 + mount_client $MOUNT3 flock,localflock + stack_trap "umount_client $MOUNT3" EXIT + mount | grep "$MOUNT3 .*,flock" && error "flock is still set" + mount | grep "$MOUNT3 .*,localflock" || error "localflock is not set" + umount_client $MOUNT3 + mount_client $MOUNT3 localflock,flock + mount | grep "$MOUNT3 .*,localflock" && error "localflock is still set" + mount | grep "$MOUNT3 .*,flock" || error "flock is not set" + umount_client $MOUNT3 + mount_client $MOUNT3 localflock,flock,noflock + flock_is_enabled $MOUNT3 && error "some flock is still enabled" || true +} +run_test 104b "Mount uses last flock argument" error_and_umount() { umount $TMP/$tdir @@ -7795,10 +7937,10 @@ test_108a() { echo "changing server nid..." $rcmd mount -t lustre -o nosvc lustre-mdt1/mdt1 $tmp/mnt/mdt1 - $rcmd lctl replace_nids lustre-MDT0000 $nid - $rcmd lctl replace_nids lustre-MDT0001 $nid - $rcmd lctl replace_nids lustre-OST0000 $nid - $rcmd lctl replace_nids lustre-OST0001 $nid + $rcmd lctl replace_nids $FSNAME-MDT0000 $nid + $rcmd lctl replace_nids $FSNAME-MDT0001 $nid + $rcmd lctl replace_nids $FSNAME-OST0000 $nid + $rcmd lctl replace_nids $FSNAME-OST0001 $nid $rcmd umount $tmp/mnt/mdt1 for facet in $facets; do @@ -7857,10 +7999,10 @@ test_108b() { echo "changing server nid..." $rcmd mount -t lustre -o nosvc,loop $tmp/images/mdt1 $tmp/mnt/mdt1 - $rcmd lctl replace_nids lustre-MDT0000 $nid - $rcmd lctl replace_nids lustre-MDT0001 $nid - $rcmd lctl replace_nids lustre-OST0000 $nid - $rcmd lctl replace_nids lustre-OST0001 $nid + $rcmd lctl replace_nids $FSNAME-MDT0000 $nid + $rcmd lctl replace_nids $FSNAME-MDT0001 $nid + $rcmd lctl replace_nids $FSNAME-OST0000 $nid + $rcmd lctl replace_nids $FSNAME-OST0001 $nid $rcmd umount $tmp/mnt/mdt1 for facet in $facets; do @@ -7870,7 +8012,7 @@ test_108b() { done for facet in $scrub_list; do - $rcmd $LCTL lfsck_start -M lustre-$facet -t scrub || + $rcmd $LCTL lfsck_start -M $FSNAME-$facet -t scrub || error "failed to start OI scrub on $facet" done @@ -8296,6 +8438,36 @@ test_111() { } run_test 111 "Adding large_dir with over 2GB directory" +test_112() { + start_mds || error "MDS start failed" + start_ost || error "OSS start failed" + echo "start ost2 service on $(facet_active_host ost2)" + start ost2 $(ostdevname 2) $(csa_add "$OST_MOUNT_OPTS" -o no_precreate) || + error "start ost2 facet failed" + local val=$(do_facet ost2 \ + "$LCTL get_param -n obdfilter.$FSNAME-OST0001*.no_precreate") + (( $val == 1 )) || error "obdfilter.$FSNAME-OST0001*.no_precreate=$val" + + mount_client $MOUNT || error "mount client failed" + wait_osc_import_state mds1 ost1 FULL + wait_osc_import_state client ost1 FULL + wait_osc_import_state client ost2 FULL + + $LFS setstripe -i 0 $DIR/$tfile.0 || + error "problem creating $tfile.0 on OST0000" + $LFS setstripe -i 1 $DIR/$tfile.1 && $LFS getstripe $DIR/$tfile.1 && + (( $($LFS getstripe -i $DIR/$tfile.1) == 1 )) && + error "allowed to create $tfile.1 on OST0001" + do_facet ost2 $LCTL set_param obdfilter.*.no_precreate=0 + sleep_maxage + $LFS setstripe -i 1 $DIR/$tfile.2 || + error "failed to create $tfile.2 on ost1 facet" + # files not cleaned with ONLY_REPEAT because of client unmount below + rm -f $DIR/$tfile.[012] + stop_ost2 || error "stop ost2 facet failed" + cleanup +} +run_test 112 "mount OST with nocreate option" cleanup_115() { @@ -8308,6 +8480,7 @@ test_115() { if [ "$mds1_FSTYPE" != ldiskfs ]; then skip "Only applicable to ldiskfs-based MDTs" fi + [ -n "$FILESET" ] && skip "Not functional for FILESET set" local dbfs_ver=$(do_facet $SINGLEMDS $DEBUGFS -V 2>&1) @@ -8321,6 +8494,14 @@ test_115() { IMAGESIZE=$((3072 << 30)) # 3072 GiB stopall + + echo "client1: " + lctl dl + mount | grep lustre + echo "mds1: " + do_facet mds1 "hostname; ifconfig; lctl dl; mount" + echo "ost1: " + do_facet ost1 "hostname; ifconfig; lctl dl; mount" # We need MDT size 3072GB, because it is smallest # partition that can store 2B inodes do_facet $SINGLEMDS "mkdir -p $TMP/$tdir" @@ -8333,9 +8514,9 @@ test_115() { local mdsdev=$(do_facet $SINGLEMDS "losetup -f") do_facet $SINGLEMDS "losetup $mdsdev $mdsimgname" - local mds_opts="$(mkfs_opts mds1 $(mdsdevname 1)) --device-size=$IMAGESIZE \ + local mds_opts="$(mkfs_opts mds1 $(mdsdevname 1)) \ --mkfsoptions='-O ea_inode,^resize_inode,meta_bg \ - -N 2247484000 -E lazy_itable_init'" + -N 2247484000 -E lazy_itable_init' --device-size=$IMAGESIZE" add mds1 $mds_opts --mgs --reformat $mdsdev || skip_env "format large MDT failed" opts="$(mkfs_opts ost1 $(ostdevname 1)) \ @@ -8403,7 +8584,10 @@ test_116() { MDSSIZE=$old_mdssize MDSISIZE=$old_mdsisize - do_facet $SINGLEMDS $MKFS $opts17t $tmpmnt/$mdtimg || + + do_facet $SINGLEMDS $MKFS \ + ${opts17t/-E /-E lazy_itable_init,lazy_journal_init,} \ + $tmpmnt/$mdtimg || error "failed to mkfs for $tmpmnt/$mdtimg" do_facet $SINGLEMDS $TUNE2FS -l $tmpmnt/$mdtimg | @@ -8496,18 +8680,22 @@ test_123ab() { local yaml local orig_val + local mgs_arg="" + + [[ $MGS_VERSION -gt $(version_code 2.13.54) ]] || + mgs_arg="--device MGS" orig_val=$(do_facet mgs $LCTL get_param jobid_name) - do_facet mgs $LCTL set_param -P jobid_name="testname" + do_facet mgs $LCTL set_param -P jobid_name="TESTNAME" - yaml=$(do_facet mgs $LCTL --device MGS llog_print params | + yaml=$(do_facet mgs $LCTL $mgs_arg llog_print params | grep jobid_name | tail -n 1) local param=$(awk '{ print $10 }' <<< "$yaml") local val=$(awk '{ print $12 }' <<< "$yaml") #return to the default do_facet mgs $LCTL set_param -P jobid_name=$orig_val - [ $val = "testname" ] || error "bad value: $val" + [ $val = "TESTNAME" ] || error "bad value: $val" [ $param = "jobid_name," ] || error "Bad param: $param" } run_test 123ab "llog_print params output values from set_param -P" @@ -8519,12 +8707,16 @@ test_123ac() { # LU-11566 local start=10 local end=50 + local mgs_arg="" + + [[ $MGS_VERSION -gt $(version_code 2.13.54) ]] || + mgs_arg="--device MGS" [ -d $MOUNT/.lustre ] || setup # - { index: 10, event: add_uuid, nid: 192.168.20.1@tcp(0x20000c0a81401, # node: 192.168.20.1@tcp } - do_facet mgs $LCTL --device MGS \ + do_facet mgs $LCTL $mgs_arg \ llog_print --start $start --end $end $FSNAME-client | tr -d , | while read DASH BRACE INDEX idx EVENT BLAH BLAH BLAH; do (( idx >= start )) || error "llog_print index $idx < $start" @@ -8564,26 +8756,30 @@ test_123ae() { # LU-11566 [ -d $MOUNT/.lustre ] || setupall local max=$($LCTL get_param -n osc.*-OST0000-*.max_dirty_mb | head -1) + local mgs_arg="" + + [[ $MGS_VERSION -gt $(version_code 2.13.54) ]] || + mgs_arg="--device MGS" if do_facet mgs "$LCTL help llog_cancel" 2>&1| grep -q -- --log_id; then # save one set_param -P record in case none exist do_facet mgs $LCTL set_param -P osc.*.max_dirty_mb=$max local log=params - local orig=$(do_facet mgs $LCTL --device MGS llog_print $log | + local orig=$(do_facet mgs $LCTL $mgs_arg llog_print $log | tail -1 | awk '{ print $4 }' | tr -d , ) do_facet mgs $LCTL set_param -P osc.*.max_dirty_mb=$max - do_facet mgs $LCTL --device MGS llog_print $log | tail -1 | + do_facet mgs $LCTL $mgs_arg llog_print $log | tail -1 | grep "parameter: osc.*.max_dirty_mb" || error "new set_param -P wasn't stored in params log" # - { index: 71, event: set_param, device: general, # param: osc.*.max_dirty_mb, value: 256 } - local id=$(do_facet mgs $LCTL --device MGS llog_print $log | + local id=$(do_facet mgs $LCTL $mgs_arg llog_print $log | tail -1 | awk '{ print $4 }' | tr -d , ) - do_facet mgs $LCTL --device MGS llog_cancel $log --log_idx=$id - local new=$(do_facet mgs $LCTL --device MGS llog_print $log | + do_facet mgs $LCTL $mgs_arg llog_cancel $log --log_idx=$id + local new=$(do_facet mgs $LCTL $mgs_arg llog_print $log | tail -1 | awk '{ print $4 }' | tr -d , ) (( new == orig )) || error "new llog_cancel now $new, not at $orig records" @@ -8612,6 +8808,55 @@ test_123ae() { # LU-11566 } run_test 123ae "llog_cancel can cancel requested record" +test_123af() { #LU-13609 + [ "$MGS_VERSION" -ge $(version_code 2.13.54) -a \ + "$MDS1_VERSION" -ge $(version_code 2.13.54) ] || + skip "Need both MGS and MDS version at least 2.13.54" + + [ -d $MOUNT/.lustre ] || setupall + stack_trap "do_facet mds1 $LCTL set_param fail_loc=0" EXIT + + local device + local facet + local cmd + local orig_clist + local orig_count + local new_clist + local new_count + + for device in "MGS" "$FSNAME-MDT0000"; do + cmd="--device $device llog_catlist" + echo "lctl $cmd ..." + if [ "$device" = "MGS" ]; then + facet="mgs" + else + facet="mds1" + fi + orig_clist=($(do_facet $facet $LCTL $cmd | awk '{ print $2 }')) + orig_count=${#orig_clist[@]} + echo "orig_clist: ${orig_clist[@]}" + + #define OBD_FAIL_CATLIST 0x131b + #fetch to llog records from the second one + do_facet $facet $LCTL set_param fail_loc=0x131b fail_val=2 + + new_clist=($(do_facet $facet $LCTL $cmd | awk '{ print $2 }')) + new_count=${#new_clist[@]} + echo "new_clist: ${new_clist[@]}" + + [ $new_count -eq $((orig_count - 1)) ] || + error "$new_count != $orig_count - 1" + for i in $(seq 0 $new_count); do + j=$((i + 1)) + [ "${orig_clist[$j]}" = "${new_clist[$i]}" ] || + error "${orig_clist[$j]} != ${new_clist[$i]}" + done + do_facet mds1 $LCTL set_param fail_loc=0 + echo "done" + done +} +run_test 123af "llog_catlist can show all config files correctly" + test_123F() { remote_mgs_nodsh && skip "remote MGS with nodsh" @@ -8623,7 +8868,7 @@ test_123F() { # set jobid_var to a different value for test local orig_val=$(do_facet mgs $LCTL get_param jobid_var) - do_facet mgs $LCTL set_param -P jobid_var="testname" + do_facet mgs $LCTL set_param -P jobid_var="TESTNAME" for i in $cfgfiles params; do do_facet mgs "lctl --device MGS llog_print ${i} >> $yaml_file" @@ -8646,8 +8891,8 @@ test_123F() { local set_val=$(do_facet mgs $LCTL get_param jobid_var) do_facet mgs $LCTL set_param -P $orig_val - [ $set_val == "jobid_var=testname" ] || - error "$set_val is not testname" + [ $set_val == "jobid_var=TESTNAME" ] || + error "$set_val is not TESTNAME" do_facet mgs rm "$yaml_file" cleanup @@ -8660,6 +8905,12 @@ test_124() [ -z $mds2failover_HOST ] && skip "needs MDT failover setup" setup + do_facet mgs $LCTL --device MGS llog_print $FSNAME-client | + grep 1.2.3.4@tcp && error "Should not be fake nid" + do_facet mgs $LCTL conf_param $FSNAME-MDT0001.failover.node=1.2.3.4@tcp\ + || error "Set params error" + do_facet mgs $LCTL --device MGS llog_print $FSNAME-client | + grep 1.2.3.4@tcp || error "Fake nid should be added" cleanup load_modules @@ -8798,7 +9049,7 @@ test_125() echo "Before: ${facet} ${device} ${pre_max_sectors} ${max_hw_sectors}" - do_facet ${facet} "libtool execute l_tunedisk ${device}" + do_facet ${facet} "l_tunedisk ${device}" # Value after l_tunedisk local post_max_sectors=$(get_max_sectors_kb ${facet} ${device}) @@ -8827,6 +9078,90 @@ test_125() } run_test 125 "check l_tunedisk only tunes OSTs and their slave devices" +test_126() { + [[ "$MDS1_VERSION" -ge $(version_code 2.13.52) ]] || + skip "Need MDS version at least 2.13.52" + + cleanup + do_rpc_nodes $(facet_active_host $SINGLEMDS) load_module ../libcfs/libcfs/libcfs + #define OBD_FAIL_OBD_SETUP 0x60d + do_facet mds1 $LCTL set_param fail_loc=0x60d + do_rpc_nodes $(facet_active_host $SINGLEMDS) load_modules & + for i in {1..40}; do + do_facet mds1 lsmod | grep -q osd_$mds1_FSTYPE && break + sleep 1 + done + clear_failloc $SINGLEMDS 20 & + start mds1 $(mdsdevname 1) $MDS_MOUNT_OPTS +} +run_test 126 "mount in parallel shouldn't cause a crash" + +test_127() { + [[ "$ost1_FSTYPE" == ldiskfs ]] || skip "ldiskfs only test" + + cleanup + setup + zconf_umount_clients $RCLIENTS $MOUNT + + wait_osp_active ost ${FSNAME}-OST0000 0 1 + local osc_tgt="$FSNAME-OST0000-osc-$($LFS getname -i $DIR)" + local avail1=($($LCTL get_param -n osc.${osc_tgt}.kbytesavail)) + + $LFS setstripe -i 0 $DIR/$tfile || error "failed creating $DIR/$tfile" + dd if=/dev/zero of=$DIR/$tfile bs=1M oflag=direct || true + + local avail2=($($LCTL get_param -n osc.${osc_tgt}.kbytesavail)) + + if ((avail2 * 100 / avail1 > 1)); then + lfs df $DIR + ls -l $DIR/$tfile + error "more than 1% space left: before=$avail1 after=$avail2" + fi + + local mbs=$(($(stat -c %s $DIR/$tfile) / (1024 * 1024))) + + dd if=/dev/zero of=$DIR/$tfile bs=1M count=$mbs conv=notrunc \ + oflag=direct || error "overwrite failed" +} +run_test 127 "direct io overwrite on full ost" + +test_128() +{ + combined_mgs_mds && skip "need separate mgs device" + [ "$ost2_FSTYPE" == zfs ] && import_zpool ost2 + + format_ost 2 + # Try to apply nolocallogs to the virgin OST. Should fail. + do_facet ost2 "$TUNEFS --nolocallogs $(ostdevname 2)" && + error "nolocallogs should not be allowed on the virgin target" + + setupall + stopall + + [ "$ost1_FSTYPE" == zfs ] && import_zpool ost1 + # Start OST without MGS (local configs) + do_facet ost1 "$TUNEFS --dryrun $(ostdevname 1)" + start_ost || error "unable to start OST1" + stop_ost || error "Unable to stop OST1" + + [ "$ost1_FSTYPE" == zfs ] && import_zpool ost1 + # Do not allow reading local configs, should fail + do_facet ost1 "$TUNEFS --nolocallogs $(ostdevname 1)" || + error "Can not set nolocallogs" + start_ost && error "OST1 started, but should fail" + + # Connect to MGS successfully, reset nolocallogs flag + [ "$ost1_FSTYPE" == zfs ] && import_zpool ost1 + start_mgs || error "unable to start MGS" + start_ost || error "unable to start OST1" + + do_facet ost1 "$TUNEFS --dryrun $(ostdevname 1)" | grep "nolocallogs" && + error "nolocallogs expected to be reset" + + stop_ost || error "Unable to stop OST1" +} +run_test 128 "Force using remote logs with --nolocallogs" + if ! combined_mgs_mds ; then stop mgs fi