Whamcloud - gitweb
git://git.whamcloud.com
/
fs
/
lustre-release.git
/ blobdiff
commit
grep
author
committer
pickaxe
?
search:
re
summary
|
shortlog
|
log
|
commit
|
commitdiff
|
tree
raw
|
inline
| side by side
LU-13717 sec: filename encryption
[fs/lustre-release.git]
/
lustre
/
tests
/
parallel-scale.sh
diff --git
a/lustre/tests/parallel-scale.sh
b/lustre/tests/parallel-scale.sh
index
af4ad5d
..
c82a302
100644
(file)
--- a/
lustre/tests/parallel-scale.sh
+++ b/
lustre/tests/parallel-scale.sh
@@
-1,17
+1,15
@@
#!/bin/bash
#!/bin/bash
-#
-#set -vx
-LUSTRE=${LUSTRE:-$(
cd $(dirname $0)/..; echo $PWD)
}
+LUSTRE=${LUSTRE:-$(
dirname $0)/..
}
. $LUSTRE/tests/test-framework.sh
init_test_env $@
. $LUSTRE/tests/test-framework.sh
init_test_env $@
-. ${CONFIG:=$LUSTRE/tests/cfg/$NAME.sh}
init_logging
init_logging
-# bug 20670
-ALWAYS_EXCEPT="parallel_grouplock $PARALLEL_SCALE_EXCEPT"
+ALWAYS_EXCEPT="$PARALLEL_SCALE_EXCEPT "
+# bug number for skipped test: LU-9429
+ALWAYS_EXCEPT+=" parallel_grouplock "
-if [
$(facet_fstype $SINGLEMDS) = zfs -o $(facet_fstype "ost1")
= zfs ]; then
+if [
"$mds1_FSTYPE" = zfs -o "$ost1_FSTYPE"
= zfs ]; then
ZFSSLOW=$SLOW
SLOW=no
ZFSSLOW=$SLOW
SLOW=no
@@
-22,11
+20,12
@@
if [ $(facet_fstype $SINGLEMDS) = zfs -o $(facet_fstype "ost1") = zfs ]; then
statahead_NUMFILES=${statahead_NUMFILES:-100000}
fi
statahead_NUMFILES=${statahead_NUMFILES:-100000}
fi
+build_test_filter
+
# common setup
# common setup
-MACHINEFILE=${MACHINEFILE:-$TMP/$(basename $0 .sh).machines}
clients=${CLIENTS:-$HOSTNAME}
generate_machine_file $clients $MACHINEFILE ||
clients=${CLIENTS:-$HOSTNAME}
generate_machine_file $clients $MACHINEFILE ||
- error "Failed to generate machine file"
+
error "Failed to generate machine file"
num_clients=$(get_node_count ${clients//,/ })
# compilbench
num_clients=$(get_node_count ${clients//,/ })
# compilbench
@@
-56,9
+55,17
@@
fi
# write_disjoint
[ "$SLOW" = "no" ] && wdisjoint_REP=${wdisjoint_REP:-100}
# write_disjoint
[ "$SLOW" = "no" ] && wdisjoint_REP=${wdisjoint_REP:-100}
+# fs_test
+if [ "$SLOW" = "no" ]; then
+ fs_test_ndirs=${fs_test_ndirs:-10000}
+ fs_test_nobj=${fs_test_nobj:-2}
+fi
+
+# xdd
+[ "$SLOW" = "no" ] && xdd_passes=${xdd_passes:-15}
+
. $LUSTRE/tests/functions.sh
. $LUSTRE/tests/functions.sh
-build_test_filter
check_and_setup_lustre
get_mpiuser_id $MPI_USER
check_and_setup_lustre
get_mpiuser_id $MPI_USER
@@
-105,6
+112,16
@@
test_iorfpp() {
}
run_test iorfpp "iorfpp"
}
run_test iorfpp "iorfpp"
+test_ior_mdtest_parallel_ssf() {
+ ior_mdtest_parallel "ssf"
+}
+run_test ior_mdtest_parallel_ssf "iormdtestssf"
+
+test_ior_mdtest_parallel_fpp() {
+ ior_mdtest_parallel "fpp"
+}
+run_test ior_mdtest_parallel_fpp "iormdtestfpp"
+
test_mib() {
run_mib
}
test_mib() {
run_mib
}
@@
-120,11
+137,18
@@
test_write_append_truncate() {
}
run_test write_append_truncate "write_append_truncate"
}
run_test write_append_truncate "write_append_truncate"
+# Argument is chunk size limit, the upper bound on write size
test_write_disjoint() {
test_write_disjoint() {
- run_write_disjoint
+ run_write_disjoint
123456
}
run_test write_disjoint "write_disjoint"
}
run_test write_disjoint "write_disjoint"
+# Make sure to exercise the tiny write code
+test_write_disjoint_tiny() {
+ run_write_disjoint 16384
+}
+run_test write_disjoint_tiny "write_disjoint_tiny"
+
test_parallel_grouplock() {
run_parallel_grouplock
}
test_parallel_grouplock() {
run_parallel_grouplock
}
@@
-135,7
+159,28
@@
test_statahead () {
}
run_test statahead "statahead test, multiple clients"
}
run_test statahead "statahead test, multiple clients"
-[ $(facet_fstype $SINGLEMDS) = zfs -o $(facet_fstype "ost1") = zfs ] &&
+test_rr_alloc () {
+ run_rr_alloc
+}
+run_test rr_alloc "Checking even file distribution over OSTs in RR policy"
+
+test_fs_test () {
+ run_fs_test
+}
+run_test fs_test "fs_test"
+
+test_fio () {
+ run_fio
+}
+run_test fio "fio"
+
+test_xdd () {
+ run_xdd
+}
+run_test xdd "xdd"
+
+# If necessary, return SLOW to its original value
+[ "$mds1_FSTYPE" = zfs -o "$ost1_FSTYPE" = zfs ] &&
SLOW=$ZFSSLOW
complete $SECONDS
SLOW=$ZFSSLOW
complete $SECONDS