From: Yu Jian Date: Fri, 6 Apr 2012 03:51:10 +0000 (+0800) Subject: LU-1213 tests: remove duplicate parallel-scale variables X-Git-Tag: 2.2.55~46 X-Git-Url: https://git.whamcloud.com/?p=fs%2Flustre-release.git;a=commitdiff_plain;h=66397f7661093aba422aadbfd288f24bd434ff75 LU-1213 tests: remove duplicate parallel-scale variables Remove the duplicate variables from parallel-scale.sh and parallel-scale-nfs.sh. Those variables are defined inside the corresponding functions in functions.sh. Signed-off-by: Yu Jian Change-Id: I77ea061289296aeb8b138631f7a55090f52093c9 Reviewed-on: http://review.whamcloud.com/2306 Tested-by: Hudson Tested-by: Maloo Reviewed-by: Cliff White Reviewed-by: Minh Diep Reviewed-by: Oleg Drokin --- diff --git a/lustre/tests/functions.sh b/lustre/tests/functions.sh index 9e6976d..912fc9b 100644 --- a/lustre/tests/functions.sh +++ b/lustre/tests/functions.sh @@ -734,8 +734,8 @@ run_write_append_truncate() { run_write_disjoint() { - WRITE_DISJOINT=${WRITE_DISJOINT:-\ - $(which write_disjoint 2> /dev/null || true)} + WRITE_DISJOINT=${WRITE_DISJOINT:-$(which write_disjoint \ + 2> /dev/null || true)} # threads per client wdisjoint_THREADS=${wdisjoint_THREADS:-4} wdisjoint_REP=${wdisjoint_REP:-10000} @@ -773,8 +773,8 @@ run_write_disjoint() { run_parallel_grouplock() { - PARALLEL_GROUPLOCK=${PARALLEL_GROUPLOCK:-\ - $(which parallel_grouplock 2> /dev/null || true)} + PARALLEL_GROUPLOCK=${PARALLEL_GROUPLOCK:-$(which parallel_grouplock \ + 2> /dev/null || true)} parallel_grouplock_MINTASKS=${parallel_grouplock_MINTASKS:-5} if [ "$NFSCLIENT" ]; then diff --git a/lustre/tests/parallel-scale-nfs.sh b/lustre/tests/parallel-scale-nfs.sh index 9955dc1..0182c7a 100755 --- a/lustre/tests/parallel-scale-nfs.sh +++ b/lustre/tests/parallel-scale-nfs.sh @@ -42,7 +42,6 @@ NFSCLIENT=true FAIL_ON_ERROR=false # common setup -# MACHINEFILE=${MACHINEFILE:-$TMP/$(basename $0 .sh).machines} clients=${CLIENTS:-$HOSTNAME} generate_machine_file $clients $MACHINEFILE || \ @@ -50,44 +49,18 @@ generate_machine_file $clients $MACHINEFILE || \ num_clients=$(get_node_count ${clients//,/ }) # compilbench -# -cbench_DIR=${cbench_DIR:-"/usr/bin"} -cbench_IDIRS=${cbench_IDIRS:-4} -# FIXME: wiki page requirements is 30, do we really need 30 ? -cbench_RUNS=${cbench_RUNS:-4} - if [ "$SLOW" = "no" ]; then cbench_IDIRS=2 cbench_RUNS=2 fi -# # metabench -# -METABENCH=${METABENCH:-$(which metabench 2> /dev/null || true)} -mbench_NFILES=${mbench_NFILES:-30400} [ "$SLOW" = "no" ] && mbench_NFILES=10000 -# threads per client -mbench_THREADS=${mbench_THREADS:-4} -# # connectathon -# -cnt_DIR=${cnt_DIR:-""} -cnt_NRUN=${cnt_NRUN:-10} [ "$SLOW" = "no" ] && cnt_NRUN=2 -# # IOR -# -IOR=${IOR:-$(which IOR 2> /dev/null || true)} -# threads per client -ior_THREADS=${ior_THREADS:-2} -ior_iteration=${ior_iteration:-1} -ior_blockSize=${ior_blockSize:-6} # Gb -ior_xferSize=${ior_xferSize:-2m} -ior_type=${ior_type:-POSIX} -ior_DURATION=${ior_DURATION:-60} # minutes [ "$SLOW" = "no" ] && ior_DURATION=30 # source the common file after all parameters are set to take affect diff --git a/lustre/tests/parallel-scale.sh b/lustre/tests/parallel-scale.sh index 66b6d50..28c42af 100644 --- a/lustre/tests/parallel-scale.sh +++ b/lustre/tests/parallel-scale.sh @@ -12,119 +12,39 @@ init_logging ALWAYS_EXCEPT="parallel_grouplock $PARALLEL_SCALE_EXCEPT" # common setup -# MACHINEFILE=${MACHINEFILE:-$TMP/$(basename $0 .sh).machines} clients=${CLIENTS:-$HOSTNAME} generate_machine_file $clients $MACHINEFILE || \ error "Failed to generate machine file" num_clients=$(get_node_count ${clients//,/ }) - # compilbench -# -cbench_DIR=${cbench_DIR:-""} -cbench_IDIRS=${cbench_IDIRS:-4} -# FIXME: wiki page requirements is 30, do we really need 30 ? -cbench_RUNS=${cbench_RUNS:-4} - if [ "$SLOW" = "no" ]; then cbench_IDIRS=2 cbench_RUNS=2 fi -# # metabench -# -METABENCH=${METABENCH:-$(which metabench 2> /dev/null || true)} -mbench_NFILES=${mbench_NFILES:-30400} [ "$SLOW" = "no" ] && mbench_NFILES=10000 -# threads per client -mbench_THREADS=${mbench_THREADS:-4} -# # simul -# -SIMUL=${SIMUL:=$(which simul 2> /dev/null || true)} -# threads per client -simul_THREADS=${simul_THREADS:-2} -simul_REP=${simul_REP:-20} [ "$SLOW" = "no" ] && simul_REP=2 -# -# mib -# -MIB=${MIB:=$(which mib 2> /dev/null || true)} -# threads per client -mib_THREADS=${mib_THREADS:-2} -mib_xferSize=${mib_xferSize:-1m} -mib_xferLimit=${mib_xferLimit:-5000} -mib_timeLimit=${mib_timeLimit:-300} - -# -# MDTEST -# -MDTEST=${MDTEST:=$(which mdtest 2> /dev/null || true)} -# threads per client -mdtest_THREADS=${mdtest_THREADS:-2} -mdtest_nFiles=${mdtest_nFiles:-"100000"} -# We devide the files by number of core -mdtest_nFiles=$((mdtest_nFiles/mdtest_THREADS/num_clients)) -mdtest_iteration=${mdtest_iteration:-1} - -# # connectathon -# -cnt_DIR=${cnt_DIR:-""} -cnt_NRUN=${cnt_NRUN:-10} [ "$SLOW" = "no" ] && cnt_NRUN=2 -# # cascading rw -# -CASC_RW=${CASC_RW:-$(which cascading_rw 2> /dev/null || true)} -# threads per client -casc_THREADS=${casc_THREADS:-2} -casc_REP=${casc_REP:-300} [ "$SLOW" = "no" ] && casc_REP=10 -# # IOR -# -IOR=${IOR:-$(which IOR 2> /dev/null || true)} -# threads per client -ior_THREADS=${ior_THREADS:-2} -ior_iteration=${ior_iteration:-1} -ior_blockSize=${ior_blockSize:-6} # Gb -ior_xferSize=${ior_xferSize:-2m} -ior_type=${ior_type:-POSIX} -ior_DURATION=${ior_DURATION:-30} # minutes [ "$SLOW" = "no" ] && ior_DURATION=5 -# # write_append_truncate -# -# threads per client -write_THREADS=${write_THREADS:-8} -write_REP=${write_REP:-10000} [ "$SLOW" = "no" ] && write_REP=100 -# # write_disjoint -# -WRITE_DISJOINT=${WRITE_DISJOINT:-$(which write_disjoint 2> /dev/null || true)} -# threads per client -wdisjoint_THREADS=${wdisjoint_THREADS:-4} -wdisjoint_REP=${wdisjoint_REP:-10000} [ "$SLOW" = "no" ] && wdisjoint_REP=100 -# -# parallel_grouplock -# -# -PARALLEL_GROUPLOCK=${PARALLEL_GROUPLOCK:-\ - $(which parallel_grouplock 2> /dev/null || true)} -parallel_grouplock_MINTASKS=${parallel_grouplock_MINTASKS:-5} - . $LUSTRE/tests/functions.sh build_test_filter @@ -199,9 +119,6 @@ test_parallel_grouplock() { } run_test parallel_grouplock "parallel_grouplock" -statahead_NUMMNTPTS=${statahead_NUMMNTPTS:-5} -statahead_NUMFILES=${statahead_NUMFILES:-500000} - test_statahead () { run_statahead }