diff options
| author | Pranith Kumar K <pkarampu@redhat.com> | 2014-05-19 13:29:06 +0530 | 
|---|---|---|
| committer | Anand Avati <avati@redhat.com> | 2014-05-19 21:00:46 -0700 | 
| commit | 7cd32c18a11d0dfeb70c276b0185e21d06a71fa5 (patch) | |
| tree | 175980b522a99d89c09034c1f74489105f89619f | |
| parent | c13c1fd5a487eeca269e8b62114f607487ee30f6 (diff) | |
tests: Use uniform timeouts
Change-Id: I479ab941b3b2da3b16f624400fbd300f08326268
BUG: 1092850
Signed-off-by: Pranith Kumar K <pkarampu@redhat.com>
Reviewed-on: http://review.gluster.org/7799
Tested-by: Gluster Build System <jenkins@build.gluster.com>
Reviewed-by: Anand Avati <avati@redhat.com>
79 files changed, 219 insertions, 215 deletions
diff --git a/tests/basic/afr/read-subvol-data.t b/tests/basic/afr/read-subvol-data.t index 7db4988faee..25ae3e63c64 100644 --- a/tests/basic/afr/read-subvol-data.t +++ b/tests/basic/afr/read-subvol-data.t @@ -24,7 +24,7 @@ TEST dd if=/dev/urandom of=$M0/afr_success_5.txt bs=1M count=1  TEST kill_brick $V0 $H0 $B0/brick0  TEST dd if=/dev/urandom of=$M0/afr_success_5.txt bs=1M count=10  TEST $CLI volume start $V0 force -EXPECT_WITHIN 5 "10485760" echo `ls -l $M0/afr_success_5.txt | awk '{ print $5}'` +EXPECT_WITHIN $PROCESS_UP_TIMEOUT "10485760" echo `ls -l $M0/afr_success_5.txt | awk '{ print $5}'`  #Cleanup  TEST umount $M0 diff --git a/tests/basic/afr/read-subvol-entry.t b/tests/basic/afr/read-subvol-entry.t index 91110b8cd70..3e7ee3f35dd 100644 --- a/tests/basic/afr/read-subvol-entry.t +++ b/tests/basic/afr/read-subvol-entry.t @@ -26,7 +26,7 @@ TEST kill_brick $V0 $H0 $B0/brick0  TEST touch $M0/abc/def/ghi  TEST $CLI volume start $V0 force -EXPECT_WITHIN 5 "ghi" echo `ls $M0/abc/def/` +EXPECT_WITHIN $PROCESS_UP_TIMEOUT "ghi" echo `ls $M0/abc/def/`  #Cleanup  TEST umount $M0 diff --git a/tests/basic/afr/self-heal.t b/tests/basic/afr/self-heal.t index df9526bcf88..575ed4c4f72 100644 --- a/tests/basic/afr/self-heal.t +++ b/tests/basic/afr/self-heal.t @@ -39,12 +39,12 @@ TEST dd if=/dev/urandom of=$M0/jkl/mno/file.txt bs=1M count=4 2>/dev/null  TEST chown  $NEW_UID:$NEW_GID $M0/def/ghi/file2.txt  TEST $CLI volume start $V0 force -EXPECT_WITHIN 20 "1" afr_child_up_status $V0 0 -EXPECT_WITHIN 20 "Y" glustershd_up_status -EXPECT_WITHIN 20 "1" afr_child_up_status_in_shd $V0 0 -EXPECT_WITHIN 20 "1" afr_child_up_status_in_shd $V0 1 +EXPECT_WITHIN $CHILD_UP_TIMEOUT "1" afr_child_up_status $V0 0 +EXPECT_WITHIN $PROCESS_UP_TIMEOUT "Y" glustershd_up_status +EXPECT_WITHIN $CHILD_UP_TIMEOUT "1" afr_child_up_status_in_shd $V0 0 +EXPECT_WITHIN $CHILD_UP_TIMEOUT "1" afr_child_up_status_in_shd $V0 1  TEST $CLI volume heal $V0 -EXPECT_WITHIN 20 "0" afr_get_pending_heal_count $V0 +EXPECT_WITHIN $HEAL_TIMEOUT "0" afr_get_pending_heal_count $V0  #check all files created/deleted on brick1 are also replicated on brick 0  #(i.e. no reverse heal has happened) @@ -68,12 +68,12 @@ TEST rm -f $M0/file  TEST mkdir $M0/file  TEST $CLI volume start $V0 force -EXPECT_WITHIN 20 "1" afr_child_up_status $V0 0 -EXPECT_WITHIN 20 "Y" glustershd_up_status -EXPECT_WITHIN 20 "1" afr_child_up_status_in_shd $V0 0 -EXPECT_WITHIN 20 "1" afr_child_up_status_in_shd $V0 1 +EXPECT_WITHIN $CHILD_UP_TIMEOUT "1" afr_child_up_status $V0 0 +EXPECT_WITHIN $PROCESS_UP_TIMEOUT "Y" glustershd_up_status +EXPECT_WITHIN $CHILD_UP_TIMEOUT "1" afr_child_up_status_in_shd $V0 0 +EXPECT_WITHIN $CHILD_UP_TIMEOUT "1" afr_child_up_status_in_shd $V0 1  TEST $CLI volume heal $V0 -EXPECT_WITHIN 20 "0" afr_get_pending_heal_count $V0 +EXPECT_WITHIN $HEAL_TIMEOUT "0" afr_get_pending_heal_count $V0  #check heal has happened in the correct direction  TEST test -d $B0/brick0/file @@ -91,12 +91,12 @@ TEST chmod 666 $M0/file  TEST kill_brick $V0 $H0 $B0/brick0  TEST chmod 777 $M0/file  TEST $CLI volume start $V0 force -EXPECT_WITHIN 20 "1" afr_child_up_status $V0 0 -EXPECT_WITHIN 20 "Y" glustershd_up_status -EXPECT_WITHIN 20 "1" afr_child_up_status_in_shd $V0 0 -EXPECT_WITHIN 20 "1" afr_child_up_status_in_shd $V0 1 +EXPECT_WITHIN $CHILD_UP_TIMEOUT "1" afr_child_up_status $V0 0 +EXPECT_WITHIN $PROCESS_UP_TIMEOUT "Y" glustershd_up_status +EXPECT_WITHIN $CHILD_UP_TIMEOUT "1" afr_child_up_status_in_shd $V0 0 +EXPECT_WITHIN $CHILD_UP_TIMEOUT "1" afr_child_up_status_in_shd $V0 1  TEST $CLI volume heal $V0 -EXPECT_WITHIN 20 "0" afr_get_pending_heal_count $V0 +EXPECT_WITHIN $HEAL_TIMEOUT "0" afr_get_pending_heal_count $V0  #check heal has happened in the correct direction  EXPECT "777" stat --printf=%a $B0/brick0/file @@ -115,12 +115,12 @@ NEW_UID=36  NEW_GID=36  TEST chown $NEW_UID:$NEW_GID $M0/file  TEST $CLI volume start $V0 force -EXPECT_WITHIN 20 "1" afr_child_up_status $V0 0 -EXPECT_WITHIN 20 "Y" glustershd_up_status -EXPECT_WITHIN 20 "1" afr_child_up_status_in_shd $V0 0 -EXPECT_WITHIN 20 "1" afr_child_up_status_in_shd $V0 1 +EXPECT_WITHIN $CHILD_UP_TIMEOUT "1" afr_child_up_status $V0 0 +EXPECT_WITHIN $PROCESS_UP_TIMEOUT "Y" glustershd_up_status +EXPECT_WITHIN $CHILD_UP_TIMEOUT "1" afr_child_up_status_in_shd $V0 0 +EXPECT_WITHIN $CHILD_UP_TIMEOUT "1" afr_child_up_status_in_shd $V0 1  TEST $CLI volume heal $V0 -EXPECT_WITHIN 20 "0" afr_get_pending_heal_count $V0 +EXPECT_WITHIN $HEAL_TIMEOUT "0" afr_get_pending_heal_count $V0  #check heal has happened in the correct direction  EXPECT "$NEW_UID$NEW_GID" stat --printf=%u%g $B0/brick0/file @@ -138,20 +138,20 @@ TEST `echo "write1">$M0/file`  TEST kill_brick $V0 $H0 $B0/brick0  TEST `echo "write2">>$M0/file`  TEST $CLI volume start $V0 force -EXPECT_WITHIN 20 "1" afr_child_up_status $V0 0 -EXPECT_WITHIN 20 "Y" glustershd_up_status -EXPECT_WITHIN 20 "1" afr_child_up_status_in_shd $V0 0 -EXPECT_WITHIN 20 "1" afr_child_up_status_in_shd $V0 1 +EXPECT_WITHIN $CHILD_UP_TIMEOUT "1" afr_child_up_status $V0 0 +EXPECT_WITHIN $PROCESS_UP_TIMEOUT "Y" glustershd_up_status +EXPECT_WITHIN $CHILD_UP_TIMEOUT "1" afr_child_up_status_in_shd $V0 0 +EXPECT_WITHIN $CHILD_UP_TIMEOUT "1" afr_child_up_status_in_shd $V0 1  TEST $CLI volume heal $V0  TEST kill_brick $V0 $H0 $B0/brick1  TEST truncate -s 0 $M0/file  TEST $CLI volume start $V0 force -EXPECT_WITHIN 20 "1" afr_child_up_status $V0 1 -EXPECT_WITHIN 20 "Y" glustershd_up_status -EXPECT_WITHIN 20 "1" afr_child_up_status_in_shd $V0 0 -EXPECT_WITHIN 20 "1" afr_child_up_status_in_shd $V0 1 +EXPECT_WITHIN $CHILD_UP_TIMEOUT "1" afr_child_up_status $V0 1 +EXPECT_WITHIN $PROCESS_UP_TIMEOUT "Y" glustershd_up_status +EXPECT_WITHIN $CHILD_UP_TIMEOUT "1" afr_child_up_status_in_shd $V0 0 +EXPECT_WITHIN $CHILD_UP_TIMEOUT "1" afr_child_up_status_in_shd $V0 1  TEST $CLI volume heal $V0 -EXPECT_WITHIN 20 "0" afr_get_pending_heal_count $V0 +EXPECT_WITHIN $HEAL_TIMEOUT "0" afr_get_pending_heal_count $V0  #check heal has happened in the correct direction  EXPECT 0 stat --printf=%s $B0/brick1/file @@ -170,11 +170,11 @@ TEST rm -f $M0/file  TEST touch $M0/file  GFID=$(gf_get_gfid_xattr $B1/brick1/file)  TEST $CLI volume start $V0 force -EXPECT_WITHIN 20 "Y" glustershd_up_status -EXPECT_WITHIN 20 "1" afr_child_up_status_in_shd $V0 0 -EXPECT_WITHIN 20 "1" afr_child_up_status_in_shd $V0 1 +EXPECT_WITHIN $PROCESS_UP_TIMEOUT "Y" glustershd_up_status +EXPECT_WITHIN $CHILD_UP_TIMEOUT "1" afr_child_up_status_in_shd $V0 0 +EXPECT_WITHIN $CHILD_UP_TIMEOUT "1" afr_child_up_status_in_shd $V0 1  TEST $CLI volume heal $V0 -EXPECT_WITHIN 20 "0" afr_get_pending_heal_count $V0 +EXPECT_WITHIN $HEAL_TIMEOUT "0" afr_get_pending_heal_count $V0  #check heal has happened in the correct direction  EXPECT "$GFID" gf_get_gfid_xattr $B0/brick0/file @@ -193,12 +193,12 @@ TEST rm -f $M0/link_to_file  TEST ln -s $M0/file $M0/link_to_file  TEST ln  $M0/file $M0/hard_link_to_file  TEST $CLI volume start $V0 force -EXPECT_WITHIN 20 "1" afr_child_up_status $V0 0 -EXPECT_WITHIN 20 "Y" glustershd_up_status -EXPECT_WITHIN 20 "1" afr_child_up_status_in_shd $V0 0 -EXPECT_WITHIN 20 "1" afr_child_up_status_in_shd $V0 1 +EXPECT_WITHIN $CHILD_UP_TIMEOUT "1" afr_child_up_status $V0 0 +EXPECT_WITHIN $PROCESS_UP_TIMEOUT "Y" glustershd_up_status +EXPECT_WITHIN $CHILD_UP_TIMEOUT "1" afr_child_up_status_in_shd $V0 0 +EXPECT_WITHIN $CHILD_UP_TIMEOUT "1" afr_child_up_status_in_shd $V0 1  TEST $CLI volume heal $V0 -EXPECT_WITHIN 20 "0" afr_get_pending_heal_count $V0 +EXPECT_WITHIN $HEAL_TIMEOUT "0" afr_get_pending_heal_count $V0  #check heal has happened in the correct direction  TEST test -f $B0/brick0/hard_link_to_file @@ -219,12 +219,12 @@ TEST kill_brick $V0 $H0 $B0/brick0  TEST setfattr -n user.myattr_1 -v "My_attribute_1_modified" $M0/file  TEST setfattr -n user.myattr_3 -v "My_attribute_3" $M0/file  TEST $CLI volume start $V0 force -EXPECT_WITHIN 20 "1" afr_child_up_status $V0 0 -EXPECT_WITHIN 20 "Y" glustershd_up_status -EXPECT_WITHIN 20 "1" afr_child_up_status_in_shd $V0 0 -EXPECT_WITHIN 20 "1" afr_child_up_status_in_shd $V0 1 +EXPECT_WITHIN $CHILD_UP_TIMEOUT "1" afr_child_up_status $V0 0 +EXPECT_WITHIN $PROCESS_UP_TIMEOUT "Y" glustershd_up_status +EXPECT_WITHIN $CHILD_UP_TIMEOUT "1" afr_child_up_status_in_shd $V0 0 +EXPECT_WITHIN $CHILD_UP_TIMEOUT "1" afr_child_up_status_in_shd $V0 1  TEST $CLI volume heal $V0 -EXPECT_WITHIN 20 "0" afr_get_pending_heal_count $V0 +EXPECT_WITHIN $HEAL_TIMEOUT "0" afr_get_pending_heal_count $V0  TEST diff <(echo "user.myattr_1=\"My_attribute_1_modified\"") <(getfattr -n user.myattr_1 $B0/brick1/file|grep user.myattr_1)  TEST diff <(echo "user.myattr_3=\"My_attribute_3\"") <(getfattr -n user.myattr_3 $B0/brick1/file|grep user.myattr_3) diff --git a/tests/basic/afr/sparse-file-self-heal.t b/tests/basic/afr/sparse-file-self-heal.t index 9b795c331cf..01b676ea900 100644 --- a/tests/basic/afr/sparse-file-self-heal.t +++ b/tests/basic/afr/sparse-file-self-heal.t @@ -39,12 +39,12 @@ TEST truncate -s 2M $M0/big2bigger  big2bigger_md5sum=$(md5sum $M0/big2bigger | awk '{print $1}')  $CLI volume start $V0 force -EXPECT_WITHIN 20 "1" afr_child_up_status $V0 0 -EXPECT_WITHIN 20 "Y" glustershd_up_status -EXPECT_WITHIN 20 "1" afr_child_up_status_in_shd $V0 0 -EXPECT_WITHIN 20 "1" afr_child_up_status_in_shd $V0 1 +EXPECT_WITHIN $CHILD_UP_TIMEOUT "1" afr_child_up_status $V0 0 +EXPECT_WITHIN $PROCESS_UP_TIMEOUT "Y" glustershd_up_status +EXPECT_WITHIN $CHILD_UP_TIMEOUT "1" afr_child_up_status_in_shd $V0 0 +EXPECT_WITHIN $CHILD_UP_TIMEOUT "1" afr_child_up_status_in_shd $V0 1  TEST gluster volume heal $V0 full -EXPECT_WITHIN 20 "0" afr_get_pending_heal_count $V0 +EXPECT_WITHIN $HEAL_TIMEOUT "0" afr_get_pending_heal_count $V0  big_md5sum_0=$(md5sum $B0/${V0}0/big | awk '{print $1}')  small_md5sum_0=$(md5sum $B0/${V0}0/small | awk '{print $1}') @@ -96,12 +96,12 @@ TEST truncate -s 2M $M0/big2bigger  big2bigger_md5sum=$(md5sum $M0/big2bigger | awk '{print $1}')  $CLI volume start $V0 force -EXPECT_WITHIN 20 "1" afr_child_up_status $V0 0 -EXPECT_WITHIN 20 "Y" glustershd_up_status -EXPECT_WITHIN 20 "1" afr_child_up_status_in_shd $V0 0 -EXPECT_WITHIN 20 "1" afr_child_up_status_in_shd $V0 1 +EXPECT_WITHIN $CHILD_UP_TIMEOUT "1" afr_child_up_status $V0 0 +EXPECT_WITHIN $PROCESS_UP_TIMEOUT "Y" glustershd_up_status +EXPECT_WITHIN $CHILD_UP_TIMEOUT "1" afr_child_up_status_in_shd $V0 0 +EXPECT_WITHIN $CHILD_UP_TIMEOUT "1" afr_child_up_status_in_shd $V0 1  TEST gluster volume heal $V0 full -EXPECT_WITHIN 20 "0" afr_get_pending_heal_count $V0 +EXPECT_WITHIN $HEAL_TIMEOUT "0" afr_get_pending_heal_count $V0  big_md5sum_0=$(md5sum $B0/${V0}0/big | awk '{print $1}')  small_md5sum_0=$(md5sum $B0/${V0}0/small | awk '{print $1}') diff --git a/tests/basic/afr/stale-file-lookup.t b/tests/basic/afr/stale-file-lookup.t index 24a478d5c4a..f2ab560cb28 100644 --- a/tests/basic/afr/stale-file-lookup.t +++ b/tests/basic/afr/stale-file-lookup.t @@ -22,7 +22,7 @@ TEST touch $M0/a  TEST kill_brick $V0 $H0 $B0/${V0}0  TEST rm -f $M0/a  TEST $CLI volume start $V0 force -EXPECT_WITHIN 20 "1" afr_child_up_status $V0 0 +EXPECT_WITHIN $CHILD_UP_TIMEOUT "1" afr_child_up_status $V0 0  TEST stat $B0/${V0}0/a  TEST ! stat $B0/${V0}1/a  TEST ! ls -l $M0/a diff --git a/tests/basic/mgmt_v3-locks.t b/tests/basic/mgmt_v3-locks.t index 22ca27b9f20..4c259caa874 100644 --- a/tests/basic/mgmt_v3-locks.t +++ b/tests/basic/mgmt_v3-locks.t @@ -79,7 +79,7 @@ TEST launch_cluster 3;  TEST $CLI_1 peer probe $H2;  TEST $CLI_1 peer probe $H3; -EXPECT_WITHIN 20 2 check_peers +EXPECT_WITHIN $PROBE_TIMEOUT 2 check_peers  two_diff_vols_create  EXPECT 'Created' volinfo_field $V0 'Status'; @@ -92,12 +92,12 @@ EXPECT 'Started' volinfo_field $V1 'Status';  same_vol_remove_brick $V0 $H2:$B2/$V0  # Checking glusterd crashed or not after same volume remove brick  # on both nodes. -EXPECT_WITHIN 20 2 check_peers +EXPECT_WITHIN $PROBE_TIMEOUT 2 check_peers  same_vol_remove_brick $V1 $H2:$B2/$V1  # Checking glusterd crashed or not after same volume remove brick  # on both nodes. -EXPECT_WITHIN 20 2 check_peers +EXPECT_WITHIN $PROBE_TIMEOUT 2 check_peers  $CLI_1 volume set $V0 diagnostics.client-log-level DEBUG &  $CLI_1 volume set $V1 diagnostics.client-log-level DEBUG @@ -105,7 +105,7 @@ kill_glusterd 3  $CLI_1 volume status $V0  $CLI_2 volume status $V1  $CLI_1 peer status -EXPECT_WITHIN 20 1 check_peers +EXPECT_WITHIN $PROBE_TIMEOUT 1 check_peers  EXPECT 'Started' volinfo_field $V0 'Status';  EXPECT 'Started' volinfo_field $V1 'Status'; @@ -113,9 +113,9 @@ TEST $glusterd_3  $CLI_1 volume status $V0  $CLI_2 volume status $V1  $CLI_1 peer status -#EXPECT_WITHIN 20 2 check_peers +#EXPECT_WITHIN $PROBE_TIMEOUT 2 check_peers  #EXPECT 'Started' volinfo_field $V0 'Status';  #EXPECT 'Started' volinfo_field $V1 'Status';  #two_diff_vols_stop_force -#EXPECT_WITHIN 20 2 check_peers +#EXPECT_WITHIN $PROBE_TIMEOUT 2 check_peers  cleanup; diff --git a/tests/basic/mount.t b/tests/basic/mount.t index 8163975d679..c77317f37f5 100755 --- a/tests/basic/mount.t +++ b/tests/basic/mount.t @@ -50,7 +50,7 @@ TEST 'mount -t fuse.glusterfs | grep -E "^$H0:$V0 .+ \(ro,"';  TEST 'grep -E "^$H0:$V0 .+ ,?ro,.+" /proc/mounts';  ## Wait for volume to register with rpc.mountd -EXPECT_WITHIN 20 "1" is_nfs_export_available; +EXPECT_WITHIN $NFS_EXPORT_TIMEOUT "1" is_nfs_export_available;  ## Mount NFS  TEST mount -t nfs -o nolock,soft,intr $H0:/$V0 $N0; diff --git a/tests/basic/quota.t b/tests/basic/quota.t index cfc4f06950d..d92b382c780 100755 --- a/tests/basic/quota.t +++ b/tests/basic/quota.t @@ -74,20 +74,20 @@ TEST ! dd if=/dev/urandom of=$M0/test_dir/1.txt bs=1M count=12  TEST rm $M0/test_dir/1.txt  # wait for marker's accounting to complete -EXPECT_WITHIN 10 "0Bytes" usage "/test_dir" +EXPECT_WITHIN $MARKER_UPDATE_TIMEOUT "0Bytes" usage "/test_dir"  TEST dd if=/dev/urandom of=$M0/test_dir/2.txt bs=1M count=8 -EXPECT_WITHIN 20 "8.0MB" usage "/test_dir" +EXPECT_WITHIN $MARKER_UPDATE_TIMEOUT "8.0MB" usage "/test_dir"  TEST rm $M0/test_dir/2.txt -EXPECT_WITHIN 10 "0Bytes" usage "/test_dir" +EXPECT_WITHIN $MARKER_UPDATE_TIMEOUT "0Bytes" usage "/test_dir"  ## rename tests  TEST dd if=/dev/urandom of=$M0/test_dir/2 bs=1M count=8 -EXPECT_WITHIN 20 "8.0MB" usage "/test_dir" +EXPECT_WITHIN $MARKER_UPDATE_TIMEOUT "8.0MB" usage "/test_dir"  TEST mv $M0/test_dir/2 $M0/test_dir/0 -EXPECT_WITHIN 10 "8.0MB" usage "/test_dir" +EXPECT_WITHIN $MARKER_UPDATE_TIMEOUT "8.0MB" usage "/test_dir"  TEST rm $M0/test_dir/0 -EXPECT_WITHIN 10 "0Bytes" usage "/test_dir" +EXPECT_WITHIN $MARKER_UPDATE_TIMEOUT "0Bytes" usage "/test_dir"  ## --------------------------- diff --git a/tests/basic/self-heald.t b/tests/basic/self-heald.t index 4468c881bac..913cfc80599 100644 --- a/tests/basic/self-heald.t +++ b/tests/basic/self-heald.t @@ -34,15 +34,11 @@ TEST ! $CLI volume heal $V0 info  TEST ! $CLI volume heal $V0  TEST $CLI volume start $V0 force  TEST $CLI volume set $V0 cluster.self-heal-daemon on -EXPECT_WITHIN 20 "Y" glustershd_up_status -EXPECT_WITHIN 20 "1" afr_child_up_status_in_shd $V0 0 -EXPECT_WITHIN 20 "1" afr_child_up_status_in_shd $V0 2 -EXPECT_WITHIN 20 "1" afr_child_up_status_in_shd $V0 4 -TEST $CLI volume heal $V0 -sleep 5 #Until the heal-statistics command implementation -#check that this heals the contents partially -TEST [ $HEAL_FILES -gt $(afr_get_pending_heal_count $V0) ] +EXPECT_WITHIN $PROCESS_UP_TIMEOUT "Y" glustershd_up_status +EXPECT_WITHIN $CHILD_UP_TIMEOUT "1" afr_child_up_status_in_shd $V0 0 +EXPECT_WITHIN $CHILD_UP_TIMEOUT "1" afr_child_up_status_in_shd $V0 2 +EXPECT_WITHIN $CHILD_UP_TIMEOUT "1" afr_child_up_status_in_shd $V0 4  TEST $CLI volume heal $V0 full -EXPECT_WITHIN 30 "0" afr_get_pending_heal_count $V0 +EXPECT_WITHIN $HEAL_TIMEOUT "0" afr_get_pending_heal_count $V0  cleanup diff --git a/tests/basic/volume-snapshot.t b/tests/basic/volume-snapshot.t index 30dfbbca195..21d6ea3b939 100755 --- a/tests/basic/volume-snapshot.t +++ b/tests/basic/volume-snapshot.t @@ -76,7 +76,7 @@ TEST setup_lvm 3  TEST $CLI_1 peer probe $H2;  TEST $CLI_1 peer probe $H3; -EXPECT_WITHIN 20 2 peer_count; +EXPECT_WITHIN $PROBE_TIMEOUT 2 peer_count;  create_volumes  EXPECT 'Created' volinfo_field $V0 'Status'; diff --git a/tests/basic/volume-status.t b/tests/basic/volume-status.t index f4196ac30f0..a91db3cc50a 100644 --- a/tests/basic/volume-status.t +++ b/tests/basic/volume-status.t @@ -24,8 +24,8 @@ TEST mount -t nfs -o vers=3,nolock,soft,intr $H0:/$V0 $N0;  TEST $CLI volume status all  TEST $CLI volume status $V0 -EXPECT_WITHIN 10 'Y' nfs_up_status -EXPECT_WITHIN 10 'Y' glustershd_up_status +EXPECT_WITHIN $PROCESS_UP_TIMEOUT 'Y' nfs_up_status +EXPECT_WITHIN $PROCESS_UP_TIMEOUT 'Y' glustershd_up_status  function test_nfs_cmds () {      local ret=0      declare -a nfs_cmds=("clients" "mem" "inode" "callpool") diff --git a/tests/bugs/859927/repl.t b/tests/bugs/859927/repl.t index 856b057fbcb..9ac524fc1bd 100755 --- a/tests/bugs/859927/repl.t +++ b/tests/bugs/859927/repl.t @@ -32,20 +32,20 @@ touch $M0/a  TEST $CLI volume set $V0 cluster.data-self-heal-algorithm full  EXPECT full volume_option $V0 cluster.data-self-heal-algorithm  create_setup_for_self_heal $M0/a -EXPECT_WITHIN 20 "1" afr_child_up_status $V0 0 +EXPECT_WITHIN $CHILD_UP_TIMEOUT "1" afr_child_up_status $V0 0  cat $file 2>&1 > /dev/null  TEST cmp $B0/${V0}1/a $B0/${V0}2/a  TEST $CLI volume set $V0 cluster.data-self-heal-algorithm diff  EXPECT diff volume_option $V0 cluster.data-self-heal-algorithm  create_setup_for_self_heal $M0/a -EXPECT_WITHIN 20 "1" afr_child_up_status $V0 0 +EXPECT_WITHIN $CHILD_UP_TIMEOUT "1" afr_child_up_status $V0 0  cat $file 2>&1 > /dev/null  TEST cmp $B0/${V0}1/a $B0/${V0}2/a  TEST $CLI volume reset $V0 cluster.data-self-heal-algorithm  create_setup_for_self_heal $M0/a -EXPECT_WITHIN 20 "1" afr_child_up_status $V0 0 +EXPECT_WITHIN $CHILD_UP_TIMEOUT "1" afr_child_up_status $V0 0  cat $file 2>&1 > /dev/null  TEST cmp $B0/${V0}1/a $B0/${V0}2/a diff --git a/tests/bugs/886998/strict-readdir.t b/tests/bugs/886998/strict-readdir.t index 0de953e8a52..57a8c1c32dc 100644 --- a/tests/bugs/886998/strict-readdir.t +++ b/tests/bugs/886998/strict-readdir.t @@ -25,16 +25,16 @@ TEST kill_brick $V0 $H0 $B0/r2d2_2  EXPECT "100" num_files_in_dir $M0  TEST $CLI volume start $V0 force -EXPECT_WITHIN 20 "1" afr_child_up_status $V0 0 -EXPECT_WITHIN 20 "1" afr_child_up_status $V0 2 +EXPECT_WITHIN $CHILD_UP_TIMEOUT "1" afr_child_up_status $V0 0 +EXPECT_WITHIN $CHILD_UP_TIMEOUT "1" afr_child_up_status $V0 2  TEST kill_brick $V0 $H0 $B0/r2d2_1  TEST kill_brick $V0 $H0 $B0/r2d2_3  EXPECT "100" num_files_in_dir $M0  TEST $CLI volume start $V0 force -EXPECT_WITHIN 20 "1" afr_child_up_status $V0 1 -EXPECT_WITHIN 20 "1" afr_child_up_status $V0 3 +EXPECT_WITHIN $CHILD_UP_TIMEOUT "1" afr_child_up_status $V0 1 +EXPECT_WITHIN $CHILD_UP_TIMEOUT "1" afr_child_up_status $V0 3  TEST $CLI volume set $V0 cluster.strict-readdir on  EXPECT "on" volinfo_field $V0 cluster.strict-readdir @@ -43,8 +43,8 @@ TEST kill_brick $V0 $H0 $B0/r2d2_2  EXPECT "100" num_files_in_dir $M0  TEST $CLI volume start $V0 force -EXPECT_WITHIN 20 "1" afr_child_up_status $V0 0 -EXPECT_WITHIN 20 "1" afr_child_up_status $V0 2 +EXPECT_WITHIN $CHILD_UP_TIMEOUT "1" afr_child_up_status $V0 0 +EXPECT_WITHIN $CHILD_UP_TIMEOUT "1" afr_child_up_status $V0 2  TEST kill_brick $V0 $H0 $B0/r2d2_1  TEST kill_brick $V0 $H0 $B0/r2d2_3 diff --git a/tests/bugs/bug-1004744.t b/tests/bugs/bug-1004744.t index 1211002e299..089cb35a184 100644 --- a/tests/bugs/bug-1004744.t +++ b/tests/bugs/bug-1004744.t @@ -36,9 +36,7 @@ TEST $CLI volume add-brick $V0 $H0:$B0/${V0}{3,4};  #perform rebalance fix-layout  TEST $CLI volume rebalance $V0 fix-layout start -EXPECT_WITHIN 1 "fix-layout in progress" rebalance_status_field $V0; - -EXPECT_WITHIN 30 "fix-layout completed" rebalance_status_field $V0; +EXPECT_WITHIN $REBALANCE_TIMEOUT "fix-layout completed" rebalance_status_field $V0;  TEST umount $M0  TEST $CLI volume stop $V0 diff --git a/tests/bugs/bug-1022055.t b/tests/bugs/bug-1022055.t index c2f4218bb20..07d0b1f2a6f 100755 --- a/tests/bugs/bug-1022055.t +++ b/tests/bugs/bug-1022055.t @@ -13,7 +13,7 @@ TEST launch_cluster 2;  TEST $CLI_1 peer probe $H2; -EXPECT_WITHIN 20 1 check_peers; +EXPECT_WITHIN $PROBE_TIMEOUT 1 check_peers;  TEST $CLI_1 volume create $V0 $H1:$B1/$V0 $H2:$B2/$V0; diff --git a/tests/bugs/bug-1032894.t b/tests/bugs/bug-1032894.t index 2a2b77fece9..ecb5952d860 100644 --- a/tests/bugs/bug-1032894.t +++ b/tests/bugs/bug-1032894.t @@ -21,13 +21,13 @@ for i in {1..10}; do echo abc > $i; done  for i in {1..10}; do rm -f $i; done  TEST $CLI volume start $V0 force -EXPECT_WITHIN 20 "1" afr_child_up_status_in_shd $V0 0 +EXPECT_WITHIN $CHILD_UP_TIMEOUT "1" afr_child_up_status_in_shd $V0 0  TEST $CLI volume set $V0 cluster.self-heal-daemon on -EXPECT_WITHIN 20 "Y" glustershd_up_status +EXPECT_WITHIN $PROCESS_UP_TIMEOUT "Y" glustershd_up_status  #Since maximum depth of the directory structure that needs healin is 2  #Trigger two self-heals. That should make sure the heal is complete  TEST $CLI volume heal $V0 -EXPECT_WITHIN 20 "0"  afr_get_index_count $B0/${V0}1 +EXPECT_WITHIN $HEAL_TIMEOUT "0"  afr_get_index_count $B0/${V0}1  cleanup diff --git a/tests/bugs/bug-1035576.t b/tests/bugs/bug-1035576.t index 938306a8503..62d431a703a 100644 --- a/tests/bugs/bug-1035576.t +++ b/tests/bugs/bug-1035576.t @@ -29,7 +29,7 @@ TEST mkdir $M0/a  TEST $CLI volume quota $V0 limit-usage /a 1GB  echo abc > $M0/a/f  $CLI volume start $V0 force -EXPECT_WITHIN 20 "1" afr_child_up_status $V0 0 +EXPECT_WITHIN $CHILD_UP_TIMEOUT "1" afr_child_up_status $V0 0  quota_limit_val1=$(get_hex_xattr trusted.glusterfs.quota.limit-set $B0/${V0}1/a)  quota_size_val1=$(get_hex_xattr trusted.glusterfs.quota.size $B0/${V0}1/a) diff --git a/tests/bugs/bug-1043886.t b/tests/bugs/bug-1043886.t index b9ee320ae96..edebdc2cd2e 100755 --- a/tests/bugs/bug-1043886.t +++ b/tests/bugs/bug-1043886.t @@ -14,7 +14,7 @@ sleep 2;  ## Mount FUSE with caching disabled  TEST glusterfs --entry-timeout=0 --attribute-timeout=0 -s $H0 --volfile-id $V0 $M0; -EXPECT_WITHIN 20 "1" is_nfs_export_available; +EXPECT_WITHIN $NFS_EXPORT_TIMEOUT "1" is_nfs_export_available;  ## Mount volume as NFS export  TEST mount_nfs $H0:/$V0 $N0 nolock; @@ -32,7 +32,7 @@ TEST $CLI volume set $V0 server.anongid $gid;  sleep 2; -EXPECT_WITHIN 20 "1" is_nfs_export_available; +EXPECT_WITHIN $NFS_EXPORT_TIMEOUT "1" is_nfs_export_available;  # create files and directories in the root of the glusterfs and nfs mount  # which is owned by root and hence the right behavior is getting EACCESS diff --git a/tests/bugs/bug-1047955.t b/tests/bugs/bug-1047955.t index e15f3ceef00..169333e9eb4 100644 --- a/tests/bugs/bug-1047955.t +++ b/tests/bugs/bug-1047955.t @@ -17,7 +17,7 @@ TEST launch_cluster 2;  TEST $CLI_1 volume create $V0 replica 2 $H1:$B1/${V0}{1,2,3,4}  TEST $CLI_1 volume start $V0;  TEST $CLI_1 peer probe $H2; -EXPECT_WITHIN 20 1 check_peers; +EXPECT_WITHIN $PROBE_TIMEOUT 1 check_peers;  TEST $CLI_2 volume remove-brick $V0 $H1:$B1/${V0}{3,4} start;  TEST $CLI_2 volume info  cleanup; diff --git a/tests/bugs/bug-1049834.t b/tests/bugs/bug-1049834.t index eb9a7fbe875..7e16fde61aa 100755 --- a/tests/bugs/bug-1049834.t +++ b/tests/bugs/bug-1049834.t @@ -11,7 +11,7 @@ TEST launch_cluster 2  TEST setup_lvm 2  TEST $CLI_1 peer probe $H2 -EXPECT_WITHIN 20 1 peer_count +EXPECT_WITHIN $PROBE_TIMEOUT 1 peer_count  TEST $CLI_1 volume create $V0 $H1:$L1 $H2:$L2  EXPECT 'Created' volinfo_field $V0 'Status' diff --git a/tests/bugs/bug-1053579.t b/tests/bugs/bug-1053579.t index afbc4c9579a..7a8ea192adc 100755 --- a/tests/bugs/bug-1053579.t +++ b/tests/bugs/bug-1053579.t @@ -28,7 +28,7 @@ TEST $CLI volume create $V0 $H0:$B0/${V0}1  TEST $CLI volume set $V0 nfs.server-aux-gids on  TEST $CLI volume start $V0 -EXPECT_WITHIN 20 "1" is_nfs_export_available +EXPECT_WITHIN $NFS_EXPORT_TIMEOUT "1" is_nfs_export_available  # mount the volume  TEST mount_nfs $H0:/$V0 $N0 nolock diff --git a/tests/bugs/bug-1058797.t b/tests/bugs/bug-1058797.t index 1e9f09af0a8..c48f19b9aa8 100644 --- a/tests/bugs/bug-1058797.t +++ b/tests/bugs/bug-1058797.t @@ -28,7 +28,7 @@ EXPECT "s" echo $setuid_bit1  #Restart volume and do lookup from mount to trigger heal  TEST $CLI volume start $V0 force -EXPECT_WITHIN 20 "1" afr_child_up_status $V0 1 +EXPECT_WITHIN $CHILD_UP_TIMEOUT "1" afr_child_up_status $V0 1  TEST dd if=$M0/file of=/dev/null  #Get file permissions from healed brick1 and verify that S_ISUID is indeed set diff --git a/tests/bugs/bug-1064768.t b/tests/bugs/bug-1064768.t index b87168150d2..b0d04eb25ec 100644 --- a/tests/bugs/bug-1064768.t +++ b/tests/bugs/bug-1064768.t @@ -8,7 +8,7 @@ TEST glusterd  TEST pidof glusterd  TEST $CLI volume create $V0 replica 2 $H0:$B0/brick0 $H0:$B0/brick1  TEST $CLI volume start $V0 -EXPECT_WITHIN 15 'Started' volinfo_field $V0 'Status'; +EXPECT_WITHIN $PROCESS_UP_TIMEOUT 'Started' volinfo_field $V0 'Status';  TEST $CLI volume profile $V0 start  TEST $CLI volume profile $V0 info @@ -16,5 +16,5 @@ TEST $CLI volume profile $V0 stop  TEST $CLI volume status  TEST $CLI volume stop $V0 -EXPECT_WITHIN 15 'Stopped' volinfo_field $V0 'Status'; +EXPECT_WITHIN $PROCESS_UP_TIMEOUT 'Stopped' volinfo_field $V0 'Status';  cleanup; diff --git a/tests/bugs/bug-1066798.t b/tests/bugs/bug-1066798.t index 635b143f05a..445ec75c936 100755 --- a/tests/bugs/bug-1066798.t +++ b/tests/bugs/bug-1066798.t @@ -56,7 +56,7 @@ done  TEST $CLI volume remove-brick $V0 $H0:${HASHED} start -EXPECT_WITHIN 20 "completed" remove_brick_status_completed_field "$V0" "$H0:${HASHED}"; +EXPECT_WITHIN $REBALANCE_TIMEOUT "completed" remove_brick_status_completed_field "$V0" "$H0:${HASHED}";  #check consistency in mount point  #And also check all the links are migrated to OTHER diff --git a/tests/bugs/bug-1070734.t b/tests/bugs/bug-1070734.t index f35189eb871..fd7bbe01cf6 100755 --- a/tests/bugs/bug-1070734.t +++ b/tests/bugs/bug-1070734.t @@ -22,7 +22,7 @@ EXPECT 'Created' volinfo_field $V0 'Status';  TEST $CLI volume start $V0;  EXPECT 'Started' volinfo_field $V0 'Status'; -EXPECT_WITHIN 20 "1" is_nfs_export_available; +EXPECT_WITHIN $NFS_EXPORT_TIMEOUT "1" is_nfs_export_available;  TEST mount_nfs $H0:/$V0 $N0;  ############################################################################ @@ -54,7 +54,7 @@ fi  TEST rm -f $N0/DIR/file;  TEST rmdir $N0/DIR;  TEST $CLI volume remove-brick $V0 $H0:${HASHED} start; -EXPECT_WITHIN 10 "completed" remove_brick_status_completed_field "$V0" \ +EXPECT_WITHIN $REBALANCE_TIMEOUT "completed" remove_brick_status_completed_field "$V0" \  "$H0:${HASHED}";  TEST mkdir $N0/DIR; diff --git a/tests/bugs/bug-1077682.t b/tests/bugs/bug-1077682.t index 2923c5f66dd..b47744e603c 100644 --- a/tests/bugs/bug-1077682.t +++ b/tests/bugs/bug-1077682.t @@ -24,7 +24,7 @@ TEST ! $CLI volume remove-brick $V0 $H0:$B0/${V0}1  TEST $CLI volume remove-brick $V0 $H0:$B0/${V0}2 force  TEST $CLI volume remove-brick $V0 $H0:$B0/${V0}3 start -EXPECT_WITHIN 10 "completed" remove_brick_status_completed_field "$V0" \ +EXPECT_WITHIN $REBALANCE_TIMEOUT "completed" remove_brick_status_completed_field "$V0" \  "$H0:$B0/${V0}3"  TEST $CLI volume remove-brick $V0 $H0:$B0/${V0}3 commit diff --git a/tests/bugs/bug-1087198.t b/tests/bugs/bug-1087198.t index 29a81c3abbb..6d3802ed088 100644 --- a/tests/bugs/bug-1087198.t +++ b/tests/bugs/bug-1087198.t @@ -30,7 +30,7 @@ EXPECT 'Created' volinfo_field $V0 'Status';  TEST $CLI volume start $V0;  EXPECT 'Started' volinfo_field $V0 'Status'; -EXPECT_WITHIN 20 "1" is_nfs_export_available; +EXPECT_WITHIN $NFS_EXPORT_TIMEOUT "1" is_nfs_export_available;  TEST mount_nfs $H0:/$V0 $N0 noac,nolock diff --git a/tests/bugs/bug-765473.t b/tests/bugs/bug-765473.t index 5fc0ec9d77e..d9af751f472 100755 --- a/tests/bugs/bug-765473.t +++ b/tests/bugs/bug-765473.t @@ -11,7 +11,7 @@ function clients_connected()      volname=$1      gluster volume status $volname clients | grep -i 'Clients connected' | sed -e 's/[^0-9]*\(.*\)/\1/g'  } -         +  ## Start and create a volume  TEST glusterd;  TEST pidof glusterd; @@ -27,7 +27,7 @@ TEST $CLI volume stop $V0  # write some content which will result in marking fd bad  fd_write $fd "more content"  TEST $CLI volume start $V0 -EXPECT_WITHIN 30 2 clients_connected $V0 +EXPECT_WITHIN $PROCESS_UP_TIMEOUT 2 clients_connected $V0  TEST ! fd_write $fd "still more content"  cleanup diff --git a/tests/bugs/bug-765564.t b/tests/bugs/bug-765564.t index 6e4087f805c..b21bade9478 100644 --- a/tests/bugs/bug-765564.t +++ b/tests/bugs/bug-765564.t @@ -76,7 +76,7 @@ TEST mv $M0/h $M0/1;  TEST $CLI volume start $V0 force; -EXPECT_WITHIN 20 "1" afr_child_up_status $V0 1; +EXPECT_WITHIN $CHILD_UP_TIMEOUT "1" afr_child_up_status $V0 1;  find $M0 | xargs stat 2>/dev/null 1>/dev/null;  TEST rm_mv_correctness; diff --git a/tests/bugs/bug-802417.t b/tests/bugs/bug-802417.t index b596df30385..3a6db22edbf 100755 --- a/tests/bugs/bug-802417.t +++ b/tests/bugs/bug-802417.t @@ -51,9 +51,9 @@ TEST write_file $M0/a_file "new_data"  ## Bring all the bricks up and kill one so we do a partial self-heal.  TEST $CLI volume start $V0 force -EXPECT_WITHIN 20 "1" afr_child_up_status $V0 0 -EXPECT_WITHIN 20 "1" afr_child_up_status $V0 1 -EXPECT_WITHIN 20 "1" afr_child_up_status $V0 2 +EXPECT_WITHIN $CHILD_UP_TIMEOUT "1" afr_child_up_status $V0 0 +EXPECT_WITHIN $CHILD_UP_TIMEOUT "1" afr_child_up_status $V0 1 +EXPECT_WITHIN $CHILD_UP_TIMEOUT "1" afr_child_up_status $V0 2  TEST kill_brick ${V0} ${H0} ${B0}/${V0}-2  TEST dd if=${M0}/a_file of=/dev/null diff --git a/tests/bugs/bug-821056.t b/tests/bugs/bug-821056.t index 5e81541ac3a..db87993ed76 100644 --- a/tests/bugs/bug-821056.t +++ b/tests/bugs/bug-821056.t @@ -27,14 +27,14 @@ realpath=$(gf_get_gfid_backend_file_path $B0/${V0}0 "a")  kill_brick $V0 $H0 $B0/${V0}0  TEST $CLI volume start $V0 force -EXPECT_WITHIN 20 "1" afr_child_up_status $V0 0 +EXPECT_WITHIN $CHILD_UP_TIMEOUT "1" afr_child_up_status $V0 0  EXPECT "Y" gf_check_file_opened_in_brick $V0 $H0 $B0/${V0}0 "$realpath"  kill_brick $V0 $H0 $B0/${V0}0  TEST gf_rm_file_and_gfid_link $B0/${V0}0 "a"  TEST $CLI volume start $V0 force -EXPECT_WITHIN 20 "1" afr_child_up_status $V0 0 +EXPECT_WITHIN $CHILD_UP_TIMEOUT "1" afr_child_up_status $V0 0  ls -l $M0/a 2>&1 > /dev/null #Make sure the file is re-created  EXPECT "N" gf_check_file_opened_in_brick $V0 $H0 $B0/${V0}0 "$realpath"  EXPECT "N" gf_check_file_opened_in_brick $V0 $H0 $B0/${V0}0 $B0/${V0}0/a @@ -43,10 +43,10 @@ for i in {1..1024}; do          echo "open sesame" >&5  done -EXPECT_WITHIN 20 "Y" gf_check_file_opened_in_brick $V0 $H0 $B0/${V0}0 $B0/${V0}0/a +EXPECT_WITHIN $REOPEN_TIMEOUT "Y" gf_check_file_opened_in_brick $V0 $H0 $B0/${V0}0 $B0/${V0}0/a  #close the fd  exec 5>&-  #Check that anon-fd based file is not leaking. -EXPECT_WITHIN 20 "N" gf_check_file_opened_in_brick $V0 $H0 $B0/${V0}0 "$realpath" +EXPECT_WITHIN $REOPEN_TIMEOUT "N" gf_check_file_opened_in_brick $V0 $H0 $B0/${V0}0 "$realpath"  cleanup; diff --git a/tests/bugs/bug-830665.t b/tests/bugs/bug-830665.t index e6804891e02..f5a5a67f8e9 100755 --- a/tests/bugs/bug-830665.t +++ b/tests/bugs/bug-830665.t @@ -43,7 +43,7 @@ TEST $CLI volume start $V0;  EXPECT 'Started' volinfo_field $V0 'Status'; -EXPECT_WITHIN 20 "1" is_nfs_export_available; +EXPECT_WITHIN $NFS_EXPORT_TIMEOUT "1" is_nfs_export_available;  ## Mount NFS  TEST mount_nfs $H0:/$V0 $N0 nolock; @@ -69,7 +69,7 @@ setfattr -n trusted.glusterfs.volume-id -v $volid $B0/${V0}-0  ## Restart and remount. Note that we use actimeo=0 so that the stat calls  ## we need for self-heal don't get blocked by the NFS client.  TEST $CLI volume start $V0; -EXPECT_WITHIN 20 "1" is_nfs_export_available; +EXPECT_WITHIN $NFS_EXPORT_TIMEOUT "1" is_nfs_export_available;  TEST mount_nfs $H0:/$V0 $N0 nolock,actimeo=0;  ## The Linux NFS client has a really charming habit of caching stuff right @@ -89,8 +89,8 @@ TEST gluster volume heal $V0 full  ## check, but we want to test whether self-heal already happened.  ## Make sure everything's in order on the recreated brick. -EXPECT_WITHIN 20 'test_data' cat $B0/${V0}-0/a_file; -EXPECT_WITHIN 20 'more_test_data' cat $B0/${V0}-0/a_dir/another_file; +EXPECT_WITHIN $HEAL_TIMEOUT 'test_data' cat $B0/${V0}-0/a_file; +EXPECT_WITHIN $HEAL_TIMEOUT 'more_test_data' cat $B0/${V0}-0/a_dir/another_file;  if [ "$EXIT_EARLY" = "1" ]; then  	exit 0; diff --git a/tests/bugs/bug-847622.t b/tests/bugs/bug-847622.t index ffc5b54194f..a3e04ca2047 100755 --- a/tests/bugs/bug-847622.t +++ b/tests/bugs/bug-847622.t @@ -10,7 +10,7 @@ TEST pidof glusterd  TEST $CLI volume create $V0 $H0:$B0/brick0  TEST $CLI volume start $V0 -EXPECT_WITHIN 20 "1" is_nfs_export_available; +EXPECT_WITHIN $NFS_EXPORT_TIMEOUT "1" is_nfs_export_available;  TEST mount_nfs $H0:/$V0 $N0 nolock  cd $N0 diff --git a/tests/bugs/bug-847624.t b/tests/bugs/bug-847624.t index 40dd13de399..fcd203c7422 100755 --- a/tests/bugs/bug-847624.t +++ b/tests/bugs/bug-847624.t @@ -11,7 +11,7 @@ TEST pidof glusterd  TEST $CLI volume create $V0 $H0:$B0/$V0  TEST $CLI volume set $V0 nfs.drc on  TEST $CLI volume start $V0 -EXPECT_WITHIN 20 "1" is_nfs_export_available; +EXPECT_WITHIN $NFS_EXPORT_TIMEOUT "1" is_nfs_export_available;  TEST mount_nfs $H0:/$V0 $N0 nolock  cd $N0  #7 diff --git a/tests/bugs/bug-853258.t b/tests/bugs/bug-853258.t index faa9d4465ed..3d4cfba00a2 100755 --- a/tests/bugs/bug-853258.t +++ b/tests/bugs/bug-853258.t @@ -16,11 +16,11 @@ mkdir -p $H0:$B0/${V0}3  # Create and start a volume.  TEST $CLI volume create $V0 $H0:$B0/${V0}0 $H0:$B0/${V0}1 $H0:$B0/${V0}2  TEST $CLI volume start $V0 -EXPECT_WITHIN 15 'Started' volinfo_field $V0 'Status'; +EXPECT_WITHIN $PROCESS_UP_TIMEOUT 'Started' volinfo_field $V0 'Status';  # Force assignment of initial ranges.  TEST $CLI volume rebalance $V0 fix-layout start -EXPECT_WITHIN 15 "fix-layout completed" rebalance_status_field $V0 +EXPECT_WITHIN $REBALANCE_TIMEOUT "fix-layout completed" rebalance_status_field $V0  # Get the original values.  xattrs="" @@ -32,7 +32,7 @@ done  TEST $CLI volume add-brick $V0 $H0:$B0/${V0}3  # Force assignment of initial ranges.  TEST $CLI volume rebalance $V0 fix-layout start -EXPECT_WITHIN 15 "fix-layout completed" rebalance_status_field $V0 +EXPECT_WITHIN $REBALANCE_TIMEOUT "fix-layout completed" rebalance_status_field $V0  for i in $(seq 0 3); do  	xattrs="$xattrs $(dht_get_layout $B0/${V0}$i)" diff --git a/tests/bugs/bug-861015-index.t b/tests/bugs/bug-861015-index.t index 4b148e6ccc5..4ca115f8263 100644 --- a/tests/bugs/bug-861015-index.t +++ b/tests/bugs/bug-861015-index.t @@ -10,7 +10,7 @@ TEST pidof glusterd  TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1,2,3,4,5}  TEST $CLI volume set $V0 ensure-durability off  TEST $CLI volume start $V0 -EXPECT_WITHIN 20 "Y" glustershd_up_status +EXPECT_WITHIN $PROCESS_UP_TIMEOUT "Y" glustershd_up_status  TEST glusterfs --volfile-id=/$V0 --volfile-server=$H0 $M0 --attribute-timeout=0 --entry-timeout=0  TEST kill_brick $V0 $H0 $B0/${V0}0  TEST kill_brick $V0 $H0 $B0/${V0}2 diff --git a/tests/bugs/bug-864222.t b/tests/bugs/bug-864222.t index 958fcb9dd78..4fd2f9235b1 100755 --- a/tests/bugs/bug-864222.t +++ b/tests/bugs/bug-864222.t @@ -10,7 +10,7 @@ TEST pidof glusterd  TEST $CLI volume create $V0 $H0:$B0/brick0  TEST $CLI volume start $V0 -EXPECT_WITHIN 20 "1" is_nfs_export_available; +EXPECT_WITHIN $NFS_EXPORT_TIMEOUT "1" is_nfs_export_available;  TEST mount_nfs $H0:/$V0 $N0 nolock  cd $N0 diff --git a/tests/bugs/bug-865825.t b/tests/bugs/bug-865825.t index 4b4b8427cd7..7ca076dcc84 100755 --- a/tests/bugs/bug-865825.t +++ b/tests/bugs/bug-865825.t @@ -62,7 +62,7 @@ sleep 10  gluster volume heal $V0 full  ## Make sure brick 2 now has the correct contents. -EXPECT_WITHIN 30 "test_data" cat $B0/${V0}-2/a_file +EXPECT_WITHIN $HEAL_TIMEOUT "test_data" cat $B0/${V0}-2/a_file  if [ "$EXIT_EARLY" = "1" ]; then  	exit 0; diff --git a/tests/bugs/bug-872923.t b/tests/bugs/bug-872923.t index 68d2e116769..e94fd905e3b 100755 --- a/tests/bugs/bug-872923.t +++ b/tests/bugs/bug-872923.t @@ -11,7 +11,7 @@ TEST $CLI volume info  TEST $CLI volume create $V0 replica 2 $H0:$B0/brick0 $H0:$B0/brick1  TEST $CLI volume start $V0 -EXPECT_WITHIN 20 "1" is_nfs_export_available; +EXPECT_WITHIN $NFS_EXPORT_TIMEOUT "1" is_nfs_export_available;  TEST mount_nfs $H0:/$V0 $N0 nolock  cd $N0 diff --git a/tests/bugs/bug-873962-spb.t b/tests/bugs/bug-873962-spb.t index 62a8318ed42..2821952ac9a 100644 --- a/tests/bugs/bug-873962-spb.t +++ b/tests/bugs/bug-873962-spb.t @@ -24,12 +24,12 @@ exec 5<$M0/a  kill_brick $V0 $H0 $B0/${V0}0  echo "hi" > $M0/a  TEST $CLI volume start $V0 force -EXPECT_WITHIN 20 "1" afr_child_up_status $V0 0 +EXPECT_WITHIN $CHILD_UP_TIMEOUT "1" afr_child_up_status $V0 0  kill_brick $V0 $H0 $B0/${V0}1  echo "bye" > $M0/a  TEST $CLI volume start $V0 force -EXPECT_WITHIN 20 "1" afr_child_up_status $V0 1 +EXPECT_WITHIN $CHILD_UP_TIMEOUT "1" afr_child_up_status $V0 1  TEST ! cat $M0/a #To mark split-brain diff --git a/tests/bugs/bug-873962.t b/tests/bugs/bug-873962.t index 0281417f07f..47b39984cd5 100755 --- a/tests/bugs/bug-873962.t +++ b/tests/bugs/bug-873962.t @@ -50,15 +50,15 @@ echo "1" > $M0/c  TEST setfattr -n trusted.mdata -v abc $M0/b  TEST setfattr -n trusted.mdata -v abc $M0/d  TEST $CLI volume start $V0 force -EXPECT_WITHIN 20 "1" afr_child_up_status $V0 1 +EXPECT_WITHIN $CHILD_UP_TIMEOUT "1" afr_child_up_status $V0 1  TEST kill_brick $V0 $H0 $B0/${V0}1  echo "2" > $M0/a  echo "2" > $M0/c  TEST setfattr -n trusted.mdata -v def $M0/b  TEST setfattr -n trusted.mdata -v def $M0/d  TEST $CLI volume start $V0 force -EXPECT_WITHIN 20 "1" afr_child_up_status $V0 0 -EXPECT_WITHIN 20 "1" afr_child_up_status $V0 1 +EXPECT_WITHIN $CHILD_UP_TIMEOUT "1" afr_child_up_status $V0 0 +EXPECT_WITHIN $CHILD_UP_TIMEOUT "1" afr_child_up_status $V0 1  TEST glusterfs --entry-timeout=0 --attribute-timeout=0 -s $H0 --volfile-id=$V0 $M1 --direct-io-mode=enable diff --git a/tests/bugs/bug-874498.t b/tests/bugs/bug-874498.t index 0b5991011d4..de58aef4d39 100644 --- a/tests/bugs/bug-874498.t +++ b/tests/bugs/bug-874498.t @@ -53,7 +53,7 @@ TEST $CLI volume heal $V0  ##Expected number of entries are 0 in the .glusterfs/indices/xattrop directory -EXPECT_WITHIN 60 '0' count_sh_entries $FILEN; +EXPECT_WITHIN $HEAL_TIMEOUT '0' count_sh_entries $FILEN;  TEST $CLI volume stop $V0;  TEST $CLI volume delete $V0; diff --git a/tests/bugs/bug-877885.t b/tests/bugs/bug-877885.t index 05b4f1cdf2b..955d517876a 100755 --- a/tests/bugs/bug-877885.t +++ b/tests/bugs/bug-877885.t @@ -17,7 +17,7 @@ $M0;  TEST touch $M0/file  TEST mkdir $M0/dir -EXPECT_WITHIN 20 "1" is_nfs_export_available; +EXPECT_WITHIN $NFS_EXPORT_TIMEOUT "1" is_nfs_export_available;  TEST mount_nfs $H0:/$V0 $N0 nolock  cd $N0 diff --git a/tests/bugs/bug-884455.t b/tests/bugs/bug-884455.t index 3b3a2241e14..e63af4334ae 100755 --- a/tests/bugs/bug-884455.t +++ b/tests/bugs/bug-884455.t @@ -66,7 +66,7 @@ TEST ls -l $M0 2>/dev/null;  TEST $CLI volume rebalance $V0 start force -EXPECT_WITHIN 30 "0" rebalance_completed +EXPECT_WITHIN $REBALANCE_TIMEOUT "0" rebalance_completed  ## check for layout overlaps.  EXPECT "0" get_layout $B0/${V0}0 $B0/${V0}1 $B0/${V0}2 diff --git a/tests/bugs/bug-887145.t b/tests/bugs/bug-887145.t index b80f4bd392f..afd39da1616 100755 --- a/tests/bugs/bug-887145.t +++ b/tests/bugs/bug-887145.t @@ -15,7 +15,7 @@ sleep 2;  ## Mount FUSE with caching disabled  TEST glusterfs --entry-timeout=0 --attribute-timeout=0 -s $H0 --volfile-id $V0 $M0; -EXPECT_WITHIN 20 "1" is_nfs_export_available; +EXPECT_WITHIN $NFS_EXPORT_TIMEOUT "1" is_nfs_export_available;  useradd tmp_user 2>/dev/null 1>/dev/null; @@ -34,7 +34,7 @@ TEST $CLI volume set $V0 server.root-squash on;  sleep 2; -EXPECT_WITHIN 20 "1" is_nfs_export_available; +EXPECT_WITHIN $NFS_EXPORT_TIMEOUT "1" is_nfs_export_available;  # create files and directories in the root of the glusterfs and nfs mount  # which is owned by root and hence the right behavior is getting EACCESS diff --git a/tests/bugs/bug-888752.t b/tests/bugs/bug-888752.t index 56d3f9ffb2b..b82c0ddb33a 100644 --- a/tests/bugs/bug-888752.t +++ b/tests/bugs/bug-888752.t @@ -13,7 +13,7 @@ cleanup  TEST launch_cluster 2;  TEST $CLI_1 peer probe $H2; -EXPECT_WITHIN 5 1 get_peer_count +EXPECT_WITHIN $PROBE_TIMEOUT 1 get_peer_count  TEST $CLI_1 volume create $V0 $H1:$B1/$V0 $H2:$B2/$V0  TEST $CLI_1 volume start $V0 diff --git a/tests/bugs/bug-889630.t b/tests/bugs/bug-889630.t index b04eb34076e..d2fcc10a4d4 100755 --- a/tests/bugs/bug-889630.t +++ b/tests/bugs/bug-889630.t @@ -21,7 +21,7 @@ cleanup;  TEST launch_cluster 2;  TEST $CLI_1 peer probe $H2; -EXPECT_WITHIN 20 1 check_peers +EXPECT_WITHIN $PROBE_TIMEOUT 1 check_peers  TEST $CLI_1 volume create $V0 $H1:$B1/$V0 $H2:$B2/$V0  TEST $CLI_1 volume start $V0 diff --git a/tests/bugs/bug-904065.t b/tests/bugs/bug-904065.t index 5d5045bb734..ff677453015 100755 --- a/tests/bugs/bug-904065.t +++ b/tests/bugs/bug-904065.t @@ -34,7 +34,7 @@ TEST $CLI volume start $V0;  EXPECT 'Started' volinfo_field $V0 'Status'  # glusterfs/nfs needs some time to start up in the background -EXPECT_WITHIN 20 1 is_nfs_export_available +EXPECT_WITHIN $NFS_EXPORT_TIMEOUT 1 is_nfs_export_available  # before mounting the rmtab should be empty  EXPECT '0' count_lines /var/lib/glusterd/nfs/rmtab @@ -73,7 +73,7 @@ EXPECT '2' count_lines $M0/rmtab  TEST gluster volume set $V0 nfs.mount-rmtab $M0/rmtab  # glusterfs/nfs needs some time to restart -EXPECT_WITHIN 20 1 is_nfs_export_available +EXPECT_WITHIN $NFS_EXPORT_TIMEOUT 1 is_nfs_export_available  # a new mount should be added to the rmtab, not overwrite exiting ones  TEST mount_nfs $H0:/$V0 $N0 nolock diff --git a/tests/bugs/bug-904300.t b/tests/bugs/bug-904300.t index 1e6917463ec..95be2724282 100755 --- a/tests/bugs/bug-904300.t +++ b/tests/bugs/bug-904300.t @@ -11,7 +11,7 @@ TEST pidof glusterd  TEST $CLI volume create $V0 $H0:$B0/$V0;  TEST $CLI volume start $V0 -EXPECT_WITHIN 20 1 is_nfs_export_available +EXPECT_WITHIN $NFS_EXPORT_TIMEOUT 1 is_nfs_export_available  TEST mount_nfs $H0:/$V0 $N0 nolock  TEST mkdir $N0/dir1 @@ -21,7 +21,7 @@ TEST umount $N0  # Case 1: Allow "dir1" to be mounted only from 127.0.0.1  # 9-12  TEST $CLI volume set $V0 export-dir \""/dir1(127.0.0.1)"\" -EXPECT_WITHIN 20 2 is_nfs_export_available +EXPECT_WITHIN $NFS_EXPORT_TIMEOUT 2 is_nfs_export_available  TEST mount_nfs localhost:/$V0/dir1 $N0 nolock  TEST umount $N0 @@ -31,7 +31,7 @@ TEST umount $N0  #         a negative test case therefore the mount should fail.  # 13-16  TEST $CLI volume set $V0 export-dir \""/dir1(8.8.8.8)"\" -EXPECT_WITHIN 20 2 is_nfs_export_available +EXPECT_WITHIN $NFS_EXPORT_TIMEOUT 2 is_nfs_export_available  TEST ! mount_nfs $H0:/$V0/dir1 $N0 nolock  TEST ! umount $N0 @@ -41,7 +41,7 @@ TEST ! umount $N0  #         instead of ip address.  # 17-20  TEST $CLI volume set $V0 export-dir \""/dir1($H0)"\" -EXPECT_WITHIN 20 2 is_nfs_export_available +EXPECT_WITHIN $NFS_EXPORT_TIMEOUT 2 is_nfs_export_available  TEST mount_nfs $H0:/$V0/dir1 $N0 nolock  TEST umount $N0 @@ -49,7 +49,7 @@ TEST umount $N0  # Case 4: Variation of test case1. Here we are checking with IP range  # 21-24  TEST $CLI volume set $V0 export-dir \""/dir1(127.0.0.0/24)"\" -EXPECT_WITHIN 20 2 is_nfs_export_available +EXPECT_WITHIN $NFS_EXPORT_TIMEOUT 2 is_nfs_export_available  TEST mount_nfs localhost:/$V0/dir1 $N0 nolock  TEST umount $N0 diff --git a/tests/bugs/bug-906646.t b/tests/bugs/bug-906646.t index b2cbf6bc32e..754f193beb8 100644 --- a/tests/bugs/bug-906646.t +++ b/tests/bugs/bug-906646.t @@ -82,7 +82,7 @@ EXPECT 1 xattr_query_check ${backend_paths_array[1]} "trusted.name"  # restart the brick process  TEST $CLI volume start $V0 force -EXPECT_WITHIN 20 "1" afr_child_up_status $V0 `expr $brick_id - 1` +EXPECT_WITHIN $CHILD_UP_TIMEOUT "1" afr_child_up_status $V0 `expr $brick_id - 1`  cat $pth >/dev/null diff --git a/tests/bugs/bug-912564.t b/tests/bugs/bug-912564.t index b24268fbc9b..4fc548c695c 100755 --- a/tests/bugs/bug-912564.t +++ b/tests/bugs/bug-912564.t @@ -38,7 +38,7 @@ mkdir -p $H0:$B0/${V0}3  TEST $CLI volume create $V0 $H0:$B0/${V0}0 $H0:$B0/${V0}1 \  			    $H0:$B0/${V0}2 $H0:$B0/${V0}3  TEST $CLI volume start $V0 -EXPECT_WITHIN 15 'Started' volinfo_field $V0 'Status'; +EXPECT_WITHIN $PROCESS_UP_TIMEOUT 'Started' volinfo_field $V0 'Status';  # Mount it.  TEST glusterfs --volfile-server=$H0 --volfile-id=$V0 $M0 diff --git a/tests/bugs/bug-913051.t b/tests/bugs/bug-913051.t index 9a59424f412..cdf1bfcd1dc 100644 --- a/tests/bugs/bug-913051.t +++ b/tests/bugs/bug-913051.t @@ -35,7 +35,7 @@ TEST rfd=`fd_available`  TEST fd_open $rfd "r" $M0/dir/b  TEST $CLI volume start $V0 force -EXPECT_WITHIN 20 "1" afr_child_up_status $V0 0 +EXPECT_WITHIN $CHILD_UP_TIMEOUT "1" afr_child_up_status $V0 0  #check that the files are not opned on brick-0  realpatha=$(gf_get_gfid_backend_file_path $B0/${V0}0 "dir/a") @@ -57,8 +57,8 @@ TEST fd_write $wfd "open sesame"  #trigger readv for attempting open-fd-fix in afr  TEST fd_cat $rfd -EXPECT_WITHIN 20 "Y" gf_check_file_opened_in_brick $V0 $H0 $B0/${V0}0 $B0/${V0}0/dir/a -EXPECT_WITHIN 20 "Y" gf_check_file_opened_in_brick $V0 $H0 $B0/${V0}0 $B0/${V0}0/dir/b +EXPECT_WITHIN $REOPEN_TIMEOUT "Y" gf_check_file_opened_in_brick $V0 $H0 $B0/${V0}0 $B0/${V0}0/dir/a +EXPECT_WITHIN $REOPEN_TIMEOUT "Y" gf_check_file_opened_in_brick $V0 $H0 $B0/${V0}0 $B0/${V0}0/dir/b  TEST fd_close $wfd  TEST fd_close $rfd diff --git a/tests/bugs/bug-913555.t b/tests/bugs/bug-913555.t index f58d7bd6dd6..2393a16ad6f 100755 --- a/tests/bugs/bug-913555.t +++ b/tests/bugs/bug-913555.t @@ -26,7 +26,7 @@ TEST launch_cluster 3; # start 3-node virtual cluster  TEST $CLI_1 peer probe $H2; # peer probe server 2 from server 1 cli  TEST $CLI_1 peer probe $H3; # peer probe server 3 from server 1 cli -EXPECT_WITHIN 20 2 check_peers +EXPECT_WITHIN $PROBE_TIMEOUT 2 check_peers  TEST $CLI_1 volume create $V0 $H1:$B1/$V0 $H2:$B2/$V0 $H3:$B3/$V0  TEST $CLI_1 volume set $V0 cluster.server-quorum-type server @@ -35,20 +35,20 @@ TEST glusterfs --volfile-server=$H1 --volfile-id=$V0 $M0  # Kill one pseudo-node, make sure the others survive and volume stays up.  TEST kill_node 3; -EXPECT_WITHIN 20 1 check_peers; +EXPECT_WITHIN $PROBE_TIMEOUT 1 check_peers;  EXPECT 0 check_fs $M0;  EXPECT 2 glusterfsd_count;  # Kill another pseudo-node, make sure the last one dies and volume goes down.  TEST kill_node 2; -EXPECT_WITHIN 20 0 check_peers +EXPECT_WITHIN $PROBE_TIMEOUT 0 check_peers  EXPECT 1 check_fs $M0;  EXPECT 0 glusterfsd_count; # the two glusterfsds of the other two glusterds                             # must be dead  TEST $glusterd_2;  TEST $glusterd_3; -EXPECT_WITHIN 20 3 glusterfsd_count; # restore quorum, all ok -EXPECT_WITHIN 5 0 check_fs $M0; +EXPECT_WITHIN $PROCESS_UP_TIMEOUT 3 glusterfsd_count; # restore quorum, all ok +EXPECT_WITHIN $PROCESS_UP_TIMEOUT 0 check_fs $M0;  cleanup diff --git a/tests/bugs/bug-915280.t b/tests/bugs/bug-915280.t index 218fa7baf89..2c3553be540 100755 --- a/tests/bugs/bug-915280.t +++ b/tests/bugs/bug-915280.t @@ -24,7 +24,7 @@ TEST $CLI volume start $V0;  EXPECT 'Started' volinfo_field $V0 'Status';  MOUNTDIR=$N0; -EXPECT_WITHIN 20 "1" is_nfs_export_available; +EXPECT_WITHIN $NFS_EXPORT_TIMEOUT "1" is_nfs_export_available;  TEST mount_nfs $H0:/$V0 $N0 nolock,timeo=30,retrans=1  TEST touch $N0/testfile diff --git a/tests/bugs/bug-915554.t b/tests/bugs/bug-915554.t index beb669f8cef..0425117a42d 100755 --- a/tests/bugs/bug-915554.t +++ b/tests/bugs/bug-915554.t @@ -60,7 +60,7 @@ TEST $CLI volume rebalance $V0 start force  # check if rebalance has completed for upto 15 secs -EXPECT_WITHIN 30 "0" rebalance_completed +EXPECT_WITHIN $REBALANCE_TIMEOUT "0" rebalance_completed  # validate the file size after the migration  SIZE2=`stat -c %s $M0/$i` diff --git a/tests/bugs/bug-916226.t b/tests/bugs/bug-916226.t index 2abfa1fc604..50d1e312012 100644 --- a/tests/bugs/bug-916226.t +++ b/tests/bugs/bug-916226.t @@ -21,6 +21,6 @@ TEST touch $M0/dir{1..10}/files{1..10};  TEST $CLI volume add-brick $V0 $H0:$B0/${V0}4 $H0:/$B0/${V0}5  TEST $CLI volume rebalance $V0 start force -EXPECT_WITHIN 60 "completed" rebalance_status_field $V0 +EXPECT_WITHIN $REBALANCE_TIMEOUT "completed" rebalance_status_field $V0  cleanup; diff --git a/tests/bugs/bug-918437-sh-mtime.t b/tests/bugs/bug-918437-sh-mtime.t index 11155ad1629..e541a81d642 100644 --- a/tests/bugs/bug-918437-sh-mtime.t +++ b/tests/bugs/bug-918437-sh-mtime.t @@ -35,15 +35,15 @@ TEST gf_rm_file_and_gfid_link $B0/gfs0/brick01 a  TEST gf_rm_file_and_gfid_link $B0/gfs0/brick02 b  TEST $CLI volume start $V0 force -EXPECT_WITHIN 20 "1" afr_child_up_status $V0 0 -EXPECT_WITHIN 20 "1" afr_child_up_status $V0 1 +EXPECT_WITHIN $CHILD_UP_TIMEOUT "1" afr_child_up_status $V0 0 +EXPECT_WITHIN $CHILD_UP_TIMEOUT "1" afr_child_up_status $V0 1  TEST $CLI volume set $V0 cluster.self-heal-daemon on  sleep 1  TEST gluster volume heal $V0 full  size=`stat -c '%s' /etc/passwd` -EXPECT_WITHIN 60 $size stat -c '%s' $B0/gfs0/brick01/a +EXPECT_WITHIN $HEAL_TIMEOUT $size stat -c '%s' $B0/gfs0/brick01/a  TEST modify_atstamp1=$(get_mtime $B0/gfs0/brick01/a)  TEST modify_atstamp2=$(get_mtime $B0/gfs0/brick02/a) diff --git a/tests/bugs/bug-921072.t b/tests/bugs/bug-921072.t index 8ccdea90399..ea5a91c2c24 100755 --- a/tests/bugs/bug-921072.t +++ b/tests/bugs/bug-921072.t @@ -11,33 +11,33 @@ TEST pidof glusterd  TEST $CLI volume create $V0 $H0:$B0/$V0  TEST $CLI volume start $V0 -EXPECT_WITHIN 20 1 is_nfs_export_available +EXPECT_WITHIN $NFS_EXPORT_TIMEOUT 1 is_nfs_export_available  TEST mount_nfs $H0:/$V0 $N0 nolock  TEST umount $N0  # based on ip addresses (1-4)  # case 1: allow only localhost ip  TEST $CLI volume set $V0 nfs.rpc-auth-allow 127.0.0.1 -EXPECT_WITHIN 20 1 is_nfs_export_available +EXPECT_WITHIN $NFS_EXPORT_TIMEOUT 1 is_nfs_export_available  TEST mount_nfs localhost:/$V0 $N0 nolock  TEST umount $N0  # case 2: allow only non-localhost ip  TEST $CLI volume set $V0 nfs.rpc-auth-allow 192.168.1.1 -EXPECT_WITHIN 20 1 is_nfs_export_available +EXPECT_WITHIN $NFS_EXPORT_TIMEOUT 1 is_nfs_export_available  #11  TEST ! mount_nfs localhost:/$V0 $N0 nolock  TEST $CLI volume reset --mode=script $V0  # case 3: reject only localhost ip  TEST $CLI volume set $V0 nfs.rpc-auth-reject 127.0.0.1 -EXPECT_WITHIN 20 1 is_nfs_export_available +EXPECT_WITHIN $NFS_EXPORT_TIMEOUT 1 is_nfs_export_available  TEST ! mount_nfs localhost:/$V0 $N0 nolock  # case 4: reject only non-localhost ip  TEST $CLI volume set $V0 nfs.rpc-auth-reject 192.168.1.1 -EXPECT_WITHIN 20 1 is_nfs_export_available +EXPECT_WITHIN $NFS_EXPORT_TIMEOUT 1 is_nfs_export_available  TEST mount_nfs localhost:/$V0 $N0 nolock  TEST umount $N0 @@ -48,21 +48,21 @@ TEST umount $N0  # CASES WITH NFS.ADDR-NAMELOOKUP ON (5-12)  TEST $CLI volume reset --mode=script $V0  TEST $CLI volume set $V0 nfs.addr-namelookup on -EXPECT_WITHIN 20 1 is_nfs_export_available +EXPECT_WITHIN $NFS_EXPORT_TIMEOUT 1 is_nfs_export_available  #20  TEST mount_nfs localhost:/$V0 $N0 nolock  TEST umount $N0  # case 5: allow only localhost  TEST $CLI volume set $V0 nfs.rpc-auth-allow localhost -EXPECT_WITHIN 20 1 is_nfs_export_available +EXPECT_WITHIN $NFS_EXPORT_TIMEOUT 1 is_nfs_export_available  TEST mount_nfs localhost:/$V0 $N0 nolock  TEST umount $N0  # case 6: allow only somehost  TEST $CLI volume set $V0 nfs.rpc-auth-allow somehost -EXPECT_WITHIN 20 1 is_nfs_export_available +EXPECT_WITHIN $NFS_EXPORT_TIMEOUT 1 is_nfs_export_available  TEST ! mount_nfs localhost:/$V0 $N0 nolock @@ -70,13 +70,13 @@ TEST ! mount_nfs localhost:/$V0 $N0 nolock  TEST $CLI volume reset --mode=script $V0  TEST $CLI volume set $V0 nfs.addr-namelookup on  TEST $CLI volume set $V0 nfs.rpc-auth-reject localhost -EXPECT_WITHIN 20 1 is_nfs_export_available +EXPECT_WITHIN $NFS_EXPORT_TIMEOUT 1 is_nfs_export_available  #30  TEST ! mount_nfs localhost:/$V0 $N0 nolock  # case 8: reject only somehost  TEST $CLI volume set $V0 nfs.rpc-auth-reject somehost -EXPECT_WITHIN 20 1 is_nfs_export_available +EXPECT_WITHIN $NFS_EXPORT_TIMEOUT 1 is_nfs_export_available  TEST mount_nfs localhost:/$V0 $N0 nolock  TEST umount $N0 @@ -86,7 +86,7 @@ TEST umount $N0  TEST $CLI volume reset --mode=script $V0  TEST $CLI volume set $V0 nfs.addr-namelookup on  TEST $CLI volume set $V0 nfs.rpc-auth-allow 127.0.0.1 -EXPECT_WITHIN 20 1 is_nfs_export_available +EXPECT_WITHIN $NFS_EXPORT_TIMEOUT 1 is_nfs_export_available  TEST mount_nfs localhost:/$V0 $N0 nolock  TEST mkdir -p $N0/subdir @@ -94,7 +94,7 @@ TEST umount $N0  # case 10: allow a non-localhost ip  TEST $CLI volume set $V0 nfs.rpc-auth-allow 192.168.1.1 -EXPECT_WITHIN 20 1 is_nfs_export_available +EXPECT_WITHIN $NFS_EXPORT_TIMEOUT 1 is_nfs_export_available  #41  TEST ! mount_nfs localhost:/$V0 $N0 nolock @@ -102,14 +102,14 @@ TEST ! mount_nfs localhost:/$V0 $N0 nolock  TEST $CLI volume reset --mode=script $V0  TEST $CLI volume set $V0 nfs.addr-namelookup on  TEST $CLI volume set $V0 nfs.rpc-auth-reject 127.0.0.1 -EXPECT_WITHIN 20 1 is_nfs_export_available +EXPECT_WITHIN $NFS_EXPORT_TIMEOUT 1 is_nfs_export_available  TEST ! mount_nfs localhost:/$V0 $N0 nolock  TEST ! mount_nfs localhost:/$V0/subdir $N0 nolock  # case 12: reject only non-localhost ip  TEST $CLI volume set $V0 nfs.rpc-auth-reject 192.168.1.1 -EXPECT_WITHIN 20 1 is_nfs_export_available +EXPECT_WITHIN $NFS_EXPORT_TIMEOUT 1 is_nfs_export_available  TEST mount_nfs localhost:/$V0 $N0 nolock  TEST umount $N0 diff --git a/tests/bugs/bug-921408.t b/tests/bugs/bug-921408.t index ef2b4fb21cd..483bc6b3d3f 100755 --- a/tests/bugs/bug-921408.t +++ b/tests/bugs/bug-921408.t @@ -31,7 +31,7 @@ addbr_rebal_till_layout_change()          do                  $CLI volume add-brick $V0 $H0:$B0/${V0}$l &>/dev/null                  $CLI volume rebalance $V0 fix-layout start &>/dev/null -                wait_check_status 15 +                wait_check_status $REBALANCE_TIMEOUT                  if [ $? -eq 1 ]                  then                          break diff --git a/tests/bugs/bug-924265.t b/tests/bugs/bug-924265.t index 13491356dbc..b08c13fddcb 100755 --- a/tests/bugs/bug-924265.t +++ b/tests/bugs/bug-924265.t @@ -24,7 +24,7 @@ TEST $CLI volume set $V0 cluster.dht-xattr-name trusted.foo.bar  # Start and mount the volume.  TEST $CLI volume start $V0 -EXPECT_WITHIN 15 'Started' volinfo_field $V0 'Status'; +EXPECT_WITHIN $PROCESS_UP_TIMEOUT 'Started' volinfo_field $V0 'Status';  TEST glusterfs --volfile-server=$H0 --volfile-id=$V0 $M0  # Create a directory and make sure it has the right xattr. diff --git a/tests/bugs/bug-927616.t b/tests/bugs/bug-927616.t index d8fe7520d8c..d19327ca268 100755 --- a/tests/bugs/bug-927616.t +++ b/tests/bugs/bug-927616.t @@ -15,7 +15,7 @@ sleep 1;  ## Mount FUSE with caching disabled  TEST glusterfs --entry-timeout=0 --attribute-timeout=0 -s $H0 --volfile-id $V0 $M0; -EXPECT_WITHIN 20 "1" is_nfs_export_available; +EXPECT_WITHIN $NFS_EXPORT_TIMEOUT "1" is_nfs_export_available;  TEST mount_nfs $H0:/$V0 $N0 nolock;  TEST mkdir $M0/dir; diff --git a/tests/bugs/bug-948686.t b/tests/bugs/bug-948686.t index db9c198a96f..c9afc2f0758 100755 --- a/tests/bugs/bug-948686.t +++ b/tests/bugs/bug-948686.t @@ -13,7 +13,7 @@ TEST launch_cluster 3; # start 3-node virtual cluster  TEST $CLI_1 peer probe $H2; # peer probe server 2 from server 1 cli  TEST $CLI_1 peer probe $H3; # peer probe server 3 from server 1 cli -EXPECT_WITHIN 20 2 check_peers; +EXPECT_WITHIN $PROBE_TIMEOUT 2 check_peers;  TEST $CLI_1 volume create $V0 replica 2 $H1:$B1/$V0 $H1:$B1/${V0}_1 $H2:$B2/$V0 $H3:$B3/$V0  TEST $CLI_1 volume start $V0 diff --git a/tests/bugs/bug-948729/bug-948729-force.t b/tests/bugs/bug-948729/bug-948729-force.t index d14e9406182..b68b8b88bdd 100644 --- a/tests/bugs/bug-948729/bug-948729-force.t +++ b/tests/bugs/bug-948729/bug-948729-force.t @@ -19,7 +19,7 @@ TEST launch_cluster 2;  TEST $CLI_1 peer probe $H2; -EXPECT_WITHIN 20 1 check_peers; +EXPECT_WITHIN $PROBE_TIMEOUT 1 check_peers;  B3=/d/backends/3  B4=/d/backends/4 diff --git a/tests/bugs/bug-948729/bug-948729-mode-script.t b/tests/bugs/bug-948729/bug-948729-mode-script.t index 5a1fdd3b1db..5ac133a5bfe 100644 --- a/tests/bugs/bug-948729/bug-948729-mode-script.t +++ b/tests/bugs/bug-948729/bug-948729-mode-script.t @@ -17,7 +17,7 @@ TEST launch_cluster 2;  TEST $CLI_1 peer probe $H2; -EXPECT_WITHIN 20 1 check_peers; +EXPECT_WITHIN $PROBE_TIMEOUT 1 check_peers;  B3=/d/backends/3  mkdir -p $B3 diff --git a/tests/bugs/bug-948729/bug-948729.t b/tests/bugs/bug-948729/bug-948729.t index 3914a454c53..8a0620a7d60 100644 --- a/tests/bugs/bug-948729/bug-948729.t +++ b/tests/bugs/bug-948729/bug-948729.t @@ -17,7 +17,7 @@ TEST launch_cluster 2;  TEST $CLI_1 peer probe $H2; -EXPECT_WITHIN 20 1 check_peers; +EXPECT_WITHIN $PROBE_TIMEOUT 1 check_peers;  B3=/d/backends/3 diff --git a/tests/bugs/bug-957877.t b/tests/bugs/bug-957877.t index 23aefea2549..e2b74bf52f5 100644 --- a/tests/bugs/bug-957877.t +++ b/tests/bugs/bug-957877.t @@ -21,7 +21,7 @@ sleep 5  TEST $CLI volume heal $V0  # Wait for self-heal to complete -EXPECT_WITHIN 30 '0' count_sh_entries $BRICK; +EXPECT_WITHIN $HEAL_TIMEOUT '0' count_sh_entries $BRICK;  TEST getfattr -n "user.foo" $B0/${V0}0/f1; diff --git a/tests/bugs/bug-958691.t b/tests/bugs/bug-958691.t index 3f6b93eb1c3..cd38d46e1e4 100644 --- a/tests/bugs/bug-958691.t +++ b/tests/bugs/bug-958691.t @@ -11,7 +11,7 @@ TEST $CLI volume create $V0 $H0:$B0/${V0}{0,1}  TEST $CLI volume start $V0;  TEST glusterfs --volfile-server=$H0 --volfile-id=$V0 $M0; -EXPECT_WITHIN 20 "1" is_nfs_export_available; +EXPECT_WITHIN $NFS_EXPORT_TIMEOUT "1" is_nfs_export_available;  TEST mount_nfs $H0:/$V0 $N0 nolock;  sleep 2; diff --git a/tests/bugs/bug-961615.t b/tests/bugs/bug-961615.t index d183e6c5276..d10eeeabb3b 100644 --- a/tests/bugs/bug-961615.t +++ b/tests/bugs/bug-961615.t @@ -26,7 +26,7 @@ dd if=/dev/zero of=$M0/10 bs=1k &  bg_pid=$!  #Now rebalance force will migrate file '10'  TEST $CLI volume rebalance $V0 start force -EXPECT_WITHIN 60 "completed" rebalance_status_field $V0 +EXPECT_WITHIN $REBALANCE_TIMEOUT "completed" rebalance_status_field $V0  #If the bug exists mount would have crashed by now  TEST ls $M0  kill -9 $bg_pid > /dev/null 2>&1 diff --git a/tests/bugs/bug-964059.t b/tests/bugs/bug-964059.t index df07f95ee99..e81e4d708bc 100755 --- a/tests/bugs/bug-964059.t +++ b/tests/bugs/bug-964059.t @@ -21,7 +21,7 @@ cleanup;  TEST launch_cluster 2;  TEST $CLI_1 peer probe $H2; -EXPECT_WITHIN 20 1 check_peers +EXPECT_WITHIN $PROBE_TIMEOUT 1 check_peers  TEST $CLI_1 volume create $V0 $H1:$B1/$V0 $H2:$B2/$V0  TEST $CLI_1 volume start $V0 diff --git a/tests/bugs/bug-966018.t b/tests/bugs/bug-966018.t index 55a01b8b7b3..7127189c63b 100644 --- a/tests/bugs/bug-966018.t +++ b/tests/bugs/bug-966018.t @@ -19,7 +19,7 @@ TEST $CLI volume set $V0 cluster.post-op-delay-secs 3  TEST $CLI volume start $V0  TEST $CLI volume profile $V0 start -EXPECT_WITHIN 20 "1" is_nfs_export_available; +EXPECT_WITHIN $NFS_EXPORT_TIMEOUT "1" is_nfs_export_available;  TEST mount_nfs $H0:/$V0 $N0 nolock;  TEST glusterfs --entry-timeout=0 --attribute-timeout=0 -s $H0 --volfile-id=$V0 $M0  echo 1 > $N0/1 && chmod +x $N0/1 diff --git a/tests/bugs/bug-973073.t b/tests/bugs/bug-973073.t index 619f1733114..3ea54132a83 100755 --- a/tests/bugs/bug-973073.t +++ b/tests/bugs/bug-973073.t @@ -34,13 +34,13 @@ TEST glusterfs -s $H0 --volfile-id $V0 $M0;  TEST $CLI volume remove-brick $V0 $H0:$B0/${V0}2 start  ## remove-brick status == rebalance_status -EXPECT_WITHIN 30 "0" remove_brick_completed +EXPECT_WITHIN $REBALANCE_TIMEOUT "0" remove_brick_completed  TEST $CLI volume remove-brick $V0 $H0:$B0/${V0}2 stop  TEST $CLI volume rebalance $V0 fix-layout start -EXPECT_WITHIN 30 "0" rebalance_completed +EXPECT_WITHIN $REBALANCE_TIMEOUT "0" rebalance_completed  TEST mkdir $M0/dir 2>/dev/null; diff --git a/tests/bugs/bug-974007.t b/tests/bugs/bug-974007.t index c8c1c862b33..241aa6ba28e 100644 --- a/tests/bugs/bug-974007.t +++ b/tests/bugs/bug-974007.t @@ -31,7 +31,7 @@ function remove_brick_start_status {  EXPECT "success"  remove_brick_start_status;  # Wait for rebalance to complete -EXPECT_WITHIN 10 "completed" remove_brick_status_completed_field "$V0" "$H0:$B0/${V0}6 $H0:$B0/${V0}1 $H0:$B0/${V0}2 $H0:$B0/${V0}5" +EXPECT_WITHIN $REBALANCE_TIMEOUT "completed" remove_brick_status_completed_field "$V0" "$H0:$B0/${V0}6 $H0:$B0/${V0}1 $H0:$B0/${V0}2 $H0:$B0/${V0}5"  # Check commit status  function remove_brick_commit_status { diff --git a/tests/bugs/bug-974972.t b/tests/bugs/bug-974972.t index 1b57307e76d..4d367a52cdf 100755 --- a/tests/bugs/bug-974972.t +++ b/tests/bugs/bug-974972.t @@ -12,22 +12,22 @@ TEST pidof glusterd  TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1}  TEST $CLI volume set $V0 self-heal-daemon off  TEST $CLI volume start $V0 -EXPECT_WITHIN 20 "1" is_nfs_export_available; +EXPECT_WITHIN $NFS_EXPORT_TIMEOUT "1" is_nfs_export_available;  TEST mount_nfs $H0:/$V0 $N0  TEST touch $N0/1  TEST kill_brick ${V0} ${H0} ${B0}/${V0}1  echo abc > $N0/1  TEST $CLI volume start $V0 force -EXPECT_WITHIN 20 "Y" nfs_up_status -EXPECT_WITHIN 20 "1" afr_child_up_status_in_nfs $V0 0 -EXPECT_WITHIN 20 "1" afr_child_up_status_in_nfs $V0 1 +EXPECT_WITHIN $PROCESS_UP_TIMEOUT "Y" nfs_up_status +EXPECT_WITHIN $CHILD_UP_TIMEOUT "1" afr_child_up_status_in_nfs $V0 0 +EXPECT_WITHIN $CHILD_UP_TIMEOUT "1" afr_child_up_status_in_nfs $V0 1  TEST kill_brick ${V0} ${H0} ${B0}/${V0}0  echo def > $N0/1  TEST $CLI volume start $V0 force -EXPECT_WITHIN 20 "Y" nfs_up_status -EXPECT_WITHIN 20 "1" afr_child_up_status_in_nfs $V0 0 -EXPECT_WITHIN 20 "1" afr_child_up_status_in_nfs $V0 1 +EXPECT_WITHIN $PROCESS_UP_TIMEOUT "Y" nfs_up_status +EXPECT_WITHIN $CHILD_UP_TIMEOUT "1" afr_child_up_status_in_nfs $V0 0 +EXPECT_WITHIN $CHILD_UP_TIMEOUT "1" afr_child_up_status_in_nfs $V0 1  #Lookup should not fail  TEST ls $N0/1 diff --git a/tests/bugs/bug-977797.t b/tests/bugs/bug-977797.t index f2252159a21..a8b1ee7b4e4 100755 --- a/tests/bugs/bug-977797.t +++ b/tests/bugs/bug-977797.t @@ -42,7 +42,7 @@ TEST `echo "GLUSTER-FILE-SYSTEM" > $M0/a/file`  TEST mkdir $M0/a/b  TEST $CLI volume start $V0 force -EXPECT_WITHIN 20 "1" afr_child_up_status $V0 0; +EXPECT_WITHIN $CHILD_UP_TIMEOUT "1" afr_child_up_status $V0 0; @@ -52,7 +52,7 @@ TEST chmod 757 $M0/a  TEST chmod 757 $M0/a/file  TEST $CLI volume start $V0 force -EXPECT_WITHIN 20 "1" afr_child_up_status $V0 1; +EXPECT_WITHIN $CHILD_UP_TIMEOUT "1" afr_child_up_status $V0 1;  TEST dd if=$M0/a/file of=/dev/null bs=1M diff --git a/tests/bugs/bug-978794.t b/tests/bugs/bug-978794.t index d22d3cde33d..8cda83efe0a 100644 --- a/tests/bugs/bug-978794.t +++ b/tests/bugs/bug-978794.t @@ -18,12 +18,12 @@ TEST touch $M0/{1..100}  for i in {1..100}; do fd[$i]=`fd_available`; fd_open ${fd[$i]} 'w' $M0/$i; done  TEST $CLI volume add-brick $V0 $H0:$B0/${V0}{2,3}  TEST $CLI volume rebalance $V0 start force -EXPECT_WITHIN 120 "completed" rebalance_status_field $V0 +EXPECT_WITHIN $REBALANCE_TIMEOUT "completed" rebalance_status_field $V0  TEST cat $M0/{1..100}  for i in {1..100}; do fd_write ${fd[$i]} 'abc'; done  TEST $CLI volume add-brick $V0 $H0:$B0/${V0}{4,5}  TEST $CLI volume rebalance $V0 start force -EXPECT_WITHIN 120 "completed" rebalance_status_field $V0 +EXPECT_WITHIN $REBALANCE_TIMEOUT "completed" rebalance_status_field $V0  for i in {1..100}; do fd_write ${fd[$i]} 'abc'; done  TEST cat $M0/{1..100}  cleanup diff --git a/tests/bugs/bug-983477.t b/tests/bugs/bug-983477.t index c19fa96c8bf..b9d3002be30 100755 --- a/tests/bugs/bug-983477.t +++ b/tests/bugs/bug-983477.t @@ -8,6 +8,7 @@  function get_use_readdirp_value {          local vol=$1          local statedump=$(generate_mount_statedump $vol) +        sleep 1          local val=$(grep "use_readdirp=" $statedump | cut -f2 -d'=' | tail -1)          rm -f $statedump          echo $val @@ -21,28 +22,28 @@ TEST $CLI volume start $V0  #If readdirp is enabled statedump should reflect it  TEST glusterfs --volfile-id=/$V0 --volfile-server=$H0 $M0 --attribute-timeout=0 --entry-timeout=0 --use-readdirp=yes  TEST cd $M0 -EXPECT_WITHIN 20 "1" get_use_readdirp_value $V0 +EXPECT "1" get_use_readdirp_value $V0  TEST cd -  TEST umount $M0  #If readdirp is enabled statedump should reflect it  TEST glusterfs --volfile-id=/$V0 --volfile-server=$H0 $M0 --attribute-timeout=0 --entry-timeout=0 --use-readdirp=no  TEST cd $M0 -EXPECT_WITHIN 20 "0" get_use_readdirp_value $V0 +EXPECT "0" get_use_readdirp_value $V0  TEST cd -  TEST umount $M0  #Since args are optional on this argument just specifying "--use-readdirp" should also turn it `on` not `off`  TEST glusterfs --volfile-id=/$V0 --volfile-server=$H0 $M0 --attribute-timeout=0 --entry-timeout=0 --use-readdirp  TEST cd $M0 -EXPECT_WITHIN 20 "1" get_use_readdirp_value $V0 +EXPECT "1" get_use_readdirp_value $V0  TEST cd -  TEST umount $M0  #By default it is enabled.  TEST glusterfs --volfile-id=/$V0 --volfile-server=$H0 $M0 --attribute-timeout=0 --entry-timeout=0  TEST cd $M0 -EXPECT_WITHIN 20 "1" get_use_readdirp_value $V0 +EXPECT "1" get_use_readdirp_value $V0  TEST cd -  TEST umount $M0 diff --git a/tests/bugs/bug-986905.t b/tests/bugs/bug-986905.t index 0fac40fb4e3..ed11bbbd03d 100755 --- a/tests/bugs/bug-986905.t +++ b/tests/bugs/bug-986905.t @@ -20,7 +20,7 @@ TEST kill_brick $V0 $H0 $B0/${V0}0  TEST touch $M0/a  TEST ln $M0/a $M0/link_a  TEST $CLI volume start $V0 force -EXPECT_WITHIN 20 "1" afr_child_up_status $V0 0 +EXPECT_WITHIN $CHILD_UP_TIMEOUT "1" afr_child_up_status $V0 0  TEST ls -l $M0  inum=$(get_inum $B0/${V0}0/a)  EXPECT "$inum" get_inum $B0/${V0}0/link_a diff --git a/tests/include.rc b/tests/include.rc index 35b0ddd5d63..882785bd1b3 100644 --- a/tests/include.rc +++ b/tests/include.rc @@ -7,6 +7,16 @@ V1=${V1:=patchy1};            # volume name to use in tests  B0=${B0:=/d/backends};        # top level of brick directories  H0=${H0:=`hostname --fqdn`};  # hostname  DEBUG=${DEBUG:=0}             # turn on debugging? + +PROCESS_UP_TIMEOUT=20 +NFS_EXPORT_TIMEOUT=20 +CHILD_UP_TIMEOUT=20 +PROBE_TIMEOUT=20 +REBALANCE_TIMEOUT=120 +REOPEN_TIMEOUT=20 +HEAL_TIMEOUT=60 +MARKER_UPDATE_TIMEOUT=20 +  statedumpdir=`gluster --print-statedumpdir`; # Default directory for statedump  CLI="gluster --mode=script --wignore"; diff --git a/tests/nfs.rc b/tests/nfs.rc index ed80c1b426b..6e4f5c0ab49 100644 --- a/tests/nfs.rc +++ b/tests/nfs.rc @@ -1,6 +1,5 @@  #!/bin/bash -  # Due to portmap registration NFS takes some time to  # export all volumes. Therefore tests should start only  # after exports are visible by showmount command. This  | 
