diff options
-rw-r--r-- | cli/src/cli-rpc-ops.c | 23 | ||||
-rw-r--r-- | tests/basic/tier/new-tier-cmds.t | 64 |
2 files changed, 63 insertions, 24 deletions
diff --git a/cli/src/cli-rpc-ops.c b/cli/src/cli-rpc-ops.c index 9bef143574c..a2519e8c0e8 100644 --- a/cli/src/cli-rpc-ops.c +++ b/cli/src/cli-rpc-ops.c @@ -1557,10 +1557,18 @@ gf_cli_print_rebalance_status (dict_t *dict, enum gf_task_types task_type) snprintf (key, 256, "status-%d", i); ret = dict_get_int32 (dict, key, (int32_t *)&status_rcd); - if (ret) { + if (ret == -ENOENT) { gf_log ("cli", GF_LOG_TRACE, "count %d %d", count, i); gf_log ("cli", GF_LOG_TRACE, "failed to get status"); - goto out; + gf_log ("cli", GF_LOG_ERROR, "node down and has failed" + " to set dict"); + continue; + /* skip this node if value not available*/ + } else if (ret) { + gf_log ("cli", GF_LOG_TRACE, "count %d %d", count, i); + gf_log ("cli", GF_LOG_TRACE, "failed to get status"); + continue; + /* skip this node if value not available*/ } if (GF_DEFRAG_STATUS_NOT_STARTED == status_rcd) @@ -1682,10 +1690,17 @@ gf_cli_print_tier_status (dict_t *dict, enum gf_task_types task_type) snprintf (key, 256, "status-%d", i); ret = dict_get_int32 (dict, key, (int32_t *)&status_rcd); - if (ret) { + if (ret == -ENOENT) { gf_log ("cli", GF_LOG_TRACE, "count: %d, %d," "failed to get status", count, i); - goto out; + gf_log ("cli", GF_LOG_ERROR, "node down and has failed" + " to set dict"); + continue; + /*skipping this node as value unavailable*/ + } else if (ret) { + gf_log ("cli", GF_LOG_TRACE, "count: %d, %d," + "failed to get status", count, i); + continue; } if (GF_DEFRAG_STATUS_NOT_STARTED == status_rcd) diff --git a/tests/basic/tier/new-tier-cmds.t b/tests/basic/tier/new-tier-cmds.t index c38ecf9d7db..ce8dbae8660 100644 --- a/tests/basic/tier/new-tier-cmds.t +++ b/tests/basic/tier/new-tier-cmds.t @@ -3,33 +3,42 @@ . $(dirname $0)/../../include.rc . $(dirname $0)/../../volume.rc . $(dirname $0)/../../tier.rc +. $(dirname $0)/../../cluster.rc # Creates a tiered volume with pure distribute hot and cold tiers # Both hot and cold tiers will have an equal number of bricks. +function check_peers { + $CLI_1 peer status | grep 'Peer in Cluster (Connected)' | wc -l +} + function create_dist_tier_vol () { - mkdir $B0/cold - mkdir $B0/hot - TEST $CLI volume create $V0 disperse 6 disperse-data 4 $H0:$B0/cold/${V0}{1..12} - TEST $CLI volume set $V0 performance.quick-read off - TEST $CLI volume set $V0 performance.io-cache off - TEST $CLI volume start $V0 - TEST $CLI volume attach-tier $V0 replica 2 $H0:$B0/hot/${V0}{0..5} - TEST $CLI volume set $V0 cluster.tier-mode test + TEST $CLI_1 volume create $V0 $H1:$B1/${V0} $H2:$B2/${V0} $H3:$B3/${V0} + TEST $CLI_1 volume start $V0 + TEST $CLI_1 volume attach-tier $V0 $H1:$B1/${V0}_h1 $H2:$B2/${V0}_h2 $H3:$B3/${V0}_h3 } function tier_detach_commit () { - $CLI volume tier $V0 detach commit | grep "success" | wc -l + $CLI_1 volume tier $V0 detach commit | grep "success" | wc -l +} + +function tier_detach_status_node_down () { + $CLI_1 volume tier $V0 detach status | wc -l +} + +function tier_status_node_down () { + $CLI_1 volume tier $V0 status | wc -l } cleanup; -#Basic checks -TEST glusterd -TEST pidof glusterd -TEST $CLI volume status +#setup cluster and test volume +TEST launch_cluster 3; # start 3-node virtual cluster +TEST $CLI_1 peer probe $H2; # peer probe server 2 from server 1 cli +TEST $CLI_1 peer probe $H3; # peer probe server 3 from server 1 cli +EXPECT_WITHIN $PROBE_TIMEOUT 2 check_peers; #Create and start a tiered volume create_dist_tier_vol @@ -37,21 +46,36 @@ create_dist_tier_vol #Issue detach tier on the tiered volume #Will throw error saying detach tier not started -EXPECT "Tier command failed" $CLI volume tier $V0 detach status +EXPECT "Tier command failed" $CLI_1 volume tier $V0 detach status #after starting detach tier the detach tier status should display the status -TEST $CLI volume tier $V0 detach start +TEST $CLI_1 volume tier $V0 detach start + +TEST $CLI_1 volume tier $V0 detach status + +#kill a node +TEST kill_node 2 + +#check if we have the rest of the node available printed in the output of detach status +EXPECT_WITHIN $PROCESS_UP_TIMEOUT "4" tier_detach_status_node_down + +#check if we have the rest of the node available printed in the output of tier status +EXPECT_WITHIN $PROCESS_UP_TIMEOUT "5" tier_status_node_down + +TEST $glusterd_2; + +EXPECT_WITHIN $PROBE_TIMEOUT 2 check_peers; -TEST $CLI volume tier $V0 detach status +TEST $CLI_1 volume tier $V0 detach status -TEST $CLI volume tier $V0 detach stop +TEST $CLI_1 volume tier $V0 detach stop #If detach tier is stopped the detach tier command will fail -EXPECT "Tier command failed" $CLI volume tier $V0 detach status +EXPECT "Tier command failed" $CLI_1 volume tier $V0 detach status -TEST $CLI volume tier $V0 detach start +TEST $CLI_1 volume tier $V0 detach start #wait for the detach to complete EXPECT_WITHIN $PROCESS_UP_TIMEOUT "1" tier_detach_commit @@ -59,7 +83,7 @@ EXPECT_WITHIN $PROCESS_UP_TIMEOUT "1" tier_detach_commit #If detach tier is committed then the detach status should fail throwing an error #saying its not a tiered volume -EXPECT "Tier command failed" $CLI volume tier $V0 detach status +EXPECT "Tier command failed" $CLI_1 volume tier $V0 detach status cleanup; |