diff options
Diffstat (limited to 'cli/src/cli-rpc-ops.c')
-rw-r--r-- | cli/src/cli-rpc-ops.c | 20062 |
1 files changed, 10014 insertions, 10048 deletions
diff --git a/cli/src/cli-rpc-ops.c b/cli/src/cli-rpc-ops.c index 40bd64b983c..fea74667e05 100644 --- a/cli/src/cli-rpc-ops.c +++ b/cli/src/cli-rpc-ops.c @@ -14,14 +14,14 @@ */ #define VOL_TOP_PERF_FILENAME_DEF_WIDTH 47 #define VOL_TOP_PERF_FILENAME_ALT_WIDTH 44 -#define VOL_TOP_PERF_SPEED_WIDTH 4 -#define VOL_TOP_PERF_TIME_WIDTH 26 +#define VOL_TOP_PERF_SPEED_WIDTH 4 +#define VOL_TOP_PERF_TIME_WIDTH 26 #define INDENT_MAIN_HEAD "%-25s %s " /* Do not show estimates if greater than this number */ -#define REBAL_ESTIMATE_SEC_UPPER_LIMIT (60*24*3600) -#define REBAL_ESTIMATE_START_TIME 600 +#define REBAL_ESTIMATE_SEC_UPPER_LIMIT (60 * 24 * 3600) +#define REBAL_ESTIMATE_START_TIME 600 #include "cli.h" #include "compat-errno.h" @@ -46,26 +46,22 @@ #include "quota-common-utils.h" #include "events.h" -enum gf_task_types { - GF_TASK_TYPE_REBALANCE, - GF_TASK_TYPE_REMOVE_BRICK -}; +enum gf_task_types { GF_TASK_TYPE_REBALANCE, GF_TASK_TYPE_REMOVE_BRICK }; extern struct rpc_clnt *global_quotad_rpc; extern rpc_clnt_prog_t cli_quotad_clnt; extern rpc_clnt_prog_t *cli_rpc_prog; -extern int cli_op_ret; -extern int connected; +extern int cli_op_ret; +extern int connected; int32_t -gf_cli_remove_brick (call_frame_t *frame, xlator_t *this, - void *data); - +gf_cli_remove_brick(call_frame_t *frame, xlator_t *this, void *data); -char *cli_vol_status_str[] = {"Created", - "Started", - "Stopped", - }; +char *cli_vol_status_str[] = { + "Created", + "Started", + "Stopped", +}; char *cli_vol_task_status_str[] = {"not started", "in progress", @@ -76,9670 +72,9631 @@ char *cli_vol_task_status_str[] = {"not started", "fix-layout stopped", "fix-layout completed", "fix-layout failed", - "unknown" -}; + "unknown"}; int32_t -gf_cli_snapshot (call_frame_t *frame, xlator_t *this, void *data); +gf_cli_snapshot(call_frame_t *frame, xlator_t *this, void *data); int32_t -gf_cli_get_volume (call_frame_t *frame, xlator_t *this, - void *data); +gf_cli_get_volume(call_frame_t *frame, xlator_t *this, void *data); int -cli_to_glusterd (gf_cli_req *req, call_frame_t *frame, fop_cbk_fn_t cbkfn, - xdrproc_t xdrproc, dict_t *dict, int procnum, xlator_t *this, - rpc_clnt_prog_t *prog, struct iobref *iobref); +cli_to_glusterd(gf_cli_req *req, call_frame_t *frame, fop_cbk_fn_t cbkfn, + xdrproc_t xdrproc, dict_t *dict, int procnum, xlator_t *this, + rpc_clnt_prog_t *prog, struct iobref *iobref); int -add_cli_cmd_timeout_to_dict (dict_t *dict); +add_cli_cmd_timeout_to_dict(dict_t *dict); rpc_clnt_prog_t cli_handshake_prog = { - .progname = "cli handshake", - .prognum = GLUSTER_HNDSK_PROGRAM, - .progver = GLUSTER_HNDSK_VERSION, + .progname = "cli handshake", + .prognum = GLUSTER_HNDSK_PROGRAM, + .progver = GLUSTER_HNDSK_VERSION, }; rpc_clnt_prog_t cli_pmap_prog = { - .progname = "cli portmap", - .prognum = GLUSTER_PMAP_PROGRAM, - .progver = GLUSTER_PMAP_VERSION, + .progname = "cli portmap", + .prognum = GLUSTER_PMAP_PROGRAM, + .progver = GLUSTER_PMAP_VERSION, }; int -gf_cli_probe_cbk (struct rpc_req *req, struct iovec *iov, - int count, void *myframe) -{ - gf_cli_rsp rsp = {0,}; - int ret = -1; - char msg[1024] = {0,}; - - GF_ASSERT (myframe); - - if (-1 == req->rpc_status) { - goto out; - } - - ret = xdr_to_generic (*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); - if (ret < 0) { - gf_log (((call_frame_t *) myframe)->this->name, GF_LOG_ERROR, - "Failed to decode xdr response"); - //rsp.op_ret = -1; - //rsp.op_errno = EINVAL; - goto out; - } - - gf_log ("cli", GF_LOG_INFO, "Received resp to probe"); - - if (rsp.op_errstr && (strlen (rsp.op_errstr) > 0)) { - snprintf (msg, sizeof (msg), "%s", rsp.op_errstr); - if (rsp.op_ret) - gf_log ("cli", GF_LOG_ERROR, "%s", msg); - } +gf_cli_probe_cbk(struct rpc_req *req, struct iovec *iov, int count, + void *myframe) +{ + gf_cli_rsp rsp = { + 0, + }; + int ret = -1; + char msg[1024] = { + 0, + }; + + GF_ASSERT(myframe); + + if (-1 == req->rpc_status) { + goto out; + } + + ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); + if (ret < 0) { + gf_log(((call_frame_t *)myframe)->this->name, GF_LOG_ERROR, + "Failed to decode xdr response"); + // rsp.op_ret = -1; + // rsp.op_errno = EINVAL; + goto out; + } + + gf_log("cli", GF_LOG_INFO, "Received resp to probe"); + + if (rsp.op_errstr && (strlen(rsp.op_errstr) > 0)) { + snprintf(msg, sizeof(msg), "%s", rsp.op_errstr); + if (rsp.op_ret) + gf_log("cli", GF_LOG_ERROR, "%s", msg); + } - if (global_state->mode & GLUSTER_MODE_XML) { - ret = cli_xml_output_str (NULL, - (rsp.op_ret)? NULL : msg, - rsp.op_ret, rsp.op_errno, - (rsp.op_ret)? msg : NULL); - if (ret) - gf_log ("cli", GF_LOG_ERROR, - "Error outputting to xml"); - goto out; - } + if (global_state->mode & GLUSTER_MODE_XML) { + ret = cli_xml_output_str(NULL, (rsp.op_ret) ? NULL : msg, rsp.op_ret, + rsp.op_errno, (rsp.op_ret) ? msg : NULL); + if (ret) + gf_log("cli", GF_LOG_ERROR, "Error outputting to xml"); + goto out; + } - if (!rsp.op_ret) - cli_out ("peer probe: success. %s", msg); - else - cli_err ("peer probe: failed: %s", msg); + if (!rsp.op_ret) + cli_out("peer probe: success. %s", msg); + else + cli_err("peer probe: failed: %s", msg); - ret = rsp.op_ret; + ret = rsp.op_ret; out: - cli_cmd_broadcast_response (ret); - return ret; + cli_cmd_broadcast_response(ret); + return ret; } int -gf_cli_deprobe_cbk (struct rpc_req *req, struct iovec *iov, - int count, void *myframe) -{ - gf_cli_rsp rsp = {0,}; - int ret = -1; - char msg[1024] = {0,}; - - GF_ASSERT (myframe); - - if (-1 == req->rpc_status) { - goto out; - } - - ret = xdr_to_generic (*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); - if (ret < 0) { - gf_log (((call_frame_t *) myframe)->this->name, GF_LOG_ERROR, - "Failed to decode xdr response"); - //rsp.op_ret = -1; - //rsp.op_errno = EINVAL; - goto out; - } - - gf_log ("cli", GF_LOG_INFO, "Received resp to deprobe"); - - if (rsp.op_ret) { - if (strlen (rsp.op_errstr) > 0) { - snprintf (msg, sizeof (msg), "%s", rsp.op_errstr); - gf_log ("cli", GF_LOG_ERROR, "%s", rsp.op_errstr); - } - } else { - snprintf (msg, sizeof (msg), "success"); - } - - if (global_state->mode & GLUSTER_MODE_XML) { - ret = cli_xml_output_str (NULL, - (rsp.op_ret)? NULL : msg, - rsp.op_ret, rsp.op_errno, - (rsp.op_ret)? msg : NULL); - if (ret) - gf_log ("cli", GF_LOG_ERROR, - "Error outputting to xml"); - goto out; - } +gf_cli_deprobe_cbk(struct rpc_req *req, struct iovec *iov, int count, + void *myframe) +{ + gf_cli_rsp rsp = { + 0, + }; + int ret = -1; + char msg[1024] = { + 0, + }; + + GF_ASSERT(myframe); + + if (-1 == req->rpc_status) { + goto out; + } + + ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); + if (ret < 0) { + gf_log(((call_frame_t *)myframe)->this->name, GF_LOG_ERROR, + "Failed to decode xdr response"); + // rsp.op_ret = -1; + // rsp.op_errno = EINVAL; + goto out; + } + + gf_log("cli", GF_LOG_INFO, "Received resp to deprobe"); + + if (rsp.op_ret) { + if (strlen(rsp.op_errstr) > 0) { + snprintf(msg, sizeof(msg), "%s", rsp.op_errstr); + gf_log("cli", GF_LOG_ERROR, "%s", rsp.op_errstr); + } + } else { + snprintf(msg, sizeof(msg), "success"); + } + + if (global_state->mode & GLUSTER_MODE_XML) { + ret = cli_xml_output_str(NULL, (rsp.op_ret) ? NULL : msg, rsp.op_ret, + rsp.op_errno, (rsp.op_ret) ? msg : NULL); + if (ret) + gf_log("cli", GF_LOG_ERROR, "Error outputting to xml"); + goto out; + } - if (!rsp.op_ret) - cli_out ("peer detach: %s", msg); - else - cli_err ("peer detach: failed: %s", msg); + if (!rsp.op_ret) + cli_out("peer detach: %s", msg); + else + cli_err("peer detach: failed: %s", msg); - ret = rsp.op_ret; + ret = rsp.op_ret; out: - cli_cmd_broadcast_response (ret); - return ret; + cli_cmd_broadcast_response(ret); + return ret; } int -gf_cli_output_peer_hostnames (dict_t *dict, int count, char *prefix) -{ - int ret = -1; - char key[256] = {0,}; - int i = 0; - char *hostname = NULL; - - cli_out ("Other names:"); - /* Starting from friend.hostname1, as friend.hostname0 will be the same - * as friend.hostname - */ - for (i = 1; i < count; i++) { - snprintf (key, sizeof (key), "%s.hostname%d", prefix, i); - ret = dict_get_str (dict, key, &hostname); - if (ret) - break; - cli_out ("%s", hostname); - hostname = NULL; - } +gf_cli_output_peer_hostnames(dict_t *dict, int count, char *prefix) +{ + int ret = -1; + char key[256] = { + 0, + }; + int i = 0; + char *hostname = NULL; + + cli_out("Other names:"); + /* Starting from friend.hostname1, as friend.hostname0 will be the same + * as friend.hostname + */ + for (i = 1; i < count; i++) { + snprintf(key, sizeof(key), "%s.hostname%d", prefix, i); + ret = dict_get_str(dict, key, &hostname); + if (ret) + break; + cli_out("%s", hostname); + hostname = NULL; + } - return ret; + return ret; } int -gf_cli_output_peer_status (dict_t *dict, int count) -{ - int ret = -1; - char *uuid_buf = NULL; - char *hostname_buf = NULL; - int32_t i = 1; - char key[256] = {0,}; - char *state = NULL; - int32_t connected = 0; - char *connected_str = NULL; - int hostname_count = 0; - - cli_out ("Number of Peers: %d", count); - i = 1; - while ( i <= count) { - snprintf (key, 256, "friend%d.uuid", i); - ret = dict_get_str (dict, key, &uuid_buf); - if (ret) - goto out; - - snprintf (key, 256, "friend%d.hostname", i); - ret = dict_get_str (dict, key, &hostname_buf); - if (ret) - goto out; +gf_cli_output_peer_status(dict_t *dict, int count) +{ + int ret = -1; + char *uuid_buf = NULL; + char *hostname_buf = NULL; + int32_t i = 1; + char key[256] = { + 0, + }; + char *state = NULL; + int32_t connected = 0; + char *connected_str = NULL; + int hostname_count = 0; + + cli_out("Number of Peers: %d", count); + i = 1; + while (i <= count) { + snprintf(key, 256, "friend%d.uuid", i); + ret = dict_get_str(dict, key, &uuid_buf); + if (ret) + goto out; - snprintf (key, 256, "friend%d.connected", i); - ret = dict_get_int32 (dict, key, &connected); - if (ret) - goto out; - if (connected) - connected_str = "Connected"; - else - connected_str = "Disconnected"; + snprintf(key, 256, "friend%d.hostname", i); + ret = dict_get_str(dict, key, &hostname_buf); + if (ret) + goto out; + snprintf(key, 256, "friend%d.connected", i); + ret = dict_get_int32(dict, key, &connected); + if (ret) + goto out; + if (connected) + connected_str = "Connected"; + else + connected_str = "Disconnected"; - snprintf (key, 256, "friend%d.state", i); - ret = dict_get_str (dict, key, &state); - if (ret) - goto out; + snprintf(key, 256, "friend%d.state", i); + ret = dict_get_str(dict, key, &state); + if (ret) + goto out; - cli_out ("\nHostname: %s\nUuid: %s\nState: %s (%s)", - hostname_buf, uuid_buf, state, connected_str); + cli_out("\nHostname: %s\nUuid: %s\nState: %s (%s)", hostname_buf, + uuid_buf, state, connected_str); - snprintf (key, sizeof (key), "friend%d.hostname_count", i); - ret = dict_get_int32 (dict, key, &hostname_count); - /* Print other addresses only if there are more than 1. - */ - if ((ret == 0) && (hostname_count > 1)) { - snprintf (key, sizeof (key), "friend%d", i); - ret = gf_cli_output_peer_hostnames (dict, - hostname_count, - key); - if (ret) { - gf_log ("cli", GF_LOG_WARNING, - "error outputting peer other names"); - goto out; - } - } - i++; + snprintf(key, sizeof(key), "friend%d.hostname_count", i); + ret = dict_get_int32(dict, key, &hostname_count); + /* Print other addresses only if there are more than 1. + */ + if ((ret == 0) && (hostname_count > 1)) { + snprintf(key, sizeof(key), "friend%d", i); + ret = gf_cli_output_peer_hostnames(dict, hostname_count, key); + if (ret) { + gf_log("cli", GF_LOG_WARNING, + "error outputting peer other names"); + goto out; + } } + i++; + } - ret = 0; + ret = 0; out: - return ret; + return ret; } int -gf_cli_output_pool_list (dict_t *dict, int count) -{ - int ret = -1; - char *uuid_buf = NULL; - char *hostname_buf = NULL; - int32_t hostname_len = 8; /*min len 8 chars*/ - int32_t i = 1; - char key[256] = {0,}; - int32_t connected = 0; - char *connected_str = NULL; - - if (count <= 0) - goto out; - - while (i <= count) { - snprintf (key, 256, "friend%d.hostname", i); - ret = dict_get_str (dict, key, &hostname_buf); - if (ret) - goto out; +gf_cli_output_pool_list(dict_t *dict, int count) +{ + int ret = -1; + char *uuid_buf = NULL; + char *hostname_buf = NULL; + int32_t hostname_len = 8; /*min len 8 chars*/ + int32_t i = 1; + char key[256] = { + 0, + }; + int32_t connected = 0; + char *connected_str = NULL; + + if (count <= 0) + goto out; + + while (i <= count) { + snprintf(key, 256, "friend%d.hostname", i); + ret = dict_get_str(dict, key, &hostname_buf); + if (ret) + goto out; - ret = strlen(hostname_buf); - if (ret > hostname_len) - hostname_len = ret; + ret = strlen(hostname_buf); + if (ret > hostname_len) + hostname_len = ret; - i++; - } + i++; + } - cli_out ("UUID\t\t\t\t\t%-*s\tState", hostname_len, "Hostname"); + cli_out("UUID\t\t\t\t\t%-*s\tState", hostname_len, "Hostname"); - i = 1; - while ( i <= count) { - snprintf (key, 256, "friend%d.uuid", i); - ret = dict_get_str (dict, key, &uuid_buf); - if (ret) - goto out; + i = 1; + while (i <= count) { + snprintf(key, 256, "friend%d.uuid", i); + ret = dict_get_str(dict, key, &uuid_buf); + if (ret) + goto out; - snprintf (key, 256, "friend%d.hostname", i); - ret = dict_get_str (dict, key, &hostname_buf); - if (ret) - goto out; + snprintf(key, 256, "friend%d.hostname", i); + ret = dict_get_str(dict, key, &hostname_buf); + if (ret) + goto out; - snprintf (key, 256, "friend%d.connected", i); - ret = dict_get_int32 (dict, key, &connected); - if (ret) - goto out; - if (connected) - connected_str = "Connected"; - else - connected_str = "Disconnected"; + snprintf(key, 256, "friend%d.connected", i); + ret = dict_get_int32(dict, key, &connected); + if (ret) + goto out; + if (connected) + connected_str = "Connected"; + else + connected_str = "Disconnected"; - cli_out ("%s\t%-*s\t%s ", uuid_buf, hostname_len, hostname_buf, - connected_str); - i++; - } + cli_out("%s\t%-*s\t%s ", uuid_buf, hostname_len, hostname_buf, + connected_str); + i++; + } - ret = 0; + ret = 0; out: - return ret; + return ret; } /* function pointer for gf_cli_output_{pool_list,peer_status} */ -typedef int (*cli_friend_output_fn) (dict_t*, int); +typedef int (*cli_friend_output_fn)(dict_t *, int); int -gf_cli_list_friends_cbk (struct rpc_req *req, struct iovec *iov, - int count, void *myframe) +gf_cli_list_friends_cbk(struct rpc_req *req, struct iovec *iov, int count, + void *myframe) { - gf1_cli_peer_list_rsp rsp = {0,}; - int ret = -1; - dict_t *dict = NULL; - char msg[1024] = {0,}; - char *cmd = NULL; - cli_friend_output_fn friend_output_fn; - call_frame_t *frame = NULL; - unsigned long flags = 0; + gf1_cli_peer_list_rsp rsp = { + 0, + }; + int ret = -1; + dict_t *dict = NULL; + char msg[1024] = { + 0, + }; + char *cmd = NULL; + cli_friend_output_fn friend_output_fn; + call_frame_t *frame = NULL; + unsigned long flags = 0; - GF_ASSERT (myframe); + GF_ASSERT(myframe); - frame = myframe; + frame = myframe; - flags = (long)frame->local; + flags = (long)frame->local; - if (flags == GF_CLI_LIST_POOL_NODES) { - cmd = "pool list"; - friend_output_fn = &gf_cli_output_pool_list; - } else { - cmd = "peer status"; - friend_output_fn = &gf_cli_output_peer_status; - } + if (flags == GF_CLI_LIST_POOL_NODES) { + cmd = "pool list"; + friend_output_fn = &gf_cli_output_pool_list; + } else { + cmd = "peer status"; + friend_output_fn = &gf_cli_output_peer_status; + } - /* 'free' the flags set by gf_cli_list_friends */ - frame->local = NULL; + /* 'free' the flags set by gf_cli_list_friends */ + frame->local = NULL; - if (-1 == req->rpc_status) { - goto out; - } + if (-1 == req->rpc_status) { + goto out; + } - ret = xdr_to_generic (*iov, &rsp, (xdrproc_t)xdr_gf1_cli_peer_list_rsp); - if (ret < 0) { - gf_log (frame->this->name, GF_LOG_ERROR, - "Failed to decode xdr response"); - //rsp.op_ret = -1; - //rsp.op_errno = EINVAL; + ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf1_cli_peer_list_rsp); + if (ret < 0) { + gf_log(frame->this->name, GF_LOG_ERROR, + "Failed to decode xdr response"); + // rsp.op_ret = -1; + // rsp.op_errno = EINVAL; + goto out; + } + + gf_log("cli", GF_LOG_DEBUG, "Received resp to list: %d", rsp.op_ret); + + if (!rsp.op_ret) { + if (!rsp.friends.friends_len) { + snprintf(msg, sizeof(msg), "%s: No peers present", cmd); + if (global_state->mode & GLUSTER_MODE_XML) { + ret = cli_xml_output_peer_status(dict, rsp.op_ret, rsp.op_errno, + msg); + if (ret) + gf_log("cli", GF_LOG_ERROR, "Error outputting to xml"); goto out; + } + cli_err("%s", msg); + ret = 0; + goto out; } - gf_log ("cli", GF_LOG_DEBUG, "Received resp to list: %d", - rsp.op_ret); - - if (!rsp.op_ret) { - - if (!rsp.friends.friends_len) { - snprintf (msg, sizeof (msg), - "%s: No peers present", cmd); - if (global_state->mode & GLUSTER_MODE_XML) { - ret = cli_xml_output_peer_status (dict, - rsp.op_ret, - rsp.op_errno, - msg); - if (ret) - gf_log ("cli", GF_LOG_ERROR, - "Error outputting to xml"); - goto out; - } - cli_err ("%s", msg); - ret = 0; - goto out; - } - - dict = dict_new (); + dict = dict_new(); - if (!dict) { - ret = -1; - goto out; - } + if (!dict) { + ret = -1; + goto out; + } - ret = dict_unserialize (rsp.friends.friends_val, - rsp.friends.friends_len, - &dict); + ret = dict_unserialize(rsp.friends.friends_val, rsp.friends.friends_len, + &dict); - if (ret) { - gf_log ("", GF_LOG_ERROR, - "Unable to allocate memory"); - goto out; - } + if (ret) { + gf_log("", GF_LOG_ERROR, "Unable to allocate memory"); + goto out; + } - if (global_state->mode & GLUSTER_MODE_XML) { - ret = cli_xml_output_peer_status (dict, rsp.op_ret, - rsp.op_errno, msg); - if (ret) - gf_log ("cli", GF_LOG_ERROR, - "Error outputting to xml"); - goto out; - } + if (global_state->mode & GLUSTER_MODE_XML) { + ret = cli_xml_output_peer_status(dict, rsp.op_ret, rsp.op_errno, + msg); + if (ret) + gf_log("cli", GF_LOG_ERROR, "Error outputting to xml"); + goto out; + } - ret = dict_get_int32 (dict, "count", &count); - if (ret) { - goto out; - } + ret = dict_get_int32(dict, "count", &count); + if (ret) { + goto out; + } - ret = friend_output_fn (dict, count); - if (ret) { - goto out; - } + ret = friend_output_fn(dict, count); + if (ret) { + goto out; + } + } else { + if (global_state->mode & GLUSTER_MODE_XML) { + ret = cli_xml_output_peer_status(dict, rsp.op_ret, rsp.op_errno, + NULL); + if (ret) + gf_log("cli", GF_LOG_ERROR, "Error outputting to xml"); } else { - if (global_state->mode & GLUSTER_MODE_XML) { - ret = cli_xml_output_peer_status (dict, rsp.op_ret, - rsp.op_errno, NULL); - if (ret) - gf_log ("cli", GF_LOG_ERROR, - "Error outputting to xml"); - } else { - ret = -1; - } - goto out; + ret = -1; } + goto out; + } - - ret = 0; + ret = 0; out: - if (ret) - cli_err ("%s: failed", cmd); + if (ret) + cli_err("%s: failed", cmd); - cli_cmd_broadcast_response (ret); + cli_cmd_broadcast_response(ret); - if (dict) - dict_unref (dict); + if (dict) + dict_unref(dict); - return ret; + return ret; } int -gf_cli_get_state_cbk (struct rpc_req *req, struct iovec *iov, - int count, void *myframe) -{ - gf_cli_rsp rsp = {0,}; - int ret = -1; - dict_t *dict = NULL; - char *daemon_name = NULL; - char *ofilepath = NULL; - - GF_VALIDATE_OR_GOTO ("cli", myframe, out); - - if (-1 == req->rpc_status) { - goto out; - } - ret = xdr_to_generic (*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); - if (ret < 0) { - gf_log (((call_frame_t *) myframe)->this->name, GF_LOG_ERROR, - "Failed to decode xdr response"); - goto out; - } - - dict = dict_new (); - - if (!dict) { - ret = -1; - goto out; - } - - ret = dict_unserialize (rsp.dict.dict_val, rsp.dict.dict_len, &dict); +gf_cli_get_state_cbk(struct rpc_req *req, struct iovec *iov, int count, + void *myframe) +{ + gf_cli_rsp rsp = { + 0, + }; + int ret = -1; + dict_t *dict = NULL; + char *daemon_name = NULL; + char *ofilepath = NULL; + + GF_VALIDATE_OR_GOTO("cli", myframe, out); + + if (-1 == req->rpc_status) { + goto out; + } + ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); + if (ret < 0) { + gf_log(((call_frame_t *)myframe)->this->name, GF_LOG_ERROR, + "Failed to decode xdr response"); + goto out; + } + + dict = dict_new(); + + if (!dict) { + ret = -1; + goto out; + } + + ret = dict_unserialize(rsp.dict.dict_val, rsp.dict.dict_len, &dict); + if (ret) + goto out; + + if (rsp.op_ret) { + if (strcmp(rsp.op_errstr, "")) + cli_err("Failed to get daemon state: %s", rsp.op_errstr); + else + cli_err( + "Failed to get daemon state. Check glusterd" + " log file for more details"); + } else { + ret = dict_get_str(dict, "daemon", &daemon_name); if (ret) - goto out; + gf_log("cli", GF_LOG_ERROR, "Couldn't get daemon name"); - if (rsp.op_ret) { - if (strcmp (rsp.op_errstr, "")) - cli_err ("Failed to get daemon state: %s", rsp.op_errstr); - else - cli_err ("Failed to get daemon state. Check glusterd" - " log file for more details"); - } else { - ret = dict_get_str (dict, "daemon", &daemon_name); - if (ret) - gf_log ("cli", GF_LOG_ERROR, "Couldn't get daemon name"); - - ret = dict_get_str (dict, "ofilepath", &ofilepath); - if (ret) - gf_log ("cli", GF_LOG_ERROR, "Couldn't get filepath"); + ret = dict_get_str(dict, "ofilepath", &ofilepath); + if (ret) + gf_log("cli", GF_LOG_ERROR, "Couldn't get filepath"); - if (daemon_name && ofilepath) - cli_out ("%s state dumped to %s", - daemon_name, ofilepath); - } + if (daemon_name && ofilepath) + cli_out("%s state dumped to %s", daemon_name, ofilepath); + } - ret = rsp.op_ret; + ret = rsp.op_ret; out: - free (rsp.dict.dict_val); - free (rsp.op_errstr); + free(rsp.dict.dict_val); + free(rsp.op_errstr); - if (dict) - dict_unref (dict); + if (dict) + dict_unref(dict); - cli_cmd_broadcast_response (ret); + cli_cmd_broadcast_response(ret); - return ret; + return ret; } void -cli_out_options ( char *substr, char *optstr, char *valstr) +cli_out_options(char *substr, char *optstr, char *valstr) { - char *ptr1 = NULL; - char *ptr2 = NULL; + char *ptr1 = NULL; + char *ptr2 = NULL; - ptr1 = substr; - ptr2 = optstr; + ptr1 = substr; + ptr2 = optstr; - while (ptr1) - { - /* Avoiding segmentation fault. */ - if (!ptr2) - return; - if (*ptr1 != *ptr2) - break; - ptr1++; - ptr2++; - } + while (ptr1) { + /* Avoiding segmentation fault. */ + if (!ptr2) + return; + if (*ptr1 != *ptr2) + break; + ptr1++; + ptr2++; + } - if (*ptr2 == '\0') - return; - cli_out ("%s: %s",ptr2 , valstr); + if (*ptr2 == '\0') + return; + cli_out("%s: %s", ptr2, valstr); } static int -_gf_cli_output_volinfo_opts (dict_t *d, char *k, - data_t *v, void *tmp) -{ - int ret = 0; - char *key = NULL; - char *ptr = NULL; - data_t *value = NULL; - - key = tmp; - - ptr = strstr (k, "option."); - if (ptr) { - value = v; - if (!value) { - ret = -1; - goto out; - } - cli_out_options (key, k, v->data); +_gf_cli_output_volinfo_opts(dict_t *d, char *k, data_t *v, void *tmp) +{ + int ret = 0; + char *key = NULL; + char *ptr = NULL; + data_t *value = NULL; + + key = tmp; + + ptr = strstr(k, "option."); + if (ptr) { + value = v; + if (!value) { + ret = -1; + goto out; } + cli_out_options(key, k, v->data); + } out: - return ret; + return ret; } static int -print_brick_details (dict_t *dict, int volcount, int start_index, - int end_index, int replica_count) -{ - char key[1024] = {0,}; - int index = start_index; - int isArbiter = 0; - int ret = -1; - char *brick = NULL; +print_brick_details(dict_t *dict, int volcount, int start_index, int end_index, + int replica_count) +{ + char key[1024] = { + 0, + }; + int index = start_index; + int isArbiter = 0; + int ret = -1; + char *brick = NULL; #ifdef HAVE_BD_XLATOR - char *caps = NULL; + char *caps = NULL; #endif - while (index <= end_index) { - snprintf (key, 1024, "volume%d.brick%d", volcount, index); - ret = dict_get_str (dict, key, &brick); - if (ret) - goto out; - snprintf (key, sizeof (key), "volume%d.brick%d.isArbiter", - volcount, index); - if (dict_get (dict, key)) - isArbiter = 1; - else - isArbiter = 0; + while (index <= end_index) { + snprintf(key, 1024, "volume%d.brick%d", volcount, index); + ret = dict_get_str(dict, key, &brick); + if (ret) + goto out; + snprintf(key, sizeof(key), "volume%d.brick%d.isArbiter", volcount, + index); + if (dict_get(dict, key)) + isArbiter = 1; + else + isArbiter = 0; - if (isArbiter) - cli_out ("Brick%d: %s (arbiter)", index, brick); - else - cli_out ("Brick%d: %s", index, brick); + if (isArbiter) + cli_out("Brick%d: %s (arbiter)", index, brick); + else + cli_out("Brick%d: %s", index, brick); #ifdef HAVE_BD_XLATOR - snprintf (key, 1024, "volume%d.vg%d", volcount, index); - ret = dict_get_str (dict, key, &caps); - if (!ret) - cli_out ("Brick%d VG: %s", index, caps); + snprintf(key, 1024, "volume%d.vg%d", volcount, index); + ret = dict_get_str(dict, key, &caps); + if (!ret) + cli_out("Brick%d VG: %s", index, caps); #endif - index++; - } - ret = 0; + index++; + } + ret = 0; out: - return ret; + return ret; } void -gf_cli_print_number_of_bricks (int type, int brick_count, int dist_count, - int stripe_count, int replica_count, - int disperse_count, int redundancy_count, - int arbiter_count) -{ - if (type == GF_CLUSTER_TYPE_STRIPE_REPLICATE) { - if (arbiter_count == 0) { - cli_out ("Number of Bricks: %d x %d x %d = %d", - (brick_count / dist_count), - stripe_count, - replica_count, - brick_count); - } else { - cli_out ("Number of Bricks: %d x %d x (%d + %d) = %d", - (brick_count / dist_count), - stripe_count, replica_count - arbiter_count, - arbiter_count, brick_count); - } - } else if (type == GF_CLUSTER_TYPE_NONE || - type == GF_CLUSTER_TYPE_TIER) { - cli_out ("Number of Bricks: %d", brick_count); - } else if (type == GF_CLUSTER_TYPE_DISPERSE) { - cli_out ("Number of Bricks: %d x (%d + %d) = %d", - (brick_count / dist_count), - disperse_count - redundancy_count, - redundancy_count, brick_count); - } else { - /* For both replicate and stripe, dist_count is - good enough */ - if (arbiter_count == 0) { - cli_out ("Number of Bricks: %d x %d = %d", - (brick_count / dist_count), - dist_count, brick_count); - } else { - cli_out ("Number of Bricks: %d x (%d + %d) = %d", - (brick_count / dist_count), - dist_count - arbiter_count, arbiter_count, - brick_count); - } - } - +gf_cli_print_number_of_bricks(int type, int brick_count, int dist_count, + int stripe_count, int replica_count, + int disperse_count, int redundancy_count, + int arbiter_count) +{ + if (type == GF_CLUSTER_TYPE_STRIPE_REPLICATE) { + if (arbiter_count == 0) { + cli_out("Number of Bricks: %d x %d x %d = %d", + (brick_count / dist_count), stripe_count, replica_count, + brick_count); + } else { + cli_out("Number of Bricks: %d x %d x (%d + %d) = %d", + (brick_count / dist_count), stripe_count, + replica_count - arbiter_count, arbiter_count, brick_count); + } + } else if (type == GF_CLUSTER_TYPE_NONE || type == GF_CLUSTER_TYPE_TIER) { + cli_out("Number of Bricks: %d", brick_count); + } else if (type == GF_CLUSTER_TYPE_DISPERSE) { + cli_out("Number of Bricks: %d x (%d + %d) = %d", + (brick_count / dist_count), disperse_count - redundancy_count, + redundancy_count, brick_count); + } else { + /* For both replicate and stripe, dist_count is + good enough */ + if (arbiter_count == 0) { + cli_out("Number of Bricks: %d x %d = %d", + (brick_count / dist_count), dist_count, brick_count); + } else { + cli_out("Number of Bricks: %d x (%d + %d) = %d", + (brick_count / dist_count), dist_count - arbiter_count, + arbiter_count, brick_count); + } + } } int -gf_cli_print_tier_info (dict_t *dict, int i, int brick_count) -{ - - int hot_brick_count = -1; - int cold_type = 0; - int cold_brick_count = 0; - int cold_replica_count = 0; - int cold_arbiter_count = 0; - int cold_disperse_count = 0; - int cold_redundancy_count = 0; - int cold_dist_count = 0; - int hot_type = 0; - int hot_replica_count = 0; - int hot_dist_count = 0; - int ret = -1; - int vol_type = -1; - char key[256] = {0,}; - - GF_ASSERT (dict); - - snprintf (key, sizeof (key), "volume%d.cold_brick_count", i); - ret = dict_get_int32 (dict, key, &cold_brick_count); - if (ret) - goto out; - - snprintf (key, sizeof (key), "volume%d.cold_type", i); - ret = dict_get_int32 (dict, key, &cold_type); - if (ret) - goto out; - - snprintf (key, sizeof (key), "volume%d.cold_dist_count", i); - ret = dict_get_int32 (dict, key, &cold_dist_count); - if (ret) - goto out; - - snprintf (key, sizeof (key), "volume%d.cold_replica_count", i); - ret = dict_get_int32 (dict, key, &cold_replica_count); - if (ret) - goto out; - - snprintf (key, sizeof (key), "volume%d.cold_arbiter_count", i); - ret = dict_get_int32 (dict, key, &cold_arbiter_count); - if (ret) - goto out; - - snprintf (key, sizeof (key), "volume%d.cold_disperse_count", i); - ret = dict_get_int32 (dict, key, &cold_disperse_count); - if (ret) - goto out; - - snprintf (key, sizeof (key), - "volume%d.cold_redundancy_count", i); - ret = dict_get_int32 (dict, key, - &cold_redundancy_count); - if (ret) - goto out; - - snprintf (key, sizeof (key), "volume%d.hot_brick_count", i); - ret = dict_get_int32 (dict, key, &hot_brick_count); - if (ret) - goto out; - - snprintf (key, sizeof (key), "volume%d.hot_type", i); - ret = dict_get_int32 (dict, key, &hot_type); - if (ret) - goto out; - snprintf (key, sizeof (key), "volume%d.hot_replica_count", i); - ret = dict_get_int32 (dict, key, &hot_replica_count); - if (ret) - goto out; - - cli_out ("Hot Tier :"); - hot_dist_count = (hot_replica_count ? - hot_replica_count : 1); - - vol_type = get_vol_type (hot_type, hot_dist_count, hot_brick_count); - cli_out ("Hot Tier Type : %s", - vol_type_str[vol_type]); - - gf_cli_print_number_of_bricks (hot_type, - hot_brick_count, hot_dist_count, 0, - hot_replica_count, 0, 0, 0); - - ret = print_brick_details (dict, i, 1, hot_brick_count, - hot_replica_count); - if (ret) - goto out; - - cli_out ("Cold Tier:"); - - vol_type = get_vol_type (cold_type, cold_dist_count, cold_brick_count); - cli_out ("Cold Tier Type : %s", - vol_type_str[vol_type]); - - gf_cli_print_number_of_bricks (cold_type, - cold_brick_count, - cold_dist_count, 0, cold_replica_count, - cold_disperse_count, cold_redundancy_count, cold_arbiter_count); - - ret = print_brick_details (dict, i, hot_brick_count+1, - brick_count, cold_replica_count); - if (ret) - goto out; -out: - return ret; +gf_cli_print_tier_info(dict_t *dict, int i, int brick_count) +{ + int hot_brick_count = -1; + int cold_type = 0; + int cold_brick_count = 0; + int cold_replica_count = 0; + int cold_arbiter_count = 0; + int cold_disperse_count = 0; + int cold_redundancy_count = 0; + int cold_dist_count = 0; + int hot_type = 0; + int hot_replica_count = 0; + int hot_dist_count = 0; + int ret = -1; + int vol_type = -1; + char key[256] = { + 0, + }; + + GF_ASSERT(dict); + + snprintf(key, sizeof(key), "volume%d.cold_brick_count", i); + ret = dict_get_int32(dict, key, &cold_brick_count); + if (ret) + goto out; + + snprintf(key, sizeof(key), "volume%d.cold_type", i); + ret = dict_get_int32(dict, key, &cold_type); + if (ret) + goto out; + + snprintf(key, sizeof(key), "volume%d.cold_dist_count", i); + ret = dict_get_int32(dict, key, &cold_dist_count); + if (ret) + goto out; + + snprintf(key, sizeof(key), "volume%d.cold_replica_count", i); + ret = dict_get_int32(dict, key, &cold_replica_count); + if (ret) + goto out; + + snprintf(key, sizeof(key), "volume%d.cold_arbiter_count", i); + ret = dict_get_int32(dict, key, &cold_arbiter_count); + if (ret) + goto out; + + snprintf(key, sizeof(key), "volume%d.cold_disperse_count", i); + ret = dict_get_int32(dict, key, &cold_disperse_count); + if (ret) + goto out; + + snprintf(key, sizeof(key), "volume%d.cold_redundancy_count", i); + ret = dict_get_int32(dict, key, &cold_redundancy_count); + if (ret) + goto out; + + snprintf(key, sizeof(key), "volume%d.hot_brick_count", i); + ret = dict_get_int32(dict, key, &hot_brick_count); + if (ret) + goto out; + + snprintf(key, sizeof(key), "volume%d.hot_type", i); + ret = dict_get_int32(dict, key, &hot_type); + if (ret) + goto out; + snprintf(key, sizeof(key), "volume%d.hot_replica_count", i); + ret = dict_get_int32(dict, key, &hot_replica_count); + if (ret) + goto out; + + cli_out("Hot Tier :"); + hot_dist_count = (hot_replica_count ? hot_replica_count : 1); + + vol_type = get_vol_type(hot_type, hot_dist_count, hot_brick_count); + cli_out("Hot Tier Type : %s", vol_type_str[vol_type]); + + gf_cli_print_number_of_bricks(hot_type, hot_brick_count, hot_dist_count, 0, + hot_replica_count, 0, 0, 0); + + ret = print_brick_details(dict, i, 1, hot_brick_count, hot_replica_count); + if (ret) + goto out; + + cli_out("Cold Tier:"); + + vol_type = get_vol_type(cold_type, cold_dist_count, cold_brick_count); + cli_out("Cold Tier Type : %s", vol_type_str[vol_type]); + + gf_cli_print_number_of_bricks(cold_type, cold_brick_count, cold_dist_count, + 0, cold_replica_count, cold_disperse_count, + cold_redundancy_count, cold_arbiter_count); + + ret = print_brick_details(dict, i, hot_brick_count + 1, brick_count, + cold_replica_count); + if (ret) + goto out; +out: + return ret; } int -gf_cli_get_volume_cbk (struct rpc_req *req, struct iovec *iov, - int count, void *myframe) -{ - int ret = -1; - int opt_count = 0; - int32_t i = 0; - int32_t j = 1; - int32_t status = 0; - int32_t type = 0; - int32_t brick_count = 0; - int32_t dist_count = 0; - int32_t stripe_count = 0; - int32_t replica_count = 0; - int32_t disperse_count = 0; - int32_t redundancy_count = 0; - int32_t arbiter_count = 0; - int32_t snap_count = 0; - int32_t vol_type = 0; - int32_t transport = 0; - char *volume_id_str = NULL; - char *volname = NULL; - dict_t *dict = NULL; - cli_local_t *local = NULL; - char key[1024] = {0}; - char err_str[2048] = {0}; - gf_cli_rsp rsp = {0}; - char *caps __attribute__((unused)) = NULL; - int k __attribute__((unused)) = 0; - call_frame_t *frame = NULL; - - GF_ASSERT (myframe); - - if (-1 == req->rpc_status) - goto out; +gf_cli_get_volume_cbk(struct rpc_req *req, struct iovec *iov, int count, + void *myframe) +{ + int ret = -1; + int opt_count = 0; + int32_t i = 0; + int32_t j = 1; + int32_t status = 0; + int32_t type = 0; + int32_t brick_count = 0; + int32_t dist_count = 0; + int32_t stripe_count = 0; + int32_t replica_count = 0; + int32_t disperse_count = 0; + int32_t redundancy_count = 0; + int32_t arbiter_count = 0; + int32_t snap_count = 0; + int32_t vol_type = 0; + int32_t transport = 0; + char *volume_id_str = NULL; + char *volname = NULL; + dict_t *dict = NULL; + cli_local_t *local = NULL; + char key[1024] = {0}; + char err_str[2048] = {0}; + gf_cli_rsp rsp = {0}; + char *caps __attribute__((unused)) = NULL; + int k __attribute__((unused)) = 0; + call_frame_t *frame = NULL; + + GF_ASSERT(myframe); + + if (-1 == req->rpc_status) + goto out; + + frame = myframe; + + GF_ASSERT(frame->local); + + local = frame->local; + + ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); + if (ret < 0) { + gf_log(frame->this->name, GF_LOG_ERROR, + "Failed to decode xdr response"); + goto out; + } + + gf_log("cli", GF_LOG_INFO, "Received resp to get vol: %d", rsp.op_ret); + + if (!rsp.dict.dict_len) { + if (global_state->mode & GLUSTER_MODE_XML) + goto xml_output; + cli_err("No volumes present"); + ret = 0; + goto out; + } - frame = myframe; + dict = dict_new(); - GF_ASSERT (frame->local); + if (!dict) { + ret = -1; + goto out; + } - local = frame->local; + ret = dict_unserialize(rsp.dict.dict_val, rsp.dict.dict_len, &dict); - ret = xdr_to_generic (*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); - if (ret < 0) { - gf_log (frame->this->name, GF_LOG_ERROR, - "Failed to decode xdr response"); - goto out; - } + if (ret) { + gf_log("cli", GF_LOG_ERROR, "Unable to allocate memory"); + goto out; + } - gf_log ("cli", GF_LOG_INFO, "Received resp to get vol: %d", - rsp.op_ret); + ret = dict_get_int32(dict, "count", &count); + if (ret) + goto out; - if (!rsp.dict.dict_len) { - if (global_state->mode & GLUSTER_MODE_XML) - goto xml_output; - cli_err ("No volumes present"); + if (!count) { + switch (local->get_vol.flags) { + case GF_CLI_GET_NEXT_VOLUME: + GF_FREE(local->get_vol.volname); + local->get_vol.volname = NULL; ret = 0; goto out; - } - - dict = dict_new (); - if (!dict) { + case GF_CLI_GET_VOLUME: + snprintf(err_str, sizeof(err_str), "Volume %s does not exist", + local->get_vol.volname); ret = -1; - goto out; + if (!(global_state->mode & GLUSTER_MODE_XML)) + goto out; } + } - ret = dict_unserialize (rsp.dict.dict_val, - rsp.dict.dict_len, - &dict); - - if (ret) { - gf_log ("cli", GF_LOG_ERROR, - "Unable to allocate memory"); - goto out; - } + if (rsp.op_ret) { + if (global_state->mode & GLUSTER_MODE_XML) + goto xml_output; + ret = -1; + goto out; + } - ret = dict_get_int32 (dict, "count", &count); - if (ret) +xml_output: + if (global_state->mode & GLUSTER_MODE_XML) { + /* For GET_NEXT_VOLUME output is already begun in + * and will also end in gf_cli_get_next_volume() + */ + if (local->get_vol.flags == GF_CLI_GET_VOLUME) { + ret = cli_xml_output_vol_info_begin(local, rsp.op_ret, rsp.op_errno, + rsp.op_errstr); + if (ret) { + gf_log("cli", GF_LOG_ERROR, "Error outputting to xml"); goto out; - - if (!count) { - switch (local->get_vol.flags) { - - case GF_CLI_GET_NEXT_VOLUME: - GF_FREE (local->get_vol.volname); - local->get_vol.volname = NULL; - ret = 0; - goto out; - - case GF_CLI_GET_VOLUME: - snprintf (err_str, sizeof (err_str), - "Volume %s does not exist", - local->get_vol.volname); - ret = -1; - if (!(global_state->mode & GLUSTER_MODE_XML)) - goto out; - } + } } - if (rsp.op_ret) { - if (global_state->mode & GLUSTER_MODE_XML) - goto xml_output; - ret = -1; + if (dict) { + ret = cli_xml_output_vol_info(local, dict); + if (ret) { + gf_log("cli", GF_LOG_ERROR, "Error outputting to xml"); goto out; + } } -xml_output: - if (global_state->mode & GLUSTER_MODE_XML) { - /* For GET_NEXT_VOLUME output is already begun in - * and will also end in gf_cli_get_next_volume() - */ - if (local->get_vol.flags == GF_CLI_GET_VOLUME) { - ret = cli_xml_output_vol_info_begin - (local, rsp.op_ret, rsp.op_errno, - rsp.op_errstr); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, - "Error outputting to xml"); - goto out; - } - } - - if (dict) { - ret = cli_xml_output_vol_info (local, dict); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, - "Error outputting to xml"); - goto out; - } - } - - if (local->get_vol.flags == GF_CLI_GET_VOLUME) { - ret = cli_xml_output_vol_info_end (local); - if (ret) - gf_log ("cli", GF_LOG_ERROR, - "Error outputting to xml"); - } - goto out; + if (local->get_vol.flags == GF_CLI_GET_VOLUME) { + ret = cli_xml_output_vol_info_end(local); + if (ret) + gf_log("cli", GF_LOG_ERROR, "Error outputting to xml"); } + goto out; + } - while ( i < count) { - cli_out (" "); - snprintf (key, 256, "volume%d.name", i); - ret = dict_get_str (dict, key, &volname); - if (ret) - goto out; + while (i < count) { + cli_out(" "); + snprintf(key, 256, "volume%d.name", i); + ret = dict_get_str(dict, key, &volname); + if (ret) + goto out; - snprintf (key, 256, "volume%d.type", i); - ret = dict_get_int32 (dict, key, &type); - if (ret) - goto out; + snprintf(key, 256, "volume%d.type", i); + ret = dict_get_int32(dict, key, &type); + if (ret) + goto out; - snprintf (key, 256, "volume%d.status", i); - ret = dict_get_int32 (dict, key, &status); - if (ret) - goto out; + snprintf(key, 256, "volume%d.status", i); + ret = dict_get_int32(dict, key, &status); + if (ret) + goto out; - snprintf (key, 256, "volume%d.brick_count", i); - ret = dict_get_int32 (dict, key, &brick_count); - if (ret) - goto out; + snprintf(key, 256, "volume%d.brick_count", i); + ret = dict_get_int32(dict, key, &brick_count); + if (ret) + goto out; - snprintf (key, 256, "volume%d.dist_count", i); - ret = dict_get_int32 (dict, key, &dist_count); - if (ret) - goto out; + snprintf(key, 256, "volume%d.dist_count", i); + ret = dict_get_int32(dict, key, &dist_count); + if (ret) + goto out; - snprintf (key, 256, "volume%d.stripe_count", i); - ret = dict_get_int32 (dict, key, &stripe_count); - if (ret) - goto out; + snprintf(key, 256, "volume%d.stripe_count", i); + ret = dict_get_int32(dict, key, &stripe_count); + if (ret) + goto out; - snprintf (key, 256, "volume%d.replica_count", i); - ret = dict_get_int32 (dict, key, &replica_count); - if (ret) - goto out; + snprintf(key, 256, "volume%d.replica_count", i); + ret = dict_get_int32(dict, key, &replica_count); + if (ret) + goto out; - snprintf (key, 256, "volume%d.disperse_count", i); - ret = dict_get_int32 (dict, key, &disperse_count); - if (ret) - goto out; + snprintf(key, 256, "volume%d.disperse_count", i); + ret = dict_get_int32(dict, key, &disperse_count); + if (ret) + goto out; - snprintf (key, 256, "volume%d.redundancy_count", i); - ret = dict_get_int32 (dict, key, &redundancy_count); - if (ret) - goto out; + snprintf(key, 256, "volume%d.redundancy_count", i); + ret = dict_get_int32(dict, key, &redundancy_count); + if (ret) + goto out; - snprintf (key, sizeof(key), "volume%d.arbiter_count", i); - ret = dict_get_int32 (dict, key, &arbiter_count); - if (ret) - goto out; + snprintf(key, sizeof(key), "volume%d.arbiter_count", i); + ret = dict_get_int32(dict, key, &arbiter_count); + if (ret) + goto out; - snprintf (key, 256, "volume%d.transport", i); - ret = dict_get_int32 (dict, key, &transport); - if (ret) - goto out; + snprintf(key, 256, "volume%d.transport", i); + ret = dict_get_int32(dict, key, &transport); + if (ret) + goto out; - snprintf (key, 256, "volume%d.volume_id", i); - ret = dict_get_str (dict, key, &volume_id_str); - if (ret) - goto out; + snprintf(key, 256, "volume%d.volume_id", i); + ret = dict_get_str(dict, key, &volume_id_str); + if (ret) + goto out; - snprintf (key, 256, "volume%d.snap_count", i); - ret = dict_get_int32 (dict, key, &snap_count); - if (ret) - goto out; + snprintf(key, 256, "volume%d.snap_count", i); + ret = dict_get_int32(dict, key, &snap_count); + if (ret) + goto out; - // Distributed (stripe/replicate/stripe-replica) setups - vol_type = get_vol_type (type, dist_count, brick_count); + // Distributed (stripe/replicate/stripe-replica) setups + vol_type = get_vol_type(type, dist_count, brick_count); - cli_out ("Volume Name: %s", volname); - cli_out ("Type: %s", vol_type_str[vol_type]); - cli_out ("Volume ID: %s", volume_id_str); - cli_out ("Status: %s", cli_vol_status_str[status]); - cli_out ("Snapshot Count: %d", snap_count); + cli_out("Volume Name: %s", volname); + cli_out("Type: %s", vol_type_str[vol_type]); + cli_out("Volume ID: %s", volume_id_str); + cli_out("Status: %s", cli_vol_status_str[status]); + cli_out("Snapshot Count: %d", snap_count); #ifdef HAVE_BD_XLATOR - k = 0; - snprintf (key, sizeof (key), "volume%d.xlator%d", i, k); - ret = dict_get_str (dict, key, &caps); - if (ret) - goto next; - do { - j = 0; - cli_out ("Xlator %d: %s", k + 1, caps); - do { - snprintf (key, sizeof (key), - "volume%d.xlator%d.caps%d", - i, k, j++); - ret = dict_get_str (dict, key, &caps); - if (ret) - break; - cli_out ("Capability %d: %s", j, caps); - } while (1); - - snprintf (key, sizeof (key), - "volume%d.xlator%d", i, ++k); - ret = dict_get_str (dict, key, &caps); - if (ret) - break; - } while (1); - -next: -#endif - gf_cli_print_number_of_bricks (type, brick_count, - dist_count, stripe_count, replica_count, - disperse_count, redundancy_count, - arbiter_count); + k = 0; + snprintf(key, sizeof(key), "volume%d.xlator%d", i, k); + ret = dict_get_str(dict, key, &caps); + if (ret) + goto next; + do { + j = 0; + cli_out("Xlator %d: %s", k + 1, caps); + do { + snprintf(key, sizeof(key), "volume%d.xlator%d.caps%d", i, k, + j++); + ret = dict_get_str(dict, key, &caps); + if (ret) + break; + cli_out("Capability %d: %s", j, caps); + } while (1); + + snprintf(key, sizeof(key), "volume%d.xlator%d", i, ++k); + ret = dict_get_str(dict, key, &caps); + if (ret) + break; + } while (1); - cli_out ("Transport-type: %s", - ((transport == 0)?"tcp": - (transport == 1)?"rdma": - "tcp,rdma")); - j = 1; + next: +#endif + gf_cli_print_number_of_bricks( + type, brick_count, dist_count, stripe_count, replica_count, + disperse_count, redundancy_count, arbiter_count); - GF_FREE (local->get_vol.volname); - local->get_vol.volname = gf_strdup (volname); + cli_out("Transport-type: %s", + ((transport == 0) ? "tcp" + : (transport == 1) ? "rdma" : "tcp,rdma")); + j = 1; - if (type == GF_CLUSTER_TYPE_TIER) { - ret = gf_cli_print_tier_info (dict, i, brick_count); - if (ret) - goto out; + GF_FREE(local->get_vol.volname); + local->get_vol.volname = gf_strdup(volname); - } else { - cli_out ("Bricks:"); - ret = print_brick_details (dict, i, j, brick_count, - replica_count); - if (ret) - goto out; - } + if (type == GF_CLUSTER_TYPE_TIER) { + ret = gf_cli_print_tier_info(dict, i, brick_count); + if (ret) + goto out; - snprintf (key, 256, "volume%d.opt_count",i); - ret = dict_get_int32 (dict, key, &opt_count); - if (ret) - goto out; + } else { + cli_out("Bricks:"); + ret = print_brick_details(dict, i, j, brick_count, replica_count); + if (ret) + goto out; + } - if (!opt_count) - goto out; + snprintf(key, 256, "volume%d.opt_count", i); + ret = dict_get_int32(dict, key, &opt_count); + if (ret) + goto out; - cli_out ("Options Reconfigured:"); + if (!opt_count) + goto out; - snprintf (key, 256, "volume%d.option.",i); + cli_out("Options Reconfigured:"); - ret = dict_foreach (dict, _gf_cli_output_volinfo_opts, key); - if (ret) - goto out; + snprintf(key, 256, "volume%d.option.", i); - i++; - } + ret = dict_foreach(dict, _gf_cli_output_volinfo_opts, key); + if (ret) + goto out; + i++; + } - ret = 0; + ret = 0; out: - if (ret) - cli_err ("%s", err_str); + if (ret) + cli_err("%s", err_str); - cli_cmd_broadcast_response (ret); + cli_cmd_broadcast_response(ret); - if (dict) - dict_unref (dict); + if (dict) + dict_unref(dict); - free (rsp.dict.dict_val); + free(rsp.dict.dict_val); - free (rsp.op_errstr); + free(rsp.op_errstr); - gf_log ("cli", GF_LOG_DEBUG, "Returning: %d", ret); - return ret; + gf_log("cli", GF_LOG_DEBUG, "Returning: %d", ret); + return ret; } int -gf_cli_create_volume_cbk (struct rpc_req *req, struct iovec *iov, - int count, void *myframe) +gf_cli_create_volume_cbk(struct rpc_req *req, struct iovec *iov, int count, + void *myframe) { - gf_cli_rsp rsp = {0,}; - int ret = -1; - cli_local_t *local = NULL; - char *volname = NULL; - dict_t *rsp_dict = NULL; - call_frame_t *frame = NULL; - - GF_ASSERT (myframe); + gf_cli_rsp rsp = { + 0, + }; + int ret = -1; + cli_local_t *local = NULL; + char *volname = NULL; + dict_t *rsp_dict = NULL; + call_frame_t *frame = NULL; - if (-1 == req->rpc_status) { - goto out; - } + GF_ASSERT(myframe); - frame = myframe; + if (-1 == req->rpc_status) { + goto out; + } - GF_ASSERT (frame->local); + frame = myframe; - local = frame->local; + GF_ASSERT(frame->local); - ret = xdr_to_generic (*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); - if (ret < 0) { - gf_log (frame->this->name, GF_LOG_ERROR, - "Failed to decode xdr response"); - goto out; - } + local = frame->local; - gf_log ("cli", GF_LOG_INFO, "Received resp to create volume"); + ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); + if (ret < 0) { + gf_log(frame->this->name, GF_LOG_ERROR, + "Failed to decode xdr response"); + goto out; + } - ret = dict_get_str (local->dict, "volname", &volname); - if (ret) - goto out; + gf_log("cli", GF_LOG_INFO, "Received resp to create volume"); - if (global_state->mode & GLUSTER_MODE_XML) { - if (rsp.op_ret == 0) { - rsp_dict = dict_new (); - ret = dict_unserialize (rsp.dict.dict_val, - rsp.dict.dict_len, - &rsp_dict); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, - "Failed rsp_dict unserialization"); - goto out; - } - } + ret = dict_get_str(local->dict, "volname", &volname); + if (ret) + goto out; - ret = cli_xml_output_vol_create (rsp_dict, rsp.op_ret, - rsp.op_errno, rsp.op_errstr); - if (ret) - gf_log ("cli", GF_LOG_ERROR, - "Error outputting to xml"); + if (global_state->mode & GLUSTER_MODE_XML) { + if (rsp.op_ret == 0) { + rsp_dict = dict_new(); + ret = dict_unserialize(rsp.dict.dict_val, rsp.dict.dict_len, + &rsp_dict); + if (ret) { + gf_log("cli", GF_LOG_ERROR, "Failed rsp_dict unserialization"); goto out; + } } - if (rsp.op_ret && strcmp (rsp.op_errstr, "")) - cli_err ("volume create: %s: failed: %s", volname, - rsp.op_errstr); - else if (rsp.op_ret) - cli_err ("volume create: %s: failed", volname); - else - cli_out ("volume create: %s: success: " - "please start the volume to access data", volname); + ret = cli_xml_output_vol_create(rsp_dict, rsp.op_ret, rsp.op_errno, + rsp.op_errstr); + if (ret) + gf_log("cli", GF_LOG_ERROR, "Error outputting to xml"); + goto out; + } - ret = rsp.op_ret; + if (rsp.op_ret && strcmp(rsp.op_errstr, "")) + cli_err("volume create: %s: failed: %s", volname, rsp.op_errstr); + else if (rsp.op_ret) + cli_err("volume create: %s: failed", volname); + else + cli_out( + "volume create: %s: success: " + "please start the volume to access data", + volname); + + ret = rsp.op_ret; out: - cli_cmd_broadcast_response (ret); - free (rsp.dict.dict_val); - free (rsp.op_errstr); - return ret; + cli_cmd_broadcast_response(ret); + free(rsp.dict.dict_val); + free(rsp.op_errstr); + return ret; } int -gf_cli_delete_volume_cbk (struct rpc_req *req, struct iovec *iov, - int count, void *myframe) +gf_cli_delete_volume_cbk(struct rpc_req *req, struct iovec *iov, int count, + void *myframe) { - gf_cli_rsp rsp = {0,}; - int ret = -1; - cli_local_t *local = NULL; - char *volname = NULL; - call_frame_t *frame = NULL; - dict_t *rsp_dict = NULL; + gf_cli_rsp rsp = { + 0, + }; + int ret = -1; + cli_local_t *local = NULL; + char *volname = NULL; + call_frame_t *frame = NULL; + dict_t *rsp_dict = NULL; - GF_ASSERT (myframe); + GF_ASSERT(myframe); - if (-1 == req->rpc_status) { - goto out; - } + if (-1 == req->rpc_status) { + goto out; + } - frame = myframe; + frame = myframe; - GF_ASSERT (frame->local); + GF_ASSERT(frame->local); - local = frame->local; + local = frame->local; - ret = xdr_to_generic (*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); - if (ret < 0) { - gf_log (frame->this->name, GF_LOG_ERROR, - "Failed to decode xdr response"); - goto out; - } + ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); + if (ret < 0) { + gf_log(frame->this->name, GF_LOG_ERROR, + "Failed to decode xdr response"); + goto out; + } - ret = dict_get_str (local->dict, "volname", &volname); - if (ret) { - gf_log (frame->this->name, GF_LOG_ERROR, - "dict get failed"); - goto out; - } + ret = dict_get_str(local->dict, "volname", &volname); + if (ret) { + gf_log(frame->this->name, GF_LOG_ERROR, "dict get failed"); + goto out; + } - gf_log ("cli", GF_LOG_INFO, "Received resp to delete volume"); + gf_log("cli", GF_LOG_INFO, "Received resp to delete volume"); - if (global_state->mode & GLUSTER_MODE_XML) { - if (rsp.op_ret == 0) { - rsp_dict = dict_new (); - ret = dict_unserialize (rsp.dict.dict_val, - rsp.dict.dict_len, - &rsp_dict); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, - "Failed rsp_dict unserialization"); - goto out; - } - } - - ret = cli_xml_output_generic_volume ("volDelete", rsp_dict, - rsp.op_ret, rsp.op_errno, - rsp.op_errstr); - if (ret) - gf_log ("cli", GF_LOG_ERROR, - "Error outputting to xml"); + if (global_state->mode & GLUSTER_MODE_XML) { + if (rsp.op_ret == 0) { + rsp_dict = dict_new(); + ret = dict_unserialize(rsp.dict.dict_val, rsp.dict.dict_len, + &rsp_dict); + if (ret) { + gf_log("cli", GF_LOG_ERROR, "Failed rsp_dict unserialization"); goto out; + } } - if (rsp.op_ret && strcmp (rsp.op_errstr, "")) - cli_err ("volume delete: %s: failed: %s", volname, - rsp.op_errstr); - else if (rsp.op_ret) - cli_err ("volume delete: %s: failed", volname); - else - cli_out ("volume delete: %s: success", volname); + ret = cli_xml_output_generic_volume("volDelete", rsp_dict, rsp.op_ret, + rsp.op_errno, rsp.op_errstr); + if (ret) + gf_log("cli", GF_LOG_ERROR, "Error outputting to xml"); + goto out; + } - ret = rsp.op_ret; + if (rsp.op_ret && strcmp(rsp.op_errstr, "")) + cli_err("volume delete: %s: failed: %s", volname, rsp.op_errstr); + else if (rsp.op_ret) + cli_err("volume delete: %s: failed", volname); + else + cli_out("volume delete: %s: success", volname); + + ret = rsp.op_ret; out: - cli_cmd_broadcast_response (ret); - free (rsp.dict.dict_val); + cli_cmd_broadcast_response(ret); + free(rsp.dict.dict_val); - gf_log ("", GF_LOG_DEBUG, "Returning with %d", ret); - return ret; + gf_log("", GF_LOG_DEBUG, "Returning with %d", ret); + return ret; } int -gf_cli3_1_uuid_get_cbk (struct rpc_req *req, struct iovec *iov, - int count, void *myframe) -{ - char *uuid_str = NULL; - gf_cli_rsp rsp = {0,}; - int ret = -1; - cli_local_t *local = NULL; - call_frame_t *frame = NULL; - dict_t *dict = NULL; - - GF_ASSERT (myframe); - - if (-1 == req->rpc_status) - goto out; - - frame = myframe; - - GF_ASSERT (frame->local); - - local = frame->local; - - ret = xdr_to_generic (*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); - if (ret < 0) { - gf_log (frame->this->name, GF_LOG_ERROR, - "Failed to decode xdr response"); - goto out; - } - - frame->local = NULL; - - gf_log ("cli", GF_LOG_INFO, "Received resp to uuid get"); - - dict = dict_new (); - if (!dict) { - ret = -1; - goto out; - } - - ret = dict_unserialize (rsp.dict.dict_val, rsp.dict.dict_len, - &dict); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Failed to unserialize " - "response for uuid get"); - goto out; - } - - ret = dict_get_str (dict, "uuid", &uuid_str); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Failed to get uuid " - "from dictionary"); - goto out; - } - - if (global_state->mode & GLUSTER_MODE_XML) { - ret = cli_xml_output_dict ("uuidGenerate", dict, rsp.op_ret, - rsp.op_errno, rsp.op_errstr); - if (ret) - gf_log ("cli", GF_LOG_ERROR, - "Error outputting to xml"); - goto out; - } - - if (rsp.op_ret) { - if (strcmp (rsp.op_errstr, "") == 0) - cli_err ("Get uuid was unsuccessful"); - else - cli_err ("%s", rsp.op_errstr); +gf_cli3_1_uuid_get_cbk(struct rpc_req *req, struct iovec *iov, int count, + void *myframe) +{ + char *uuid_str = NULL; + gf_cli_rsp rsp = { + 0, + }; + int ret = -1; + cli_local_t *local = NULL; + call_frame_t *frame = NULL; + dict_t *dict = NULL; + + GF_ASSERT(myframe); + + if (-1 == req->rpc_status) + goto out; + + frame = myframe; + + GF_ASSERT(frame->local); + + local = frame->local; + + ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); + if (ret < 0) { + gf_log(frame->this->name, GF_LOG_ERROR, + "Failed to decode xdr response"); + goto out; + } + + frame->local = NULL; + + gf_log("cli", GF_LOG_INFO, "Received resp to uuid get"); + + dict = dict_new(); + if (!dict) { + ret = -1; + goto out; + } + + ret = dict_unserialize(rsp.dict.dict_val, rsp.dict.dict_len, &dict); + if (ret) { + gf_log("cli", GF_LOG_ERROR, + "Failed to unserialize " + "response for uuid get"); + goto out; + } + + ret = dict_get_str(dict, "uuid", &uuid_str); + if (ret) { + gf_log("cli", GF_LOG_ERROR, + "Failed to get uuid " + "from dictionary"); + goto out; + } + + if (global_state->mode & GLUSTER_MODE_XML) { + ret = cli_xml_output_dict("uuidGenerate", dict, rsp.op_ret, + rsp.op_errno, rsp.op_errstr); + if (ret) + gf_log("cli", GF_LOG_ERROR, "Error outputting to xml"); + goto out; + } - } else { - cli_out ("UUID: %s", uuid_str); + if (rsp.op_ret) { + if (strcmp(rsp.op_errstr, "") == 0) + cli_err("Get uuid was unsuccessful"); + else + cli_err("%s", rsp.op_errstr); - } - ret = rsp.op_ret; + } else { + cli_out("UUID: %s", uuid_str); + } + ret = rsp.op_ret; out: - cli_cmd_broadcast_response (ret); - cli_local_wipe (local); - if (rsp.dict.dict_val) - free (rsp.dict.dict_val); + cli_cmd_broadcast_response(ret); + cli_local_wipe(local); + if (rsp.dict.dict_val) + free(rsp.dict.dict_val); - gf_log ("", GF_LOG_DEBUG, "Returning with %d", ret); - return ret; + gf_log("", GF_LOG_DEBUG, "Returning with %d", ret); + return ret; } int -gf_cli3_1_uuid_reset_cbk (struct rpc_req *req, struct iovec *iov, - int count, void *myframe) +gf_cli3_1_uuid_reset_cbk(struct rpc_req *req, struct iovec *iov, int count, + void *myframe) { - gf_cli_rsp rsp = {0,}; - int ret = -1; - cli_local_t *local = NULL; - call_frame_t *frame = NULL; + gf_cli_rsp rsp = { + 0, + }; + int ret = -1; + cli_local_t *local = NULL; + call_frame_t *frame = NULL; - GF_ASSERT (myframe); + GF_ASSERT(myframe); - if (-1 == req->rpc_status) { - goto out; - } + if (-1 == req->rpc_status) { + goto out; + } - frame = myframe; + frame = myframe; - GF_ASSERT (frame->local); + GF_ASSERT(frame->local); - local = frame->local; + local = frame->local; - ret = xdr_to_generic (*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); - if (ret < 0) { - gf_log (frame->this->name, GF_LOG_ERROR, - "Failed to decode xdr response"); - goto out; - } + ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); + if (ret < 0) { + gf_log(frame->this->name, GF_LOG_ERROR, + "Failed to decode xdr response"); + goto out; + } - frame->local = NULL; + frame->local = NULL; - gf_log ("cli", GF_LOG_INFO, "Received resp to uuid reset"); + gf_log("cli", GF_LOG_INFO, "Received resp to uuid reset"); - if (global_state->mode & GLUSTER_MODE_XML) { - ret = cli_xml_output_dict ("uuidReset", NULL, rsp.op_ret, - rsp.op_errno, rsp.op_errstr); - if (ret) - gf_log ("cli", GF_LOG_ERROR, - "Error outputting to xml"); - goto out; - } + if (global_state->mode & GLUSTER_MODE_XML) { + ret = cli_xml_output_dict("uuidReset", NULL, rsp.op_ret, rsp.op_errno, + rsp.op_errstr); + if (ret) + gf_log("cli", GF_LOG_ERROR, "Error outputting to xml"); + goto out; + } - if (rsp.op_ret && strcmp (rsp.op_errstr, "")) - cli_err ("%s", rsp.op_errstr); - else - cli_out ("resetting the peer uuid has been %s", - (rsp.op_ret) ? "unsuccessful": "successful"); - ret = rsp.op_ret; + if (rsp.op_ret && strcmp(rsp.op_errstr, "")) + cli_err("%s", rsp.op_errstr); + else + cli_out("resetting the peer uuid has been %s", + (rsp.op_ret) ? "unsuccessful" : "successful"); + ret = rsp.op_ret; out: - cli_cmd_broadcast_response (ret); - cli_local_wipe (local); - if (rsp.dict.dict_val) - free (rsp.dict.dict_val); + cli_cmd_broadcast_response(ret); + cli_local_wipe(local); + if (rsp.dict.dict_val) + free(rsp.dict.dict_val); - gf_log ("", GF_LOG_DEBUG, "Returning with %d", ret); - return ret; + gf_log("", GF_LOG_DEBUG, "Returning with %d", ret); + return ret; } int -gf_cli_start_volume_cbk (struct rpc_req *req, struct iovec *iov, - int count, void *myframe) +gf_cli_start_volume_cbk(struct rpc_req *req, struct iovec *iov, int count, + void *myframe) { - gf_cli_rsp rsp = {0,}; - int ret = -1; - cli_local_t *local = NULL; - char *volname = NULL; - call_frame_t *frame = NULL; - dict_t *rsp_dict = NULL; + gf_cli_rsp rsp = { + 0, + }; + int ret = -1; + cli_local_t *local = NULL; + char *volname = NULL; + call_frame_t *frame = NULL; + dict_t *rsp_dict = NULL; - GF_ASSERT (myframe); + GF_ASSERT(myframe); - if (-1 == req->rpc_status) { - goto out; - } + if (-1 == req->rpc_status) { + goto out; + } - frame = myframe; + frame = myframe; - GF_ASSERT (frame->local); + GF_ASSERT(frame->local); - local = frame->local; + local = frame->local; - ret = xdr_to_generic (*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); - if (ret < 0) { - gf_log (frame->this->name, GF_LOG_ERROR, - "Failed to decode xdr response"); - goto out; - } + ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); + if (ret < 0) { + gf_log(frame->this->name, GF_LOG_ERROR, + "Failed to decode xdr response"); + goto out; + } - ret = dict_get_str (local->dict, "volname", &volname); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, "dict get failed"); - goto out; - } + ret = dict_get_str(local->dict, "volname", &volname); + if (ret) { + gf_log("cli", GF_LOG_ERROR, "dict get failed"); + goto out; + } - gf_log ("cli", GF_LOG_INFO, "Received resp to start volume"); - - if (global_state->mode & GLUSTER_MODE_XML) { - if (rsp.op_ret == 0) { - rsp_dict = dict_new (); - ret = dict_unserialize (rsp.dict.dict_val, - rsp.dict.dict_len, - &rsp_dict); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, - "Failed rsp_dict unserialization"); - goto out; - } - } + gf_log("cli", GF_LOG_INFO, "Received resp to start volume"); - ret = cli_xml_output_generic_volume ("volStart", rsp_dict, - rsp.op_ret, rsp.op_errno, - rsp.op_errstr); - if (ret) - gf_log ("cli", GF_LOG_ERROR, - "Error outputting to xml"); + if (global_state->mode & GLUSTER_MODE_XML) { + if (rsp.op_ret == 0) { + rsp_dict = dict_new(); + ret = dict_unserialize(rsp.dict.dict_val, rsp.dict.dict_len, + &rsp_dict); + if (ret) { + gf_log("cli", GF_LOG_ERROR, "Failed rsp_dict unserialization"); goto out; + } } - if (rsp.op_ret && strcmp (rsp.op_errstr, "")) - cli_err ("volume start: %s: failed: %s", volname, - rsp.op_errstr); - else if (rsp.op_ret) - cli_err ("volume start: %s: failed", volname); - else - cli_out ("volume start: %s: success", volname); + ret = cli_xml_output_generic_volume("volStart", rsp_dict, rsp.op_ret, + rsp.op_errno, rsp.op_errstr); + if (ret) + gf_log("cli", GF_LOG_ERROR, "Error outputting to xml"); + goto out; + } - ret = rsp.op_ret; + if (rsp.op_ret && strcmp(rsp.op_errstr, "")) + cli_err("volume start: %s: failed: %s", volname, rsp.op_errstr); + else if (rsp.op_ret) + cli_err("volume start: %s: failed", volname); + else + cli_out("volume start: %s: success", volname); + + ret = rsp.op_ret; out: - cli_cmd_broadcast_response (ret); - free (rsp.dict.dict_val); - free (rsp.op_errstr); - return ret; + cli_cmd_broadcast_response(ret); + free(rsp.dict.dict_val); + free(rsp.op_errstr); + return ret; } int -gf_cli_stop_volume_cbk (struct rpc_req *req, struct iovec *iov, - int count, void *myframe) +gf_cli_stop_volume_cbk(struct rpc_req *req, struct iovec *iov, int count, + void *myframe) { - gf_cli_rsp rsp = {0,}; - int ret = -1; - cli_local_t *local = NULL; - char *volname = NULL; - call_frame_t *frame = NULL; - dict_t *rsp_dict = NULL; + gf_cli_rsp rsp = { + 0, + }; + int ret = -1; + cli_local_t *local = NULL; + char *volname = NULL; + call_frame_t *frame = NULL; + dict_t *rsp_dict = NULL; - GF_ASSERT (myframe); + GF_ASSERT(myframe); - if (-1 == req->rpc_status) { - goto out; - } + if (-1 == req->rpc_status) { + goto out; + } - frame = myframe; + frame = myframe; - GF_ASSERT (frame->local); + GF_ASSERT(frame->local); - local = frame->local; + local = frame->local; - ret = xdr_to_generic (*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); - if (ret < 0) { - gf_log (frame->this->name, GF_LOG_ERROR, - "Failed to decode xdr response"); - goto out; - } + ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); + if (ret < 0) { + gf_log(frame->this->name, GF_LOG_ERROR, + "Failed to decode xdr response"); + goto out; + } - ret = dict_get_str (local->dict, "volname", &volname); - if (ret) { - gf_log (frame->this->name, GF_LOG_ERROR, - "Unable to get volname from dict"); - goto out; - } + ret = dict_get_str(local->dict, "volname", &volname); + if (ret) { + gf_log(frame->this->name, GF_LOG_ERROR, + "Unable to get volname from dict"); + goto out; + } - gf_log ("cli", GF_LOG_INFO, "Received resp to stop volume"); + gf_log("cli", GF_LOG_INFO, "Received resp to stop volume"); - if (global_state->mode & GLUSTER_MODE_XML) { - if (rsp.op_ret == 0) { - rsp_dict = dict_new (); - ret = dict_unserialize (rsp.dict.dict_val, - rsp.dict.dict_len, - &rsp_dict); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, - "Failed rsp_dict unserialization"); - goto out; - } - } - - ret = cli_xml_output_generic_volume ("volStop", rsp_dict, - rsp.op_ret, rsp.op_errno, - rsp.op_errstr); - if (ret) - gf_log ("cli", GF_LOG_ERROR, - "Error outputting to xml"); + if (global_state->mode & GLUSTER_MODE_XML) { + if (rsp.op_ret == 0) { + rsp_dict = dict_new(); + ret = dict_unserialize(rsp.dict.dict_val, rsp.dict.dict_len, + &rsp_dict); + if (ret) { + gf_log("cli", GF_LOG_ERROR, "Failed rsp_dict unserialization"); goto out; + } } - if (rsp.op_ret && strcmp (rsp.op_errstr, "")) - cli_err ("volume stop: %s: failed: %s", volname, rsp.op_errstr); - else if (rsp.op_ret) - cli_err ("volume stop: %s: failed", volname); - else - cli_out ("volume stop: %s: success", volname); + ret = cli_xml_output_generic_volume("volStop", rsp_dict, rsp.op_ret, + rsp.op_errno, rsp.op_errstr); + if (ret) + gf_log("cli", GF_LOG_ERROR, "Error outputting to xml"); + goto out; + } - ret = rsp.op_ret; + if (rsp.op_ret && strcmp(rsp.op_errstr, "")) + cli_err("volume stop: %s: failed: %s", volname, rsp.op_errstr); + else if (rsp.op_ret) + cli_err("volume stop: %s: failed", volname); + else + cli_out("volume stop: %s: success", volname); + + ret = rsp.op_ret; out: - cli_cmd_broadcast_response (ret); - free (rsp.op_errstr); - free (rsp.dict.dict_val); + cli_cmd_broadcast_response(ret); + free(rsp.op_errstr); + free(rsp.dict.dict_val); - return ret; + return ret; } int -gf_cli_print_rebalance_status (dict_t *dict, enum gf_task_types task_type, - gf_boolean_t is_tier) -{ - int ret = -1; - int count = 0; - int i = 1; - char key[256] = {0,}; - gf_defrag_status_t status_rcd = GF_DEFRAG_STATUS_NOT_STARTED; - uint64_t files = 0; - uint64_t size = 0; - uint64_t lookup = 0; - char *node_name = NULL; - uint64_t failures = 0; - uint64_t skipped = 0; - double elapsed = 0; - char *status_str = NULL; - char *size_str = NULL; - int32_t hrs = 0; - uint32_t min = 0; - uint32_t sec = 0; - gf_boolean_t down = _gf_false; - gf_boolean_t fix_layout = _gf_false; - uint64_t max_time = 0; - uint64_t max_elapsed = 0; - uint64_t time_left = 0; - gf_boolean_t show_estimates = _gf_false; - - - ret = dict_get_int32 (dict, "count", &count); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, "count not set"); - goto out; - } - - snprintf (key, sizeof (key), "status-1"); - - ret = dict_get_int32 (dict, key, (int32_t *)&status_rcd); - if (ret) { - gf_log ("cli", GF_LOG_TRACE, "count %d %d", count, 1); - gf_log ("cli", GF_LOG_TRACE, "failed to get status"); - goto out; - } - - /* Fix layout will be sent to all nodes for the volume - so every status should be of type - GF_DEFRAG_STATUS_LAYOUT_FIX* - */ - - if ((task_type == GF_TASK_TYPE_REBALANCE) - && (status_rcd >= GF_DEFRAG_STATUS_LAYOUT_FIX_STARTED)) { - fix_layout = _gf_true; - } - - if (fix_layout) { - cli_out ("%35s %41s %27s", "Node", "status", - "run time in h:m:s"); - cli_out ("%35s %41s %27s", "---------", "-----------", - "------------"); - } else { - cli_out ("%40s %16s %13s %13s %13s %13s %20s %18s", - "Node", "Rebalanced-files", "size", "scanned", - "failures", "skipped", "status", "run time in" - " h:m:s"); - cli_out ("%40s %16s %13s %13s %13s %13s %20s %18s", - "---------", "-----------", "-----------", - "-----------", "-----------", "-----------", - "------------", "--------------"); - } - - for (i = 1; i <= count; i++) { - /* Reset the variables to prevent carryover of values */ - node_name = NULL; - files = 0; - size = 0; - lookup = 0; - skipped = 0; - status_str = NULL; - elapsed = 0; - time_left = 0; - - /* Check if status is NOT_STARTED, and continue early */ - snprintf (key, sizeof (key), "status-%d", i); - - ret = dict_get_int32 (dict, key, (int32_t *)&status_rcd); - if (ret == -ENOENT) { - gf_log ("cli", GF_LOG_TRACE, "count %d %d", count, i); - gf_log ("cli", GF_LOG_TRACE, "failed to get status"); - gf_log ("cli", GF_LOG_ERROR, "node down and has failed" - " to set dict"); - down = _gf_true; - continue; - /* skip this node if value not available*/ - } else if (ret) { - gf_log ("cli", GF_LOG_TRACE, "count %d %d", count, i); - gf_log ("cli", GF_LOG_TRACE, "failed to get status"); - continue; - /* skip this node if value not available*/ - } - - if (GF_DEFRAG_STATUS_NOT_STARTED == status_rcd) - continue; - - if (GF_DEFRAG_STATUS_STARTED == status_rcd) - show_estimates = _gf_true; - - snprintf (key, 256, "node-name-%d", i); - ret = dict_get_str (dict, key, &node_name); - if (ret) - gf_log ("cli", GF_LOG_TRACE, "failed to get node-name"); +gf_cli_print_rebalance_status(dict_t *dict, enum gf_task_types task_type, + gf_boolean_t is_tier) +{ + int ret = -1; + int count = 0; + int i = 1; + char key[256] = { + 0, + }; + gf_defrag_status_t status_rcd = GF_DEFRAG_STATUS_NOT_STARTED; + uint64_t files = 0; + uint64_t size = 0; + uint64_t lookup = 0; + char *node_name = NULL; + uint64_t failures = 0; + uint64_t skipped = 0; + double elapsed = 0; + char *status_str = NULL; + char *size_str = NULL; + int32_t hrs = 0; + uint32_t min = 0; + uint32_t sec = 0; + gf_boolean_t down = _gf_false; + gf_boolean_t fix_layout = _gf_false; + uint64_t max_time = 0; + uint64_t max_elapsed = 0; + uint64_t time_left = 0; + gf_boolean_t show_estimates = _gf_false; + + ret = dict_get_int32(dict, "count", &count); + if (ret) { + gf_log("cli", GF_LOG_ERROR, "count not set"); + goto out; + } + + snprintf(key, sizeof(key), "status-1"); + + ret = dict_get_int32(dict, key, (int32_t *)&status_rcd); + if (ret) { + gf_log("cli", GF_LOG_TRACE, "count %d %d", count, 1); + gf_log("cli", GF_LOG_TRACE, "failed to get status"); + goto out; + } + + /* Fix layout will be sent to all nodes for the volume + so every status should be of type + GF_DEFRAG_STATUS_LAYOUT_FIX* + */ + + if ((task_type == GF_TASK_TYPE_REBALANCE) && + (status_rcd >= GF_DEFRAG_STATUS_LAYOUT_FIX_STARTED)) { + fix_layout = _gf_true; + } + + if (fix_layout) { + cli_out("%35s %41s %27s", "Node", "status", "run time in h:m:s"); + cli_out("%35s %41s %27s", "---------", "-----------", "------------"); + } else { + cli_out("%40s %16s %13s %13s %13s %13s %20s %18s", "Node", + "Rebalanced-files", "size", "scanned", "failures", "skipped", + "status", + "run time in" + " h:m:s"); + cli_out("%40s %16s %13s %13s %13s %13s %20s %18s", "---------", + "-----------", "-----------", "-----------", "-----------", + "-----------", "------------", "--------------"); + } + + for (i = 1; i <= count; i++) { + /* Reset the variables to prevent carryover of values */ + node_name = NULL; + files = 0; + size = 0; + lookup = 0; + skipped = 0; + status_str = NULL; + elapsed = 0; + time_left = 0; + + /* Check if status is NOT_STARTED, and continue early */ + snprintf(key, sizeof(key), "status-%d", i); + + ret = dict_get_int32(dict, key, (int32_t *)&status_rcd); + if (ret == -ENOENT) { + gf_log("cli", GF_LOG_TRACE, "count %d %d", count, i); + gf_log("cli", GF_LOG_TRACE, "failed to get status"); + gf_log("cli", GF_LOG_ERROR, + "node down and has failed" + " to set dict"); + down = _gf_true; + continue; + /* skip this node if value not available*/ + } else if (ret) { + gf_log("cli", GF_LOG_TRACE, "count %d %d", count, i); + gf_log("cli", GF_LOG_TRACE, "failed to get status"); + continue; + /* skip this node if value not available*/ + } + + if (GF_DEFRAG_STATUS_NOT_STARTED == status_rcd) + continue; + + if (GF_DEFRAG_STATUS_STARTED == status_rcd) + show_estimates = _gf_true; + + snprintf(key, 256, "node-name-%d", i); + ret = dict_get_str(dict, key, &node_name); + if (ret) + gf_log("cli", GF_LOG_TRACE, "failed to get node-name"); - snprintf (key, sizeof (key), "files-%d", i); - ret = dict_get_uint64 (dict, key, &files); - if (ret) - gf_log ("cli", GF_LOG_TRACE, - "failed to get file count"); + snprintf(key, sizeof(key), "files-%d", i); + ret = dict_get_uint64(dict, key, &files); + if (ret) + gf_log("cli", GF_LOG_TRACE, "failed to get file count"); - snprintf (key, sizeof (key), "size-%d", i); - ret = dict_get_uint64 (dict, key, &size); - if (ret) - gf_log ("cli", GF_LOG_TRACE, - "failed to get size of xfer"); + snprintf(key, sizeof(key), "size-%d", i); + ret = dict_get_uint64(dict, key, &size); + if (ret) + gf_log("cli", GF_LOG_TRACE, "failed to get size of xfer"); - snprintf (key, sizeof (key), "lookups-%d", i); - ret = dict_get_uint64 (dict, key, &lookup); - if (ret) - gf_log ("cli", GF_LOG_TRACE, - "failed to get lookedup file count"); + snprintf(key, sizeof(key), "lookups-%d", i); + ret = dict_get_uint64(dict, key, &lookup); + if (ret) + gf_log("cli", GF_LOG_TRACE, "failed to get lookedup file count"); - snprintf (key, sizeof (key), "failures-%d", i); - ret = dict_get_uint64 (dict, key, &failures); - if (ret) - gf_log ("cli", GF_LOG_TRACE, - "failed to get failures count"); + snprintf(key, sizeof(key), "failures-%d", i); + ret = dict_get_uint64(dict, key, &failures); + if (ret) + gf_log("cli", GF_LOG_TRACE, "failed to get failures count"); - snprintf (key, sizeof (key), "skipped-%d", i); - ret = dict_get_uint64 (dict, key, &skipped); - if (ret) - gf_log ("cli", GF_LOG_TRACE, - "failed to get skipped count"); + snprintf(key, sizeof(key), "skipped-%d", i); + ret = dict_get_uint64(dict, key, &skipped); + if (ret) + gf_log("cli", GF_LOG_TRACE, "failed to get skipped count"); - /* For remove-brick include skipped count into failure count*/ - if (task_type != GF_TASK_TYPE_REBALANCE) { - failures += skipped; - skipped = 0; - } + /* For remove-brick include skipped count into failure count*/ + if (task_type != GF_TASK_TYPE_REBALANCE) { + failures += skipped; + skipped = 0; + } - snprintf (key, sizeof (key), "run-time-%d", i); - ret = dict_get_double (dict, key, &elapsed); - if (ret) - gf_log ("cli", GF_LOG_TRACE, "failed to get run-time"); + snprintf(key, sizeof(key), "run-time-%d", i); + ret = dict_get_double(dict, key, &elapsed); + if (ret) + gf_log("cli", GF_LOG_TRACE, "failed to get run-time"); - snprintf (key, sizeof (key), "time-left-%d", i); - ret = dict_get_uint64 (dict, key, &time_left); - if (ret) - gf_log ("cli", GF_LOG_TRACE, - "failed to get time left"); + snprintf(key, sizeof(key), "time-left-%d", i); + ret = dict_get_uint64(dict, key, &time_left); + if (ret) + gf_log("cli", GF_LOG_TRACE, "failed to get time left"); - if (elapsed > max_elapsed) - max_elapsed = elapsed; + if (elapsed > max_elapsed) + max_elapsed = elapsed; - if (time_left > max_time) - max_time = time_left; + if (time_left > max_time) + max_time = time_left; - /* Check for array bound */ - if (status_rcd >= GF_DEFRAG_STATUS_MAX) - status_rcd = GF_DEFRAG_STATUS_MAX; + /* Check for array bound */ + if (status_rcd >= GF_DEFRAG_STATUS_MAX) + status_rcd = GF_DEFRAG_STATUS_MAX; - status_str = cli_vol_task_status_str[status_rcd]; - size_str = gf_uint64_2human_readable(size); - hrs = elapsed / 3600; - min = ((uint64_t) elapsed % 3600) / 60; - sec = ((uint64_t) elapsed % 3600) % 60; + status_str = cli_vol_task_status_str[status_rcd]; + size_str = gf_uint64_2human_readable(size); + hrs = elapsed / 3600; + min = ((uint64_t)elapsed % 3600) / 60; + sec = ((uint64_t)elapsed % 3600) % 60; - if (fix_layout) { - cli_out ("%35s %50s %8d:%d:%d", node_name, status_str, - hrs, min, sec); - } else { - if (size_str) { - cli_out ("%40s %16"PRIu64 " %13s" " %13"PRIu64 - " %13" PRIu64" %13"PRIu64 " %20s " - "%8d:%02d:%02d", node_name, files, - size_str, lookup, failures, skipped, - status_str, hrs, min, sec); - } else { - cli_out ("%40s %16"PRIu64 " %13"PRIu64 " %13" - PRIu64 " %13"PRIu64" %13"PRIu64 " %20s" - " %8d:%02d:%02d", node_name, files, - size, lookup, failures, skipped, - status_str, hrs, min, sec); - } - } - GF_FREE(size_str); - } - if (is_tier && down) - cli_out ("WARNING: glusterd might be down on one or more nodes." - " Please check the nodes that are down using \'gluster" - " peer status\' and start the glusterd on those nodes," - " else tier detach commit might fail!"); - - /* Max time will be non-zero if rebalance is still running */ - if (max_time) { - hrs = max_time / 3600; - min = (max_time % 3600) / 60; - sec = (max_time % 3600) % 60; - - if (hrs < REBAL_ESTIMATE_SEC_UPPER_LIMIT) { - cli_out ("Estimated time left for rebalance to " - "complete : %8d:%02d:%02d", hrs, min, sec); - } else { - cli_out ("Estimated time left for rebalance to " - "complete : > 2 months. Please try again " - "later."); - } + if (fix_layout) { + cli_out("%35s %50s %8d:%d:%d", node_name, status_str, hrs, min, + sec); } else { - /* Rebalance will return 0 if it could not calculate the - * estimates or if it is complete. - */ - if (!show_estimates) { - goto out; - } - if (max_elapsed <= REBAL_ESTIMATE_START_TIME) { - cli_out ("The estimated time for rebalance to complete " - "will be unavailable for the first 10 " - "minutes."); - } else { - cli_out ("Rebalance estimated time unavailable. Please " - "try again later."); - } + if (size_str) { + cli_out("%40s %16" PRIu64 + " %13s" + " %13" PRIu64 " %13" PRIu64 " %13" PRIu64 + " %20s " + "%8d:%02d:%02d", + node_name, files, size_str, lookup, failures, skipped, + status_str, hrs, min, sec); + } else { + cli_out("%40s %16" PRIu64 " %13" PRIu64 " %13" PRIu64 + " %13" PRIu64 " %13" PRIu64 + " %20s" + " %8d:%02d:%02d", + node_name, files, size, lookup, failures, skipped, + status_str, hrs, min, sec); + } + } + GF_FREE(size_str); + } + if (is_tier && down) + cli_out( + "WARNING: glusterd might be down on one or more nodes." + " Please check the nodes that are down using \'gluster" + " peer status\' and start the glusterd on those nodes," + " else tier detach commit might fail!"); + + /* Max time will be non-zero if rebalance is still running */ + if (max_time) { + hrs = max_time / 3600; + min = (max_time % 3600) / 60; + sec = (max_time % 3600) % 60; + + if (hrs < REBAL_ESTIMATE_SEC_UPPER_LIMIT) { + cli_out( + "Estimated time left for rebalance to " + "complete : %8d:%02d:%02d", + hrs, min, sec); + } else { + cli_out( + "Estimated time left for rebalance to " + "complete : > 2 months. Please try again " + "later."); + } + } else { + /* Rebalance will return 0 if it could not calculate the + * estimates or if it is complete. + */ + if (!show_estimates) { + goto out; + } + if (max_elapsed <= REBAL_ESTIMATE_START_TIME) { + cli_out( + "The estimated time for rebalance to complete " + "will be unavailable for the first 10 " + "minutes."); + } else { + cli_out( + "Rebalance estimated time unavailable. Please " + "try again later."); } + } out: - return ret; + return ret; } int -gf_cli_print_tier_status (dict_t *dict, enum gf_task_types task_type) -{ - int ret = -1; - int count = 0; - int i = 1; - uint64_t promoted = 0; - uint64_t demoted = 0; - char key[256] = {0,}; - char *node_name = NULL; - gf_defrag_status_t status_rcd = GF_DEFRAG_STATUS_NOT_STARTED; - char *status_str = NULL; - gf_boolean_t down = _gf_false; - double elapsed = 0; - int hrs = 0; - int min = 0; - int sec = 0; - - ret = dict_get_int32 (dict, "count", &count); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, "count not set"); - goto out; - } - - cli_out ("%-20s %-20s %-20s %-20s %-20s", "Node", "Promoted files", - "Demoted files", "Status", "run time in h:m:s"); - cli_out ("%-20s %-20s %-20s %-20s %-20s", "---------", "---------", - "---------", "---------", "---------"); - - for (i = 1; i <= count; i++) { - /* Reset the variables to prevent carryover of values */ - node_name = NULL; - promoted = 0; - demoted = 0; - - /* Check if status is NOT_STARTED, and continue early */ - snprintf (key, sizeof (key), "status-%d", i); - - ret = dict_get_int32 (dict, key, (int32_t *)&status_rcd); - if (ret == -ENOENT) { - gf_log ("cli", GF_LOG_TRACE, "count: %d, %d," - "failed to get status", count, i); - gf_log ("cli", GF_LOG_ERROR, "node down and has failed" - " to set dict"); - down = _gf_true; - continue; - /*skipping this node as value unavailable*/ - } else if (ret) { - gf_log ("cli", GF_LOG_TRACE, "count: %d, %d," - "failed to get status", count, i); - continue; - } - - if (GF_DEFRAG_STATUS_NOT_STARTED == status_rcd) - continue; - - snprintf (key, sizeof (key), "node-name-%d", i); - ret = dict_get_str (dict, key, &node_name); - if (ret) - gf_log ("cli", GF_LOG_TRACE, "failed to get node-name"); +gf_cli_print_tier_status(dict_t *dict, enum gf_task_types task_type) +{ + int ret = -1; + int count = 0; + int i = 1; + uint64_t promoted = 0; + uint64_t demoted = 0; + char key[256] = { + 0, + }; + char *node_name = NULL; + gf_defrag_status_t status_rcd = GF_DEFRAG_STATUS_NOT_STARTED; + char *status_str = NULL; + gf_boolean_t down = _gf_false; + double elapsed = 0; + int hrs = 0; + int min = 0; + int sec = 0; + + ret = dict_get_int32(dict, "count", &count); + if (ret) { + gf_log("cli", GF_LOG_ERROR, "count not set"); + goto out; + } + + cli_out("%-20s %-20s %-20s %-20s %-20s", "Node", "Promoted files", + "Demoted files", "Status", "run time in h:m:s"); + cli_out("%-20s %-20s %-20s %-20s %-20s", "---------", "---------", + "---------", "---------", "---------"); + + for (i = 1; i <= count; i++) { + /* Reset the variables to prevent carryover of values */ + node_name = NULL; + promoted = 0; + demoted = 0; + + /* Check if status is NOT_STARTED, and continue early */ + snprintf(key, sizeof(key), "status-%d", i); + + ret = dict_get_int32(dict, key, (int32_t *)&status_rcd); + if (ret == -ENOENT) { + gf_log("cli", GF_LOG_TRACE, + "count: %d, %d," + "failed to get status", + count, i); + gf_log("cli", GF_LOG_ERROR, + "node down and has failed" + " to set dict"); + down = _gf_true; + continue; + /*skipping this node as value unavailable*/ + } else if (ret) { + gf_log("cli", GF_LOG_TRACE, + "count: %d, %d," + "failed to get status", + count, i); + continue; + } + + if (GF_DEFRAG_STATUS_NOT_STARTED == status_rcd) + continue; + + snprintf(key, sizeof(key), "node-name-%d", i); + ret = dict_get_str(dict, key, &node_name); + if (ret) + gf_log("cli", GF_LOG_TRACE, "failed to get node-name"); - snprintf (key, sizeof (key), "promoted-%d", i); - ret = dict_get_uint64 (dict, key, &promoted); - if (ret) - gf_log ("cli", GF_LOG_TRACE, - "failed to get promoted count"); + snprintf(key, sizeof(key), "promoted-%d", i); + ret = dict_get_uint64(dict, key, &promoted); + if (ret) + gf_log("cli", GF_LOG_TRACE, "failed to get promoted count"); - snprintf (key, sizeof (key), "demoted-%d", i); - ret = dict_get_uint64 (dict, key, &demoted); - if (ret) - gf_log ("cli", GF_LOG_TRACE, - "failed to get demoted count"); + snprintf(key, sizeof(key), "demoted-%d", i); + ret = dict_get_uint64(dict, key, &demoted); + if (ret) + gf_log("cli", GF_LOG_TRACE, "failed to get demoted count"); - snprintf (key, sizeof (key), "run-time-%d", i); - ret = dict_get_double (dict, key, &elapsed); - if (ret) - gf_log ("cli", GF_LOG_TRACE, "failed to get run-time"); + snprintf(key, sizeof(key), "run-time-%d", i); + ret = dict_get_double(dict, key, &elapsed); + if (ret) + gf_log("cli", GF_LOG_TRACE, "failed to get run-time"); - /* Check for array bound */ - if (status_rcd >= GF_DEFRAG_STATUS_MAX) - status_rcd = GF_DEFRAG_STATUS_MAX; + /* Check for array bound */ + if (status_rcd >= GF_DEFRAG_STATUS_MAX) + status_rcd = GF_DEFRAG_STATUS_MAX; - hrs = elapsed / 3600; - min = ((int) elapsed % 3600) / 60; - sec = ((int) elapsed % 3600) % 60; + hrs = elapsed / 3600; + min = ((int)elapsed % 3600) / 60; + sec = ((int)elapsed % 3600) % 60; - status_str = cli_vol_task_status_str[status_rcd]; - cli_out ("%-20s %-20"PRIu64" %-20"PRIu64" %-20s" - " %d:%d:%d", node_name, promoted, demoted, - status_str, hrs, min, sec); - } - if (down) - cli_out ("WARNING: glusterd might be down on one or more nodes." - " Please check the nodes that are down using \'gluster" - " peer status\' and start the glusterd on those nodes."); + status_str = cli_vol_task_status_str[status_rcd]; + cli_out("%-20s %-20" PRIu64 " %-20" PRIu64 + " %-20s" + " %d:%d:%d", + node_name, promoted, demoted, status_str, hrs, min, sec); + } + if (down) + cli_out( + "WARNING: glusterd might be down on one or more nodes." + " Please check the nodes that are down using \'gluster" + " peer status\' and start the glusterd on those nodes."); out: - return ret; + return ret; } int -gf_cli_defrag_volume_cbk (struct rpc_req *req, struct iovec *iov, - int count, void *myframe) -{ - gf_cli_rsp rsp = {0,}; - cli_local_t *local = NULL; - char *volname = NULL; - call_frame_t *frame = NULL; - int cmd = 0; - int ret = -1; - dict_t *dict = NULL; - char msg[1024] = {0,}; - char *task_id_str = NULL; - - if (-1 == req->rpc_status) { - goto out; - } - - GF_ASSERT (myframe); - - frame = myframe; - - GF_ASSERT (frame->local); - - local = frame->local; - - ret = xdr_to_generic (*iov, &rsp, - (xdrproc_t)xdr_gf_cli_rsp); +gf_cli_defrag_volume_cbk(struct rpc_req *req, struct iovec *iov, int count, + void *myframe) +{ + gf_cli_rsp rsp = { + 0, + }; + cli_local_t *local = NULL; + char *volname = NULL; + call_frame_t *frame = NULL; + int cmd = 0; + int ret = -1; + dict_t *dict = NULL; + char msg[1024] = { + 0, + }; + char *task_id_str = NULL; + + if (-1 == req->rpc_status) { + goto out; + } + + GF_ASSERT(myframe); + + frame = myframe; + + GF_ASSERT(frame->local); + + local = frame->local; + + ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); + if (ret < 0) { + gf_log(frame->this->name, GF_LOG_ERROR, + "Failed to decode xdr response"); + goto out; + } + + ret = dict_get_str(local->dict, "volname", &volname); + if (ret) { + gf_log(frame->this->name, GF_LOG_ERROR, "Failed to get volname"); + goto out; + } + + ret = dict_get_int32(local->dict, "rebalance-command", (int32_t *)&cmd); + if (ret) { + gf_log("cli", GF_LOG_ERROR, "Failed to get command"); + goto out; + } + + if (rsp.dict.dict_len) { + /* Unserialize the dictionary */ + dict = dict_new(); + + ret = dict_unserialize(rsp.dict.dict_val, rsp.dict.dict_len, &dict); if (ret < 0) { - gf_log (frame->this->name, GF_LOG_ERROR, - "Failed to decode xdr response"); - goto out; + gf_log("glusterd", GF_LOG_ERROR, + "failed to " + "unserialize req-buffer to dictionary"); + goto out; } + } - ret = dict_get_str (local->dict, "volname", &volname); + if (!((cmd == GF_DEFRAG_CMD_STOP) || (cmd == GF_DEFRAG_CMD_STATUS) || + (cmd == GF_DEFRAG_CMD_STATUS_TIER)) && + !(global_state->mode & GLUSTER_MODE_XML)) { + ret = dict_get_str(dict, GF_REBALANCE_TID_KEY, &task_id_str); if (ret) { - gf_log (frame->this->name, GF_LOG_ERROR, - "Failed to get volname"); - goto out; - } - - ret = dict_get_int32 (local->dict, "rebalance-command", - (int32_t *)&cmd); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, - "Failed to get command"); - goto out; - } - - if (rsp.dict.dict_len) { - /* Unserialize the dictionary */ - dict = dict_new (); - - ret = dict_unserialize (rsp.dict.dict_val, - rsp.dict.dict_len, - &dict); - if (ret < 0) { - gf_log ("glusterd", GF_LOG_ERROR, - "failed to " - "unserialize req-buffer to dictionary"); - goto out; - } - } - - if (!((cmd == GF_DEFRAG_CMD_STOP) || - (cmd == GF_DEFRAG_CMD_STATUS) || - (cmd == GF_DEFRAG_CMD_STATUS_TIER)) && - !(global_state->mode & GLUSTER_MODE_XML)) { - ret = dict_get_str (dict, GF_REBALANCE_TID_KEY, &task_id_str); - if (ret) { - gf_log ("cli", GF_LOG_WARNING, - "failed to get %s from dict", - GF_REBALANCE_TID_KEY); - goto out; - } - if (rsp.op_ret && strcmp (rsp.op_errstr, "")) { - snprintf (msg, sizeof (msg), "%s", rsp.op_errstr); - } else { - if (!rsp.op_ret) { - /* append errstr in the cli msg for successful - * case since unlock failures can be highlighted - * event though rebalance command was successful - */ - if (cmd == GF_DEFRAG_CMD_START_TIER) { - snprintf (msg, sizeof (msg), "Tier " - "start is successful on %s.", - volname); - } else if (cmd == GF_DEFRAG_CMD_STOP_TIER) { - snprintf (msg, sizeof (msg), "Tier " - "daemon stopped " - "on %s.", volname); - } else { - snprintf (msg, sizeof (msg), - "Rebalance on %s has been " - "started successfully. Use " - "rebalance status command to" - " check status of the " - "rebalance process.\nID: %s", - volname, task_id_str); - } - } else { - snprintf (msg, sizeof (msg), - "Starting rebalance on volume %s has " - "been unsuccessful.", volname); - } - } - goto done; - } - - if (cmd == GF_DEFRAG_CMD_STOP) { - if (rsp.op_ret == -1) { - if (strcmp (rsp.op_errstr, "")) - snprintf (msg, sizeof (msg), - "%s", rsp.op_errstr); - else - snprintf (msg, sizeof (msg), - "rebalance volume %s stop failed", - volname); - goto done; - } else { - /* append errstr in the cli msg for successful case - * since unlock failures can be highlighted event though - * rebalance command was successful */ - snprintf (msg, sizeof (msg), - "rebalance process may be in the middle of a " - "file migration.\nThe process will be fully " - "stopped once the migration of the file is " - "complete.\nPlease check rebalance process " - "for completion before doing any further " - "brick related tasks on the volume.\n%s", - rsp.op_errstr); - } + gf_log("cli", GF_LOG_WARNING, "failed to get %s from dict", + GF_REBALANCE_TID_KEY); + goto out; } - if (cmd == GF_DEFRAG_CMD_STATUS || cmd == GF_DEFRAG_CMD_STATUS_TIER) { - if (rsp.op_ret == -1) { - if (strcmp (rsp.op_errstr, "")) - snprintf (msg, sizeof (msg), - "%s", rsp.op_errstr); - else - snprintf (msg, sizeof (msg), - "Failed to get the status of " - "rebalance process"); - goto done; + if (rsp.op_ret && strcmp(rsp.op_errstr, "")) { + snprintf(msg, sizeof(msg), "%s", rsp.op_errstr); + } else { + if (!rsp.op_ret) { + /* append errstr in the cli msg for successful + * case since unlock failures can be highlighted + * event though rebalance command was successful + */ + if (cmd == GF_DEFRAG_CMD_START_TIER) { + snprintf(msg, sizeof(msg), + "Tier " + "start is successful on %s.", + volname); + } else if (cmd == GF_DEFRAG_CMD_STOP_TIER) { + snprintf(msg, sizeof(msg), + "Tier " + "daemon stopped " + "on %s.", + volname); } else { - snprintf (msg, sizeof (msg), "%s", rsp.op_errstr); - } + snprintf(msg, sizeof(msg), + "Rebalance on %s has been " + "started successfully. Use " + "rebalance status command to" + " check status of the " + "rebalance process.\nID: %s", + volname, task_id_str); + } + } else { + snprintf(msg, sizeof(msg), + "Starting rebalance on volume %s has " + "been unsuccessful.", + volname); + } + } + goto done; + } + + if (cmd == GF_DEFRAG_CMD_STOP) { + if (rsp.op_ret == -1) { + if (strcmp(rsp.op_errstr, "")) + snprintf(msg, sizeof(msg), "%s", rsp.op_errstr); + else + snprintf(msg, sizeof(msg), "rebalance volume %s stop failed", + volname); + goto done; + } else { + /* append errstr in the cli msg for successful case + * since unlock failures can be highlighted event though + * rebalance command was successful */ + snprintf(msg, sizeof(msg), + "rebalance process may be in the middle of a " + "file migration.\nThe process will be fully " + "stopped once the migration of the file is " + "complete.\nPlease check rebalance process " + "for completion before doing any further " + "brick related tasks on the volume.\n%s", + rsp.op_errstr); + } + } + if (cmd == GF_DEFRAG_CMD_STATUS || cmd == GF_DEFRAG_CMD_STATUS_TIER) { + if (rsp.op_ret == -1) { + if (strcmp(rsp.op_errstr, "")) + snprintf(msg, sizeof(msg), "%s", rsp.op_errstr); + else + snprintf(msg, sizeof(msg), + "Failed to get the status of " + "rebalance process"); + goto done; + } else { + snprintf(msg, sizeof(msg), "%s", rsp.op_errstr); } + } - if (global_state->mode & GLUSTER_MODE_XML) { - ret = cli_xml_output_vol_rebalance (cmd, dict, rsp.op_ret, - rsp.op_errno, - rsp.op_errstr); - goto out; - } + if (global_state->mode & GLUSTER_MODE_XML) { + ret = cli_xml_output_vol_rebalance(cmd, dict, rsp.op_ret, rsp.op_errno, + rsp.op_errstr); + goto out; + } - if (cmd == GF_DEFRAG_CMD_STATUS_TIER) - ret = gf_cli_print_tier_status (dict, GF_TASK_TYPE_REBALANCE); - else if (cmd == GF_DEFRAG_CMD_DETACH_STATUS) - ret = gf_cli_print_rebalance_status (dict, - GF_TASK_TYPE_REBALANCE, - _gf_true); - else - ret = gf_cli_print_rebalance_status (dict, - GF_TASK_TYPE_REBALANCE, - _gf_false); + if (cmd == GF_DEFRAG_CMD_STATUS_TIER) + ret = gf_cli_print_tier_status(dict, GF_TASK_TYPE_REBALANCE); + else if (cmd == GF_DEFRAG_CMD_DETACH_STATUS) + ret = gf_cli_print_rebalance_status(dict, GF_TASK_TYPE_REBALANCE, + _gf_true); + else + ret = gf_cli_print_rebalance_status(dict, GF_TASK_TYPE_REBALANCE, + _gf_false); - if (ret) - gf_log ("cli", GF_LOG_ERROR, - "Failed to print rebalance status"); + if (ret) + gf_log("cli", GF_LOG_ERROR, "Failed to print rebalance status"); done: - if (global_state->mode & GLUSTER_MODE_XML) - cli_xml_output_str ("volRebalance", msg, - rsp.op_ret, rsp.op_errno, - rsp.op_errstr); - else { - if (rsp.op_ret) - - if (cmd == GF_DEFRAG_CMD_START_TIER || cmd == - GF_DEFRAG_CMD_STATUS_TIER) { - cli_err ("Tiering Migration Functionality: %s:" - " failed%s%s", volname, - strlen (msg) ? ": " : "", msg); - } else - cli_err ("volume rebalance: %s: failed%s%s", - volname, strlen (msg) ? ": " : "", - msg); - else - if (cmd == GF_DEFRAG_CMD_START_TIER || cmd == - GF_DEFRAG_CMD_STATUS_TIER) { - cli_out ("Tiering Migration Functionality: %s:" - " success%s%s", volname, - strlen (msg) ? ": " : "", msg); - } else - cli_out ("volume rebalance: %s: success%s%s", - volname, strlen (msg) ? ": " : "", - msg); - } - ret = rsp.op_ret; + if (global_state->mode & GLUSTER_MODE_XML) + cli_xml_output_str("volRebalance", msg, rsp.op_ret, rsp.op_errno, + rsp.op_errstr); + else { + if (rsp.op_ret) -out: - free (rsp.op_errstr); //malloced by xdr - free (rsp.dict.dict_val); //malloced by xdr - if (dict) - dict_unref (dict); - cli_cmd_broadcast_response (ret); - return ret; + if (cmd == GF_DEFRAG_CMD_START_TIER || + cmd == GF_DEFRAG_CMD_STATUS_TIER) { + cli_err( + "Tiering Migration Functionality: %s:" + " failed%s%s", + volname, strlen(msg) ? ": " : "", msg); + } else + cli_err("volume rebalance: %s: failed%s%s", volname, + strlen(msg) ? ": " : "", msg); + else if (cmd == GF_DEFRAG_CMD_START_TIER || + cmd == GF_DEFRAG_CMD_STATUS_TIER) { + cli_out( + "Tiering Migration Functionality: %s:" + " success%s%s", + volname, strlen(msg) ? ": " : "", msg); + } else + cli_out("volume rebalance: %s: success%s%s", volname, + strlen(msg) ? ": " : "", msg); + } + ret = rsp.op_ret; + +out: + free(rsp.op_errstr); // malloced by xdr + free(rsp.dict.dict_val); // malloced by xdr + if (dict) + dict_unref(dict); + cli_cmd_broadcast_response(ret); + return ret; } int -gf_cli_rename_volume_cbk (struct rpc_req *req, struct iovec *iov, - int count, void *myframe) -{ - gf_cli_rsp rsp = {0,}; - int ret = -1; - char msg[1024] = {0,}; - - GF_ASSERT (myframe); - - if (-1 == req->rpc_status) { - goto out; - } - - ret = xdr_to_generic (*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); - if (ret < 0) { - gf_log (((call_frame_t *) myframe)->this->name, GF_LOG_ERROR, - "Failed to decode xdr response"); - goto out; - } - +gf_cli_rename_volume_cbk(struct rpc_req *req, struct iovec *iov, int count, + void *myframe) +{ + gf_cli_rsp rsp = { + 0, + }; + int ret = -1; + char msg[1024] = { + 0, + }; + + GF_ASSERT(myframe); + + if (-1 == req->rpc_status) { + goto out; + } + + ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); + if (ret < 0) { + gf_log(((call_frame_t *)myframe)->this->name, GF_LOG_ERROR, + "Failed to decode xdr response"); + goto out; + } + + gf_log("cli", GF_LOG_INFO, "Received resp to probe"); + snprintf(msg, sizeof(msg), "Rename volume %s", + (rsp.op_ret) ? "unsuccessful" : "successful"); + + if (global_state->mode & GLUSTER_MODE_XML) { + ret = cli_xml_output_str("volRename", msg, rsp.op_ret, rsp.op_errno, + rsp.op_errstr); + if (ret) + gf_log("cli", GF_LOG_ERROR, "Error outputting to xml"); + goto out; + } - gf_log ("cli", GF_LOG_INFO, "Received resp to probe"); - snprintf (msg, sizeof (msg), "Rename volume %s", - (rsp.op_ret) ? "unsuccessful": "successful"); + if (rsp.op_ret) + cli_err("volume rename: failed"); + else + cli_out("volume rename: success"); - if (global_state->mode & GLUSTER_MODE_XML) { - ret = cli_xml_output_str ("volRename", msg, rsp.op_ret, - rsp.op_errno, rsp.op_errstr); - if (ret) - gf_log ("cli", GF_LOG_ERROR, - "Error outputting to xml"); - goto out; - } - - if (rsp.op_ret) - cli_err ("volume rename: failed"); - else - cli_out ("volume rename: success"); - - ret = rsp.op_ret; + ret = rsp.op_ret; out: - cli_cmd_broadcast_response (ret); - return ret; + cli_cmd_broadcast_response(ret); + return ret; } int -gf_cli_reset_volume_cbk (struct rpc_req *req, struct iovec *iov, - int count, void *myframe) +gf_cli_reset_volume_cbk(struct rpc_req *req, struct iovec *iov, int count, + void *myframe) { - gf_cli_rsp rsp = {0,}; - int ret = -1; - char msg[1024] = {0,}; - - GF_ASSERT (myframe); - - if (-1 == req->rpc_status) { - goto out; - } - - ret = xdr_to_generic (*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); - if (ret < 0) { - gf_log (((call_frame_t *) myframe)->this->name, GF_LOG_ERROR, - "Failed to decode xdr response"); - goto out; - } - - gf_log ("cli", GF_LOG_INFO, "Received resp to reset"); - - if (strcmp (rsp.op_errstr, "")) - snprintf (msg, sizeof (msg), "%s", rsp.op_errstr); - else - snprintf (msg, sizeof (msg), "reset volume %s", - (rsp.op_ret) ? "unsuccessful": "successful"); - - if (global_state->mode & GLUSTER_MODE_XML) { - ret = cli_xml_output_str ("volReset", msg, rsp.op_ret, - rsp.op_errno, rsp.op_errstr); - if (ret) - gf_log ("cli", GF_LOG_ERROR, - "Error outputting to xml"); - goto out; - } + gf_cli_rsp rsp = { + 0, + }; + int ret = -1; + char msg[1024] = { + 0, + }; + + GF_ASSERT(myframe); + + if (-1 == req->rpc_status) { + goto out; + } + + ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); + if (ret < 0) { + gf_log(((call_frame_t *)myframe)->this->name, GF_LOG_ERROR, + "Failed to decode xdr response"); + goto out; + } + + gf_log("cli", GF_LOG_INFO, "Received resp to reset"); + + if (strcmp(rsp.op_errstr, "")) + snprintf(msg, sizeof(msg), "%s", rsp.op_errstr); + else + snprintf(msg, sizeof(msg), "reset volume %s", + (rsp.op_ret) ? "unsuccessful" : "successful"); + + if (global_state->mode & GLUSTER_MODE_XML) { + ret = cli_xml_output_str("volReset", msg, rsp.op_ret, rsp.op_errno, + rsp.op_errstr); + if (ret) + gf_log("cli", GF_LOG_ERROR, "Error outputting to xml"); + goto out; + } - if (rsp.op_ret) - cli_err ("volume reset: failed: %s", msg); - else - cli_out ("volume reset: success: %s", msg); + if (rsp.op_ret) + cli_err("volume reset: failed: %s", msg); + else + cli_out("volume reset: success: %s", msg); - ret = rsp.op_ret; + ret = rsp.op_ret; out: - cli_cmd_broadcast_response (ret); - return ret; + cli_cmd_broadcast_response(ret); + return ret; } char * -is_server_debug_xlator (void *myframe) -{ - call_frame_t *frame = NULL; - cli_local_t *local = NULL; - char **words = NULL; - char *key = NULL; - char *value = NULL; - char *debug_xlator = NULL; - - frame = myframe; - local = frame->local; - words = (char **)local->words; - - while (*words != NULL) { - if (strstr (*words, "trace") == NULL && - strstr (*words, "error-gen") == NULL) { - words++; - continue; - } - - key = *words; +is_server_debug_xlator(void *myframe) +{ + call_frame_t *frame = NULL; + cli_local_t *local = NULL; + char **words = NULL; + char *key = NULL; + char *value = NULL; + char *debug_xlator = NULL; + + frame = myframe; + local = frame->local; + words = (char **)local->words; + + while (*words != NULL) { + if (strstr(*words, "trace") == NULL && + strstr(*words, "error-gen") == NULL) { + words++; + continue; + } + + key = *words; + words++; + value = *words; + if (value == NULL) + break; + if (strstr(value, "client")) { + words++; + continue; + } else { + if (!(strstr(value, "posix") || strstr(value, "acl") || + strstr(value, "locks") || strstr(value, "io-threads") || + strstr(value, "marker") || strstr(value, "index"))) { words++; - value = *words; - if (value == NULL) - break; - if (strstr (value, "client")) { - words++; - continue; - } else { - if (!(strstr (value, "posix") || strstr (value, "acl") - || strstr (value, "locks") || - strstr (value, "io-threads") || - strstr (value, "marker") || - strstr (value, "index"))) { - words++; - continue; - } else { - debug_xlator = gf_strdup (key); - break; - } - } + continue; + } else { + debug_xlator = gf_strdup(key); + break; + } } + } - return debug_xlator; + return debug_xlator; } int -gf_cli_set_volume_cbk (struct rpc_req *req, struct iovec *iov, - int count, void *myframe) -{ - gf_cli_rsp rsp = {0,}; - int ret = -1; - dict_t *dict = NULL; - char *help_str = NULL; - char msg[1024] = {0,}; - char *debug_xlator = NULL; - char tmp_str[512] = {0,}; - - GF_ASSERT (myframe); - - if (-1 == req->rpc_status) { - goto out; - } - - ret = xdr_to_generic (*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); - if (ret < 0) { - gf_log (((call_frame_t *) myframe)->this->name, GF_LOG_ERROR, - "Failed to decode xdr response"); - goto out; - } - - gf_log ("cli", GF_LOG_INFO, "Received resp to set"); - - dict = dict_new (); - - if (!dict) { - ret = -1; - goto out; - } - - ret = dict_unserialize (rsp.dict.dict_val, rsp.dict.dict_len, &dict); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, - "failed to unserialize volume set respone dict"); - goto out; - } - - /* For brick processes graph change does not happen on the fly. - * The process has to be restarted. So this is a check from the - * volume set option such that if debug xlators such as trace/errorgen - * are provided in the set command, warn the user. - */ - debug_xlator = is_server_debug_xlator (myframe); - - if (dict_get_str (dict, "help-str", &help_str) && !msg[0]) - snprintf (msg, sizeof (msg), "Set volume %s", - (rsp.op_ret) ? "unsuccessful": "successful"); - if (rsp.op_ret == 0 && debug_xlator) { - snprintf (tmp_str, sizeof (tmp_str), "\n%s translator has been " - "added to the server volume file. Please restart the" - " volume for enabling the translator", debug_xlator); - } - - if ((global_state->mode & GLUSTER_MODE_XML) && (help_str == NULL)) { - ret = cli_xml_output_str ("volSet", msg, rsp.op_ret, - rsp.op_errno, rsp.op_errstr); - if (ret) - gf_log ("cli", GF_LOG_ERROR, - "Error outputting to xml"); - goto out; - } +gf_cli_set_volume_cbk(struct rpc_req *req, struct iovec *iov, int count, + void *myframe) +{ + gf_cli_rsp rsp = { + 0, + }; + int ret = -1; + dict_t *dict = NULL; + char *help_str = NULL; + char msg[1024] = { + 0, + }; + char *debug_xlator = NULL; + char tmp_str[512] = { + 0, + }; + + GF_ASSERT(myframe); + + if (-1 == req->rpc_status) { + goto out; + } + + ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); + if (ret < 0) { + gf_log(((call_frame_t *)myframe)->this->name, GF_LOG_ERROR, + "Failed to decode xdr response"); + goto out; + } + + gf_log("cli", GF_LOG_INFO, "Received resp to set"); + + dict = dict_new(); + + if (!dict) { + ret = -1; + goto out; + } + + ret = dict_unserialize(rsp.dict.dict_val, rsp.dict.dict_len, &dict); + if (ret) { + gf_log("cli", GF_LOG_ERROR, + "failed to unserialize volume set respone dict"); + goto out; + } + + /* For brick processes graph change does not happen on the fly. + * The process has to be restarted. So this is a check from the + * volume set option such that if debug xlators such as trace/errorgen + * are provided in the set command, warn the user. + */ + debug_xlator = is_server_debug_xlator(myframe); + + if (dict_get_str(dict, "help-str", &help_str) && !msg[0]) + snprintf(msg, sizeof(msg), "Set volume %s", + (rsp.op_ret) ? "unsuccessful" : "successful"); + if (rsp.op_ret == 0 && debug_xlator) { + snprintf(tmp_str, sizeof(tmp_str), + "\n%s translator has been " + "added to the server volume file. Please restart the" + " volume for enabling the translator", + debug_xlator); + } + + if ((global_state->mode & GLUSTER_MODE_XML) && (help_str == NULL)) { + ret = cli_xml_output_str("volSet", msg, rsp.op_ret, rsp.op_errno, + rsp.op_errstr); + if (ret) + gf_log("cli", GF_LOG_ERROR, "Error outputting to xml"); + goto out; + } - if (rsp.op_ret) { - if (strcmp (rsp.op_errstr, "")) - cli_err ("volume set: failed: %s", rsp.op_errstr); - else - cli_err ("volume set: failed"); + if (rsp.op_ret) { + if (strcmp(rsp.op_errstr, "")) + cli_err("volume set: failed: %s", rsp.op_errstr); + else + cli_err("volume set: failed"); + } else { + if (help_str == NULL) { + if (debug_xlator == NULL) + cli_out("volume set: success"); + else + cli_out("volume set: success%s", tmp_str); } else { - if (help_str == NULL) { - if (debug_xlator == NULL) - cli_out ("volume set: success"); - else - cli_out ("volume set: success%s", tmp_str); - }else { - cli_out ("%s", help_str); - } + cli_out("%s", help_str); } + } - ret = rsp.op_ret; + ret = rsp.op_ret; out: - if (dict) - dict_unref (dict); - GF_FREE (debug_xlator); - cli_cmd_broadcast_response (ret); - return ret; + if (dict) + dict_unref(dict); + GF_FREE(debug_xlator); + cli_cmd_broadcast_response(ret); + return ret; } int -gf_cli_add_tier_brick_cbk (struct rpc_req *req, struct iovec *iov, - int count, void *myframe) -{ - gf_cli_rsp rsp = {0,}; - int ret = -1; - char msg[1024] = {0,}; - - GF_VALIDATE_OR_GOTO ("cli", myframe, out); - - if (-1 == req->rpc_status) { - goto out; - } - - ret = xdr_to_generic (*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); - if (ret < 0) { - gf_log (((call_frame_t *) myframe)->this->name, GF_LOG_ERROR, - "Failed to decode xdr response"); - goto out; - } - - - gf_log ("cli", GF_LOG_INFO, "Received resp to attach tier"); - - if (rsp.op_ret && strcmp (rsp.op_errstr, "")) - snprintf (msg, sizeof (msg), "%s", rsp.op_errstr); - else - snprintf (msg, sizeof (msg), "Attach tier %s", - (rsp.op_ret) ? "unsuccessful" : "successful"); - - if (global_state->mode & GLUSTER_MODE_XML) { - ret = cli_xml_output_str ("volAttachTier", msg, rsp.op_ret, - rsp.op_errno, rsp.op_errstr); - if (ret) - gf_log ("cli", GF_LOG_ERROR, - "Error outputting to xml"); - goto out; - } +gf_cli_add_tier_brick_cbk(struct rpc_req *req, struct iovec *iov, int count, + void *myframe) +{ + gf_cli_rsp rsp = { + 0, + }; + int ret = -1; + char msg[1024] = { + 0, + }; + + GF_VALIDATE_OR_GOTO("cli", myframe, out); + + if (-1 == req->rpc_status) { + goto out; + } + + ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); + if (ret < 0) { + gf_log(((call_frame_t *)myframe)->this->name, GF_LOG_ERROR, + "Failed to decode xdr response"); + goto out; + } + + gf_log("cli", GF_LOG_INFO, "Received resp to attach tier"); + + if (rsp.op_ret && strcmp(rsp.op_errstr, "")) + snprintf(msg, sizeof(msg), "%s", rsp.op_errstr); + else + snprintf(msg, sizeof(msg), "Attach tier %s", + (rsp.op_ret) ? "unsuccessful" : "successful"); + + if (global_state->mode & GLUSTER_MODE_XML) { + ret = cli_xml_output_str("volAttachTier", msg, rsp.op_ret, rsp.op_errno, + rsp.op_errstr); + if (ret) + gf_log("cli", GF_LOG_ERROR, "Error outputting to xml"); + goto out; + } - if (rsp.op_ret) - cli_err ("volume attach-tier: failed: %s", msg); - else - cli_out ("volume attach-tier: success"); - ret = rsp.op_ret; + if (rsp.op_ret) + cli_err("volume attach-tier: failed: %s", msg); + else + cli_out("volume attach-tier: success"); + ret = rsp.op_ret; out: - cli_cmd_broadcast_response (ret); - free (rsp.dict.dict_val); - free (rsp.op_errstr); - return ret; + cli_cmd_broadcast_response(ret); + free(rsp.dict.dict_val); + free(rsp.op_errstr); + return ret; } int -gf_cli_attach_tier_cbk (struct rpc_req *req, struct iovec *iov, - int count, void *myframe) -{ - gf_cli_rsp rsp = {0,}; - int ret = -1; - char msg[1024] = {0,}; - - GF_VALIDATE_OR_GOTO ("cli", myframe, out); - - if (-1 == req->rpc_status) { - goto out; - } - - ret = xdr_to_generic (*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); - if (ret < 0) { - gf_log (((call_frame_t *) myframe)->this->name, GF_LOG_ERROR, - "Failed to decode xdr response"); - goto out; - } - - - gf_log ("cli", GF_LOG_INFO, "Received resp to attach tier"); - - if (rsp.op_ret && strcmp (rsp.op_errstr, "")) - snprintf (msg, sizeof (msg), "%s", rsp.op_errstr); - else - snprintf (msg, sizeof (msg), "Attach tier %s", - (rsp.op_ret) ? "unsuccessful" : "successful"); - - if (global_state->mode & GLUSTER_MODE_XML) { - ret = cli_xml_output_str ("volAttachTier", msg, rsp.op_ret, - rsp.op_errno, rsp.op_errstr); - if (ret) - gf_log ("cli", GF_LOG_ERROR, - "Error outputting to xml"); - goto out; - } +gf_cli_attach_tier_cbk(struct rpc_req *req, struct iovec *iov, int count, + void *myframe) +{ + gf_cli_rsp rsp = { + 0, + }; + int ret = -1; + char msg[1024] = { + 0, + }; + + GF_VALIDATE_OR_GOTO("cli", myframe, out); + + if (-1 == req->rpc_status) { + goto out; + } + + ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); + if (ret < 0) { + gf_log(((call_frame_t *)myframe)->this->name, GF_LOG_ERROR, + "Failed to decode xdr response"); + goto out; + } + + gf_log("cli", GF_LOG_INFO, "Received resp to attach tier"); + + if (rsp.op_ret && strcmp(rsp.op_errstr, "")) + snprintf(msg, sizeof(msg), "%s", rsp.op_errstr); + else + snprintf(msg, sizeof(msg), "Attach tier %s", + (rsp.op_ret) ? "unsuccessful" : "successful"); + + if (global_state->mode & GLUSTER_MODE_XML) { + ret = cli_xml_output_str("volAttachTier", msg, rsp.op_ret, rsp.op_errno, + rsp.op_errstr); + if (ret) + gf_log("cli", GF_LOG_ERROR, "Error outputting to xml"); + goto out; + } - if (rsp.op_ret) - cli_err ("volume attach-tier: failed: %s", msg); - else - cli_out ("volume attach-tier: success"); - ret = rsp.op_ret; + if (rsp.op_ret) + cli_err("volume attach-tier: failed: %s", msg); + else + cli_out("volume attach-tier: success"); + ret = rsp.op_ret; out: - cli_cmd_broadcast_response (ret); - free (rsp.dict.dict_val); - free (rsp.op_errstr); - return ret; + cli_cmd_broadcast_response(ret); + free(rsp.dict.dict_val); + free(rsp.op_errstr); + return ret; } int -gf_cli_remove_tier_brick_cbk (struct rpc_req *req, struct iovec *iov, - int count, void *myframe) -{ - - gf_cli_rsp rsp = {0,}; - int ret = -1; - char msg[1024] = {0,}; - char *cmd_str = "unknown"; - cli_local_t *local = NULL; - call_frame_t *frame = NULL; - char *task_id_str = NULL; - dict_t *rsp_dict = NULL; - int32_t command = 0; - - GF_ASSERT (myframe); - - if (-1 == req->rpc_status) { - goto out; - } - - frame = myframe; - - GF_ASSERT (frame->local); - - local = frame->local; - - ret = xdr_to_generic (*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); - if (ret < 0) { - gf_log (frame->this->name, GF_LOG_ERROR, - "Failed to decode xdr response"); - goto out; - } - - ret = dict_get_int32 (local->dict, "command", &command); +gf_cli_remove_tier_brick_cbk(struct rpc_req *req, struct iovec *iov, int count, + void *myframe) +{ + gf_cli_rsp rsp = { + 0, + }; + int ret = -1; + char msg[1024] = { + 0, + }; + char *cmd_str = "unknown"; + cli_local_t *local = NULL; + call_frame_t *frame = NULL; + char *task_id_str = NULL; + dict_t *rsp_dict = NULL; + int32_t command = 0; + + GF_ASSERT(myframe); + + if (-1 == req->rpc_status) { + goto out; + } + + frame = myframe; + + GF_ASSERT(frame->local); + + local = frame->local; + + ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); + if (ret < 0) { + gf_log(frame->this->name, GF_LOG_ERROR, + "Failed to decode xdr response"); + goto out; + } + + ret = dict_get_int32(local->dict, "command", &command); + if (ret) { + gf_log("", GF_LOG_ERROR, "failed to get command"); + goto out; + } + + if (rsp.dict.dict_len) { + rsp_dict = dict_new(); + if (!rsp_dict) { + ret = -1; + goto out; + } + + ret = dict_unserialize(rsp.dict.dict_val, rsp.dict.dict_len, &rsp_dict); if (ret) { - gf_log ("", GF_LOG_ERROR, "failed to get command"); - goto out; + gf_log("cli", GF_LOG_ERROR, "Failed to unserialize rsp_dict"); + goto out; } + } - - if (rsp.dict.dict_len) { - rsp_dict = dict_new (); - if (!rsp_dict) { - ret = -1; - goto out; - } - - ret = dict_unserialize (rsp.dict.dict_val, rsp.dict.dict_len, - &rsp_dict); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, - "Failed to unserialize rsp_dict"); - goto out; - } - } - - switch (command) { + switch (command) { case GF_DEFRAG_CMD_DETACH_START: - cmd_str = "start"; - - ret = dict_get_str (rsp_dict, GF_REMOVE_BRICK_TID_KEY, - &task_id_str); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, - "remove-brick-id is not present in dict"); - } - break; + cmd_str = "start"; + + ret = dict_get_str(rsp_dict, GF_REMOVE_BRICK_TID_KEY, &task_id_str); + if (ret) { + gf_log("cli", GF_LOG_ERROR, + "remove-brick-id is not present in dict"); + } + break; case GF_DEFRAG_CMD_DETACH_COMMIT: - cmd_str = "commit"; - break; + cmd_str = "commit"; + break; case GF_DEFRAG_CMD_DETACH_COMMIT_FORCE: - cmd_str = "commit force"; - break; + cmd_str = "commit force"; + break; case GF_DEFRAG_CMD_DETACH_STOP: - cmd_str = "stop"; - break; + cmd_str = "stop"; + break; case GF_DEFRAG_CMD_DETACH_STATUS: - cmd_str = "status"; - break; + cmd_str = "status"; + break; default: - cmd_str = "unknown"; - break; - } + cmd_str = "unknown"; + break; + } - gf_log ("cli", GF_LOG_INFO, "Received resp to detach tier"); + gf_log("cli", GF_LOG_INFO, "Received resp to detach tier"); - if (rsp.op_ret && strcmp (rsp.op_errstr, "")) - snprintf (msg, sizeof (msg), "%s", rsp.op_errstr); - else - snprintf (msg, sizeof (msg), "Detach tier %s %s", cmd_str, - (rsp.op_ret) ? "unsuccessful" : "successful"); + if (rsp.op_ret && strcmp(rsp.op_errstr, "")) + snprintf(msg, sizeof(msg), "%s", rsp.op_errstr); + else + snprintf(msg, sizeof(msg), "Detach tier %s %s", cmd_str, + (rsp.op_ret) ? "unsuccessful" : "successful"); - ret = rsp.op_ret; - if (global_state->mode & GLUSTER_MODE_XML) { - ret = cli_xml_output_vol_remove_brick_detach_tier ( - _gf_true, rsp_dict, - rsp.op_ret, rsp.op_errno, - msg, "volDetachTier"); + ret = rsp.op_ret; + if (global_state->mode & GLUSTER_MODE_XML) { + ret = cli_xml_output_vol_remove_brick_detach_tier( + _gf_true, rsp_dict, rsp.op_ret, rsp.op_errno, msg, "volDetachTier"); - if (ret) - gf_log ("cli", GF_LOG_ERROR, - "Error outputting to xml"); - goto out; - } else { - if (rsp.op_ret) { - if (strcmp (rsp.op_errstr, "")) - snprintf (msg, sizeof (msg), "volume tier " - "detach %s: failed: %s", cmd_str, - rsp.op_errstr); - else - snprintf (msg, sizeof (msg), "volume tier " - "detach %s: failed", cmd_str); - - cli_err ("%s", msg); - goto out; - - } else { - cli_out ("volume detach tier %s: success", cmd_str); - if (GF_DEFRAG_CMD_DETACH_START == command && - task_id_str != NULL) - cli_out ("ID: %s", task_id_str); - if (GF_DEFRAG_CMD_DETACH_COMMIT == command) - cli_out ("Check the detached bricks to ensure " - "all files are migrated.\nIf files " - "with data are found on the brick " - "path, copy them via a gluster mount " - "point before re-purposing the " - "removed brick. "); - } - } - if (command == GF_DEFRAG_CMD_DETACH_STOP || - command == GF_DEFRAG_CMD_DETACH_STATUS) - ret = gf_cli_print_rebalance_status (rsp_dict, - GF_TASK_TYPE_REMOVE_BRICK, _gf_true); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Failed to print remove-brick " - "rebalance status"); - goto out; - } - - if ((command == GF_DEFRAG_CMD_DETACH_STOP) && (rsp.op_ret == 0)) { - cli_out ("'detach tier' process may be in the middle of a " - "file migration.\nThe process will be fully stopped " - "once the migration of the file is complete.\nPlease " - "check detach tier process for completion before " - "doing any further brick related tasks on the " - "volume."); - } - ret = rsp.op_ret; - -out: - cli_cmd_broadcast_response (ret); - free (rsp.dict.dict_val); - free (rsp.op_errstr); + if (ret) + gf_log("cli", GF_LOG_ERROR, "Error outputting to xml"); + goto out; + } else { + if (rsp.op_ret) { + if (strcmp(rsp.op_errstr, "")) + snprintf(msg, sizeof(msg), + "volume tier " + "detach %s: failed: %s", + cmd_str, rsp.op_errstr); + else + snprintf(msg, sizeof(msg), + "volume tier " + "detach %s: failed", + cmd_str); + + cli_err("%s", msg); + goto out; - return ret; + } else { + cli_out("volume detach tier %s: success", cmd_str); + if (GF_DEFRAG_CMD_DETACH_START == command && task_id_str != NULL) + cli_out("ID: %s", task_id_str); + if (GF_DEFRAG_CMD_DETACH_COMMIT == command) + cli_out( + "Check the detached bricks to ensure " + "all files are migrated.\nIf files " + "with data are found on the brick " + "path, copy them via a gluster mount " + "point before re-purposing the " + "removed brick. "); + } + } + if (command == GF_DEFRAG_CMD_DETACH_STOP || + command == GF_DEFRAG_CMD_DETACH_STATUS) + ret = gf_cli_print_rebalance_status(rsp_dict, GF_TASK_TYPE_REMOVE_BRICK, + _gf_true); + if (ret) { + gf_log("cli", GF_LOG_ERROR, + "Failed to print remove-brick " + "rebalance status"); + goto out; + } + + if ((command == GF_DEFRAG_CMD_DETACH_STOP) && (rsp.op_ret == 0)) { + cli_out( + "'detach tier' process may be in the middle of a " + "file migration.\nThe process will be fully stopped " + "once the migration of the file is complete.\nPlease " + "check detach tier process for completion before " + "doing any further brick related tasks on the " + "volume."); + } + ret = rsp.op_ret; + +out: + cli_cmd_broadcast_response(ret); + free(rsp.dict.dict_val); + free(rsp.op_errstr); + + return ret; } int -gf_cli_detach_tier_status_cbk (struct rpc_req *req, struct iovec *iov, - int count, void *myframe) -{ - gf_cli_rsp rsp = {0,}; - int ret = -1; - dict_t *dict = NULL; - char msg[1024] = {0,}; - int32_t command = 0; - gf1_op_commands cmd = GF_OP_CMD_NONE; - cli_local_t *local = NULL; - call_frame_t *frame = NULL; - char *cmd_str = "unknown"; - - GF_ASSERT (myframe); - - if (-1 == req->rpc_status) { - goto out; - } +gf_cli_detach_tier_status_cbk(struct rpc_req *req, struct iovec *iov, int count, + void *myframe) +{ + gf_cli_rsp rsp = { + 0, + }; + int ret = -1; + dict_t *dict = NULL; + char msg[1024] = { + 0, + }; + int32_t command = 0; + gf1_op_commands cmd = GF_OP_CMD_NONE; + cli_local_t *local = NULL; + call_frame_t *frame = NULL; + char *cmd_str = "unknown"; - frame = myframe; + GF_ASSERT(myframe); - GF_ASSERT (frame->local); + if (-1 == req->rpc_status) { + goto out; + } - local = frame->local; + frame = myframe; - ret = xdr_to_generic (*iov, &rsp, - (xdrproc_t)xdr_gf_cli_rsp); - if (ret < 0) { - gf_log (frame->this->name, GF_LOG_ERROR, - "Failed to decode xdr response"); - goto out; - } + GF_ASSERT(frame->local); - ret = dict_get_int32 (local->dict, "command", &command); - if (ret) - goto out; + local = frame->local; + + ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); + if (ret < 0) { + gf_log(frame->this->name, GF_LOG_ERROR, + "Failed to decode xdr response"); + goto out; + } + + ret = dict_get_int32(local->dict, "command", &command); + if (ret) + goto out; - cmd = command; + cmd = command; - switch (cmd) { + switch (cmd) { case GF_OP_CMD_STOP_DETACH_TIER: - cmd_str = "stop"; - break; + cmd_str = "stop"; + break; case GF_OP_CMD_STATUS: - cmd_str = "status"; - break; + cmd_str = "status"; + break; default: - break; - } - - ret = rsp.op_ret; - if (rsp.op_ret == -1) { - if (strcmp (rsp.op_errstr, "")) - snprintf (msg, sizeof (msg), "volume tier detach %s: " - "failed: %s", cmd_str, rsp.op_errstr); - else - snprintf (msg, sizeof (msg), "volume tier detach %s: " - "failed", cmd_str); + break; + } + + ret = rsp.op_ret; + if (rsp.op_ret == -1) { + if (strcmp(rsp.op_errstr, "")) + snprintf(msg, sizeof(msg), + "volume tier detach %s: " + "failed: %s", + cmd_str, rsp.op_errstr); + else + snprintf(msg, sizeof(msg), + "volume tier detach %s: " + "failed", + cmd_str); - if (global_state->mode & GLUSTER_MODE_XML) - goto xml_output; + if (global_state->mode & GLUSTER_MODE_XML) + goto xml_output; - cli_err ("%s", msg); - goto out; - } + cli_err("%s", msg); + goto out; + } - if (rsp.dict.dict_len) { - /* Unserialize the dictionary */ - dict = dict_new (); + if (rsp.dict.dict_len) { + /* Unserialize the dictionary */ + dict = dict_new(); - ret = dict_unserialize (rsp.dict.dict_val, - rsp.dict.dict_len, - &dict); - if (ret < 0) { - strncpy (msg, "failed to unserialize req-buffer to " - "dictionary", sizeof (msg)); + ret = dict_unserialize(rsp.dict.dict_val, rsp.dict.dict_len, &dict); + if (ret < 0) { + strncpy(msg, + "failed to unserialize req-buffer to " + "dictionary", + sizeof(msg)); - if (global_state->mode & GLUSTER_MODE_XML) { - rsp.op_ret = -1; - goto xml_output; - } + if (global_state->mode & GLUSTER_MODE_XML) { + rsp.op_ret = -1; + goto xml_output; + } - gf_log ("cli", GF_LOG_ERROR, "%s", msg); - goto out; - } + gf_log("cli", GF_LOG_ERROR, "%s", msg); + goto out; } + } xml_output: - if (global_state->mode & GLUSTER_MODE_XML) { - if (strcmp (rsp.op_errstr, "")) { - ret = cli_xml_output_vol_remove_brick_detach_tier ( - _gf_true, dict, - rsp.op_ret, - rsp.op_errno, - rsp.op_errstr, - "volDetachTier"); - } else { - ret = cli_xml_output_vol_remove_brick_detach_tier - (_gf_true, dict, - rsp.op_ret, - rsp.op_errno, - msg, - "volDetachTier"); - } - goto out; - } + if (global_state->mode & GLUSTER_MODE_XML) { + if (strcmp(rsp.op_errstr, "")) { + ret = cli_xml_output_vol_remove_brick_detach_tier( + _gf_true, dict, rsp.op_ret, rsp.op_errno, rsp.op_errstr, + "volDetachTier"); + } else { + ret = cli_xml_output_vol_remove_brick_detach_tier( + _gf_true, dict, rsp.op_ret, rsp.op_errno, msg, "volDetachTier"); + } + goto out; + } + + ret = gf_cli_print_rebalance_status(dict, GF_TASK_TYPE_REMOVE_BRICK, + _gf_true); + if (ret) { + gf_log("cli", GF_LOG_ERROR, + "Failed to print remove-brick " + "rebalance status"); + goto out; + } + + if ((cmd == GF_OP_CMD_STOP_DETACH_TIER) && (rsp.op_ret == 0)) { + cli_out( + "'detach tier' process may be in the middle of a " + "file migration.\nThe process will be fully stopped " + "once the migration of the file is complete.\nPlease " + "check detach tier process for completion before " + "doing any further brick related tasks on the " + "volume."); + } + +out: + free(rsp.dict.dict_val); /* malloced by xdr */ + if (dict) + dict_unref(dict); + cli_cmd_broadcast_response(ret); + return ret; +} - ret = gf_cli_print_rebalance_status (dict, GF_TASK_TYPE_REMOVE_BRICK, - _gf_true); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Failed to print remove-brick " - "rebalance status"); - goto out; - } +int +gf_cli_add_brick_cbk(struct rpc_req *req, struct iovec *iov, int count, + void *myframe) +{ + gf_cli_rsp rsp = { + 0, + }; + int ret = -1; + char msg[1024] = { + 0, + }; + + GF_ASSERT(myframe); + + if (-1 == req->rpc_status) { + goto out; + } + + ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); + if (ret < 0) { + gf_log(((call_frame_t *)myframe)->this->name, GF_LOG_ERROR, + "Failed to decode xdr response"); + goto out; + } + + gf_log("cli", GF_LOG_INFO, "Received resp to add brick"); + + if (rsp.op_ret && strcmp(rsp.op_errstr, "")) + snprintf(msg, sizeof(msg), "%s", rsp.op_errstr); + else + snprintf(msg, sizeof(msg), "Add Brick %s", + (rsp.op_ret) ? "unsuccessful" : "successful"); + + if (global_state->mode & GLUSTER_MODE_XML) { + ret = cli_xml_output_str("volAddBrick", msg, rsp.op_ret, rsp.op_errno, + rsp.op_errstr); + if (ret) + gf_log("cli", GF_LOG_ERROR, "Error outputting to xml"); + goto out; + } - if ((cmd == GF_OP_CMD_STOP_DETACH_TIER) && (rsp.op_ret == 0)) { - cli_out ("'detach tier' process may be in the middle of a " - "file migration.\nThe process will be fully stopped " - "once the migration of the file is complete.\nPlease " - "check detach tier process for completion before " - "doing any further brick related tasks on the " - "volume."); - } + if (rsp.op_ret) + cli_err("volume add-brick: failed: %s", rsp.op_errstr); + else + cli_out("volume add-brick: success"); + ret = rsp.op_ret; out: - free (rsp.dict.dict_val); /* malloced by xdr */ - if (dict) - dict_unref (dict); - cli_cmd_broadcast_response (ret); - return ret; + cli_cmd_broadcast_response(ret); + free(rsp.dict.dict_val); + free(rsp.op_errstr); + return ret; } - int -gf_cli_add_brick_cbk (struct rpc_req *req, struct iovec *iov, - int count, void *myframe) +gf_cli3_remove_brick_status_cbk(struct rpc_req *req, struct iovec *iov, + int count, void *myframe) { - gf_cli_rsp rsp = {0,}; - int ret = -1; - char msg[1024] = {0,}; - - GF_ASSERT (myframe); - - if (-1 == req->rpc_status) { - goto out; - } - - ret = xdr_to_generic (*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); - if (ret < 0) { - gf_log (((call_frame_t *) myframe)->this->name, GF_LOG_ERROR, - "Failed to decode xdr response"); - goto out; - } - - - gf_log ("cli", GF_LOG_INFO, "Received resp to add brick"); + gf_cli_rsp rsp = { + 0, + }; + int ret = -1; + dict_t *dict = NULL; + char msg[1024] = { + 0, + }; + int32_t command = 0; + gf1_op_commands cmd = GF_OP_CMD_NONE; + cli_local_t *local = NULL; + call_frame_t *frame = NULL; + char *cmd_str = "unknown"; - if (rsp.op_ret && strcmp (rsp.op_errstr, "")) - snprintf (msg, sizeof (msg), "%s", rsp.op_errstr); - else - snprintf (msg, sizeof (msg), "Add Brick %s", - (rsp.op_ret) ? "unsuccessful": "successful"); + GF_ASSERT(myframe); - if (global_state->mode & GLUSTER_MODE_XML) { - ret = cli_xml_output_str ("volAddBrick", msg, rsp.op_ret, - rsp.op_errno, rsp.op_errstr); - if (ret) - gf_log ("cli", GF_LOG_ERROR, - "Error outputting to xml"); - goto out; - } + if (-1 == req->rpc_status) { + goto out; + } - if (rsp.op_ret) - cli_err ("volume add-brick: failed: %s", rsp.op_errstr); - else - cli_out ("volume add-brick: success"); - ret = rsp.op_ret; + frame = myframe; -out: - cli_cmd_broadcast_response (ret); - free (rsp.dict.dict_val); - free (rsp.op_errstr); - return ret; -} + GF_ASSERT(frame->local); -int -gf_cli3_remove_brick_status_cbk (struct rpc_req *req, struct iovec *iov, - int count, void *myframe) -{ - gf_cli_rsp rsp = {0,}; - int ret = -1; - dict_t *dict = NULL; - char msg[1024] = {0,}; - int32_t command = 0; - gf1_op_commands cmd = GF_OP_CMD_NONE; - cli_local_t *local = NULL; - call_frame_t *frame = NULL; - char *cmd_str = "unknown"; - - GF_ASSERT (myframe); - - if (-1 == req->rpc_status) { - goto out; - } + local = frame->local; - frame = myframe; + ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); + if (ret < 0) { + gf_log(frame->this->name, GF_LOG_ERROR, + "Failed to decode xdr response"); + goto out; + } - GF_ASSERT (frame->local); + ret = dict_get_int32(local->dict, "command", &command); + if (ret) + goto out; - local = frame->local; + cmd = command; - ret = xdr_to_generic (*iov, &rsp, - (xdrproc_t)xdr_gf_cli_rsp); - if (ret < 0) { - gf_log (frame->this->name, GF_LOG_ERROR, - "Failed to decode xdr response"); - goto out; - } - - ret = dict_get_int32 (local->dict, "command", &command); - if (ret) - goto out; - - cmd = command; - - switch (cmd) { + switch (cmd) { case GF_OP_CMD_STOP: - cmd_str = "stop"; - break; + cmd_str = "stop"; + break; case GF_OP_CMD_STATUS: - cmd_str = "status"; - break; + cmd_str = "status"; + break; default: - break; - } + break; + } + + ret = rsp.op_ret; + if (rsp.op_ret == -1) { + if (strcmp(rsp.op_errstr, "")) + snprintf(msg, sizeof(msg), + "volume remove-brick %s: " + "failed: %s", + cmd_str, rsp.op_errstr); + else + snprintf(msg, sizeof(msg), + "volume remove-brick %s: " + "failed", + cmd_str); - ret = rsp.op_ret; - if (rsp.op_ret == -1) { - if (strcmp (rsp.op_errstr, "")) - snprintf (msg, sizeof (msg), "volume remove-brick %s: " - "failed: %s", cmd_str, rsp.op_errstr); - else - snprintf (msg, sizeof (msg), "volume remove-brick %s: " - "failed", cmd_str); + if (global_state->mode & GLUSTER_MODE_XML) + goto xml_output; - if (global_state->mode & GLUSTER_MODE_XML) - goto xml_output; + cli_err("%s", msg); + goto out; + } - cli_err ("%s", msg); - goto out; - } + if (rsp.dict.dict_len) { + /* Unserialize the dictionary */ + dict = dict_new(); - if (rsp.dict.dict_len) { - /* Unserialize the dictionary */ - dict = dict_new (); - - ret = dict_unserialize (rsp.dict.dict_val, - rsp.dict.dict_len, - &dict); - if (ret < 0) { - strncpy (msg, "failed to unserialize req-buffer to " - "dictionary", sizeof (msg)); + ret = dict_unserialize(rsp.dict.dict_val, rsp.dict.dict_len, &dict); + if (ret < 0) { + strncpy(msg, + "failed to unserialize req-buffer to " + "dictionary", + sizeof(msg)); - if (global_state->mode & GLUSTER_MODE_XML) { - rsp.op_ret = -1; - goto xml_output; - } + if (global_state->mode & GLUSTER_MODE_XML) { + rsp.op_ret = -1; + goto xml_output; + } - gf_log ("cli", GF_LOG_ERROR, "%s", msg); - goto out; - } + gf_log("cli", GF_LOG_ERROR, "%s", msg); + goto out; } + } xml_output: - if (global_state->mode & GLUSTER_MODE_XML) { - if (strcmp (rsp.op_errstr, "")) { - ret = cli_xml_output_vol_remove_brick_detach_tier ( - _gf_true, dict, - rsp.op_ret, - rsp.op_errno, - rsp.op_errstr, - "volRemoveBrick"); - } else { - ret = cli_xml_output_vol_remove_brick_detach_tier ( - _gf_true, dict, - rsp.op_ret, - rsp.op_errno, - msg, - "volRemoveBrick"); - } - goto out; - } - - ret = gf_cli_print_rebalance_status (dict, GF_TASK_TYPE_REMOVE_BRICK, - _gf_false); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Failed to print remove-brick " - "rebalance status"); - goto out; - } - - if ((cmd == GF_OP_CMD_STOP) && (rsp.op_ret == 0)) { - cli_out ("'remove-brick' process may be in the middle of a " - "file migration.\nThe process will be fully stopped " - "once the migration of the file is complete.\nPlease " - "check remove-brick process for completion before " - "doing any further brick related tasks on the " - "volume."); - } - -out: - free (rsp.dict.dict_val); //malloced by xdr - if (dict) - dict_unref (dict); - cli_cmd_broadcast_response (ret); - return ret; + if (global_state->mode & GLUSTER_MODE_XML) { + if (strcmp(rsp.op_errstr, "")) { + ret = cli_xml_output_vol_remove_brick_detach_tier( + _gf_true, dict, rsp.op_ret, rsp.op_errno, rsp.op_errstr, + "volRemoveBrick"); + } else { + ret = cli_xml_output_vol_remove_brick_detach_tier( + _gf_true, dict, rsp.op_ret, rsp.op_errno, msg, + "volRemoveBrick"); + } + goto out; + } + + ret = gf_cli_print_rebalance_status(dict, GF_TASK_TYPE_REMOVE_BRICK, + _gf_false); + if (ret) { + gf_log("cli", GF_LOG_ERROR, + "Failed to print remove-brick " + "rebalance status"); + goto out; + } + + if ((cmd == GF_OP_CMD_STOP) && (rsp.op_ret == 0)) { + cli_out( + "'remove-brick' process may be in the middle of a " + "file migration.\nThe process will be fully stopped " + "once the migration of the file is complete.\nPlease " + "check remove-brick process for completion before " + "doing any further brick related tasks on the " + "volume."); + } + +out: + free(rsp.dict.dict_val); // malloced by xdr + if (dict) + dict_unref(dict); + cli_cmd_broadcast_response(ret); + return ret; } - int -gf_cli_remove_brick_cbk (struct rpc_req *req, struct iovec *iov, - int count, void *myframe) -{ - gf_cli_rsp rsp = {0,}; - int ret = -1; - char msg[1024] = {0,}; - gf1_op_commands cmd = GF_OP_CMD_NONE; - char *cmd_str = "unknown"; - cli_local_t *local = NULL; - call_frame_t *frame = NULL; - char *task_id_str = NULL; - dict_t *rsp_dict = NULL; - - GF_ASSERT (myframe); - - if (-1 == req->rpc_status) { - goto out; - } - - frame = myframe; - - GF_ASSERT (frame->local); - - local = frame->local; - - ret = xdr_to_generic (*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); - if (ret < 0) { - gf_log (frame->this->name, GF_LOG_ERROR, - "Failed to decode xdr response"); - goto out; - } - - ret = dict_get_int32 (local->dict, "command", (int32_t *)&cmd); +gf_cli_remove_brick_cbk(struct rpc_req *req, struct iovec *iov, int count, + void *myframe) +{ + gf_cli_rsp rsp = { + 0, + }; + int ret = -1; + char msg[1024] = { + 0, + }; + gf1_op_commands cmd = GF_OP_CMD_NONE; + char *cmd_str = "unknown"; + cli_local_t *local = NULL; + call_frame_t *frame = NULL; + char *task_id_str = NULL; + dict_t *rsp_dict = NULL; + + GF_ASSERT(myframe); + + if (-1 == req->rpc_status) { + goto out; + } + + frame = myframe; + + GF_ASSERT(frame->local); + + local = frame->local; + + ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); + if (ret < 0) { + gf_log(frame->this->name, GF_LOG_ERROR, + "Failed to decode xdr response"); + goto out; + } + + ret = dict_get_int32(local->dict, "command", (int32_t *)&cmd); + if (ret) { + gf_log("", GF_LOG_ERROR, "failed to get command"); + goto out; + } + + if (rsp.dict.dict_len) { + rsp_dict = dict_new(); + if (!rsp_dict) { + ret = -1; + goto out; + } + + ret = dict_unserialize(rsp.dict.dict_val, rsp.dict.dict_len, &rsp_dict); if (ret) { - gf_log ("", GF_LOG_ERROR, "failed to get command"); - goto out; - } - - if (rsp.dict.dict_len) { - rsp_dict = dict_new (); - if (!rsp_dict) { - ret = -1; - goto out; - } - - ret = dict_unserialize (rsp.dict.dict_val, rsp.dict.dict_len, - &rsp_dict); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, - "Failed to unserialize rsp_dict"); - goto out; - } + gf_log("cli", GF_LOG_ERROR, "Failed to unserialize rsp_dict"); + goto out; } + } - switch (cmd) { + switch (cmd) { case GF_OP_CMD_DETACH_START: case GF_OP_CMD_START: - cmd_str = "start"; - - ret = dict_get_str (rsp_dict, GF_REMOVE_BRICK_TID_KEY, &task_id_str); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, - "remove-brick-id is not present in dict"); - } - break; + cmd_str = "start"; + + ret = dict_get_str(rsp_dict, GF_REMOVE_BRICK_TID_KEY, &task_id_str); + if (ret) { + gf_log("cli", GF_LOG_ERROR, + "remove-brick-id is not present in dict"); + } + break; case GF_OP_CMD_COMMIT: - cmd_str = "commit"; - break; + cmd_str = "commit"; + break; case GF_OP_CMD_COMMIT_FORCE: - cmd_str = "commit force"; - break; + cmd_str = "commit force"; + break; default: - cmd_str = "unknown"; - break; - } - - gf_log ("cli", GF_LOG_INFO, "Received resp to remove brick"); - - if (rsp.op_ret && strcmp (rsp.op_errstr, "")) - snprintf (msg, sizeof (msg), "%s", rsp.op_errstr); - else - snprintf (msg, sizeof (msg), "Remove Brick %s %s", cmd_str, - (rsp.op_ret) ? "unsuccessful": "successful"); - - if (global_state->mode & GLUSTER_MODE_XML) { - ret = cli_xml_output_vol_remove_brick_detach_tier ( - _gf_false, rsp_dict, - rsp.op_ret, rsp.op_errno, - msg, "volRemoveBrick"); - if (ret) - gf_log ("cli", GF_LOG_ERROR, - "Error outputting to xml"); - goto out; - } + cmd_str = "unknown"; + break; + } + + gf_log("cli", GF_LOG_INFO, "Received resp to remove brick"); + + if (rsp.op_ret && strcmp(rsp.op_errstr, "")) + snprintf(msg, sizeof(msg), "%s", rsp.op_errstr); + else + snprintf(msg, sizeof(msg), "Remove Brick %s %s", cmd_str, + (rsp.op_ret) ? "unsuccessful" : "successful"); + + if (global_state->mode & GLUSTER_MODE_XML) { + ret = cli_xml_output_vol_remove_brick_detach_tier( + _gf_false, rsp_dict, rsp.op_ret, rsp.op_errno, msg, + "volRemoveBrick"); + if (ret) + gf_log("cli", GF_LOG_ERROR, "Error outputting to xml"); + goto out; + } - if (rsp.op_ret) { - cli_err ("volume remove-brick %s: failed: %s", cmd_str, - msg); - } else { - cli_out ("volume remove-brick %s: success", cmd_str); - if (GF_OP_CMD_START == cmd && task_id_str != NULL) - cli_out ("ID: %s", task_id_str); - if (GF_OP_CMD_COMMIT == cmd) - cli_out ("Check the removed bricks to ensure all files " - "are migrated.\nIf files with data are " - "found on the brick path, copy them via a " - "gluster mount point before re-purposing the " - "removed brick. "); - } + if (rsp.op_ret) { + cli_err("volume remove-brick %s: failed: %s", cmd_str, msg); + } else { + cli_out("volume remove-brick %s: success", cmd_str); + if (GF_OP_CMD_START == cmd && task_id_str != NULL) + cli_out("ID: %s", task_id_str); + if (GF_OP_CMD_COMMIT == cmd) + cli_out( + "Check the removed bricks to ensure all files " + "are migrated.\nIf files with data are " + "found on the brick path, copy them via a " + "gluster mount point before re-purposing the " + "removed brick. "); + } - ret = rsp.op_ret; + ret = rsp.op_ret; out: - cli_cmd_broadcast_response (ret); - free (rsp.dict.dict_val); - free (rsp.op_errstr); + cli_cmd_broadcast_response(ret); + free(rsp.dict.dict_val); + free(rsp.op_errstr); - return ret; + return ret; } int -gf_cli_reset_brick_cbk (struct rpc_req *req, struct iovec *iov, - int count, void *myframe) -{ - gf_cli_rsp rsp = {0,}; - int ret = -1; - cli_local_t *local = NULL; - call_frame_t *frame = NULL; - char *rb_operation_str = NULL; - dict_t *rsp_dict = NULL; - char msg[1024] = {0,}; - char *reset_op = NULL; - - GF_ASSERT (myframe); - - if (-1 == req->rpc_status) { - goto out; - } - - frame = myframe; - - GF_ASSERT (frame->local); - - local = frame->local; - - ret = xdr_to_generic (*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); +gf_cli_reset_brick_cbk(struct rpc_req *req, struct iovec *iov, int count, + void *myframe) +{ + gf_cli_rsp rsp = { + 0, + }; + int ret = -1; + cli_local_t *local = NULL; + call_frame_t *frame = NULL; + char *rb_operation_str = NULL; + dict_t *rsp_dict = NULL; + char msg[1024] = { + 0, + }; + char *reset_op = NULL; + + GF_ASSERT(myframe); + + if (-1 == req->rpc_status) { + goto out; + } + + frame = myframe; + + GF_ASSERT(frame->local); + + local = frame->local; + + ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); + if (ret < 0) { + gf_log(frame->this->name, GF_LOG_ERROR, + "Failed to decode xdr response"); + goto out; + } + + ret = dict_get_str(local->dict, "operation", &reset_op); + if (ret) { + gf_log(frame->this->name, GF_LOG_ERROR, "dict_get on operation failed"); + goto out; + } + + if (rsp.dict.dict_len) { + /* Unserialize the dictionary */ + rsp_dict = dict_new(); + + ret = dict_unserialize(rsp.dict.dict_val, rsp.dict.dict_len, &rsp_dict); if (ret < 0) { - gf_log (frame->this->name, GF_LOG_ERROR, - "Failed to decode xdr response"); - goto out; - } - - ret = dict_get_str (local->dict, "operation", &reset_op); - if (ret) { - gf_log (frame->this->name, GF_LOG_ERROR, - "dict_get on operation failed"); - goto out; - } - - if (rsp.dict.dict_len) { - /* Unserialize the dictionary */ - rsp_dict = dict_new (); - - ret = dict_unserialize (rsp.dict.dict_val, - rsp.dict.dict_len, - &rsp_dict); - if (ret < 0) { - gf_log (frame->this->name, GF_LOG_ERROR, "failed to " - "unserialize rsp buffer to dictionary"); - goto out; - } - } - - if (strcmp (reset_op, "GF_RESET_OP_START") && - strcmp (reset_op, "GF_RESET_OP_COMMIT") && - strcmp (reset_op, "GF_RESET_OP_COMMIT_FORCE")) { - rb_operation_str = gf_strdup ("Unknown operation"); - ret = -1; - goto out; - } - - if (rsp.op_ret && (strcmp (rsp.op_errstr, ""))) { - rb_operation_str = gf_strdup (rsp.op_errstr); - } else { - if (!strcmp (reset_op, "GF_RESET_OP_START")) { - if (rsp.op_ret) - rb_operation_str = gf_strdup ("reset-brick " - "start " - "operation " - "failed"); - else - rb_operation_str = gf_strdup ("reset-brick " - "start " - "operation " - "successful"); - } else if (!strcmp (reset_op, "GF_RESET_OP_COMMIT")) { - - if (rsp.op_ret) - rb_operation_str = gf_strdup ("reset-brick " - "commit " - "operation " - "failed"); - else - rb_operation_str = gf_strdup ("reset-brick " - "commit " - "operation " - "successful"); - } else if (!strcmp (reset_op, "GF_RESET_OP_COMMIT_FORCE")) { - - if (rsp.op_ret) - rb_operation_str = gf_strdup ("reset-brick " - "commit " - "force operation " - "failed"); - else - rb_operation_str = gf_strdup ("reset-brick " - "commit " - "force operation " - "successful"); - } - } - - gf_log ("cli", GF_LOG_INFO, "Received resp to reset brick"); - snprintf (msg, sizeof (msg), "%s", - rb_operation_str ? rb_operation_str : "Unknown operation"); - - if (global_state->mode & GLUSTER_MODE_XML) { - ret = cli_xml_output_vol_replace_brick (rsp_dict, - rsp.op_ret, - rsp.op_errno, msg); - if (ret) - gf_log ("cli", GF_LOG_ERROR, - "Error outputting to xml"); - goto out; - } + gf_log(frame->this->name, GF_LOG_ERROR, + "failed to " + "unserialize rsp buffer to dictionary"); + goto out; + } + } + + if (strcmp(reset_op, "GF_RESET_OP_START") && + strcmp(reset_op, "GF_RESET_OP_COMMIT") && + strcmp(reset_op, "GF_RESET_OP_COMMIT_FORCE")) { + rb_operation_str = gf_strdup("Unknown operation"); + ret = -1; + goto out; + } + + if (rsp.op_ret && (strcmp(rsp.op_errstr, ""))) { + rb_operation_str = gf_strdup(rsp.op_errstr); + } else { + if (!strcmp(reset_op, "GF_RESET_OP_START")) { + if (rsp.op_ret) + rb_operation_str = gf_strdup( + "reset-brick " + "start " + "operation " + "failed"); + else + rb_operation_str = gf_strdup( + "reset-brick " + "start " + "operation " + "successful"); + } else if (!strcmp(reset_op, "GF_RESET_OP_COMMIT")) { + if (rsp.op_ret) + rb_operation_str = gf_strdup( + "reset-brick " + "commit " + "operation " + "failed"); + else + rb_operation_str = gf_strdup( + "reset-brick " + "commit " + "operation " + "successful"); + } else if (!strcmp(reset_op, "GF_RESET_OP_COMMIT_FORCE")) { + if (rsp.op_ret) + rb_operation_str = gf_strdup( + "reset-brick " + "commit " + "force operation " + "failed"); + else + rb_operation_str = gf_strdup( + "reset-brick " + "commit " + "force operation " + "successful"); + } + } + + gf_log("cli", GF_LOG_INFO, "Received resp to reset brick"); + snprintf(msg, sizeof(msg), "%s", + rb_operation_str ? rb_operation_str : "Unknown operation"); + + if (global_state->mode & GLUSTER_MODE_XML) { + ret = cli_xml_output_vol_replace_brick(rsp_dict, rsp.op_ret, + rsp.op_errno, msg); + if (ret) + gf_log("cli", GF_LOG_ERROR, "Error outputting to xml"); + goto out; + } - if (rsp.op_ret) - cli_err ("volume reset-brick: failed: %s", msg); - else - cli_out ("volume reset-brick: success: %s", msg); - ret = rsp.op_ret; + if (rsp.op_ret) + cli_err("volume reset-brick: failed: %s", msg); + else + cli_out("volume reset-brick: success: %s", msg); + ret = rsp.op_ret; out: - if (frame) - frame->local = NULL; + if (frame) + frame->local = NULL; - if (local) - cli_local_wipe (local); + if (local) + cli_local_wipe(local); - if (rb_operation_str) - GF_FREE (rb_operation_str); + if (rb_operation_str) + GF_FREE(rb_operation_str); - cli_cmd_broadcast_response (ret); - free (rsp.dict.dict_val); - if (rsp_dict) - dict_unref (rsp_dict); + cli_cmd_broadcast_response(ret); + free(rsp.dict.dict_val); + if (rsp_dict) + dict_unref(rsp_dict); - return ret; + return ret; } int -gf_cli_replace_brick_cbk (struct rpc_req *req, struct iovec *iov, - int count, void *myframe) -{ - gf_cli_rsp rsp = {0,}; - int ret = -1; - cli_local_t *local = NULL; - call_frame_t *frame = NULL; - char *rb_operation_str = NULL; - dict_t *rsp_dict = NULL; - char msg[1024] = {0,}; - char *replace_op = NULL; - - GF_ASSERT (myframe); - - if (-1 == req->rpc_status) { - goto out; - } - - frame = myframe; - - GF_ASSERT (frame->local); - - local = frame->local; - - ret = xdr_to_generic (*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); +gf_cli_replace_brick_cbk(struct rpc_req *req, struct iovec *iov, int count, + void *myframe) +{ + gf_cli_rsp rsp = { + 0, + }; + int ret = -1; + cli_local_t *local = NULL; + call_frame_t *frame = NULL; + char *rb_operation_str = NULL; + dict_t *rsp_dict = NULL; + char msg[1024] = { + 0, + }; + char *replace_op = NULL; + + GF_ASSERT(myframe); + + if (-1 == req->rpc_status) { + goto out; + } + + frame = myframe; + + GF_ASSERT(frame->local); + + local = frame->local; + + ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); + if (ret < 0) { + gf_log(frame->this->name, GF_LOG_ERROR, + "Failed to decode xdr response"); + goto out; + } + + ret = dict_get_str(local->dict, "operation", &replace_op); + if (ret) { + gf_log(frame->this->name, GF_LOG_ERROR, "dict_get on operation failed"); + goto out; + } + + if (rsp.dict.dict_len) { + /* Unserialize the dictionary */ + rsp_dict = dict_new(); + + ret = dict_unserialize(rsp.dict.dict_val, rsp.dict.dict_len, &rsp_dict); if (ret < 0) { - gf_log (frame->this->name, GF_LOG_ERROR, - "Failed to decode xdr response"); - goto out; - } - - ret = dict_get_str (local->dict, "operation", &replace_op); - if (ret) { - gf_log (frame->this->name, GF_LOG_ERROR, - "dict_get on operation failed"); - goto out; - } - - if (rsp.dict.dict_len) { - /* Unserialize the dictionary */ - rsp_dict = dict_new (); - - ret = dict_unserialize (rsp.dict.dict_val, - rsp.dict.dict_len, - &rsp_dict); - if (ret < 0) { - gf_log (frame->this->name, GF_LOG_ERROR, "failed to " - "unserialize rsp buffer to dictionary"); - goto out; - } - } - - if (!strcmp (replace_op, "GF_REPLACE_OP_COMMIT_FORCE")) { - - if (rsp.op_ret || ret) - rb_operation_str = gf_strdup ("replace-brick commit " - "force operation failed"); - else - rb_operation_str = gf_strdup ("replace-brick commit " - "force operation " - "successful"); - } else { - gf_log (frame->this->name, GF_LOG_DEBUG, "Unknown operation"); - } - - if (rsp.op_ret && (strcmp (rsp.op_errstr, ""))) { - rb_operation_str = gf_strdup (rsp.op_errstr); - } - - gf_log ("cli", GF_LOG_INFO, "Received resp to replace brick"); - snprintf (msg, sizeof (msg), "%s", - rb_operation_str ? rb_operation_str : "Unknown operation"); - - if (global_state->mode & GLUSTER_MODE_XML) { - ret = cli_xml_output_vol_replace_brick (rsp_dict, - rsp.op_ret, - rsp.op_errno, msg); - if (ret) - gf_log ("cli", GF_LOG_ERROR, - "Error outputting to xml"); - goto out; - } - - if (rsp.op_ret) - cli_err ("volume replace-brick: failed: %s", msg); + gf_log(frame->this->name, GF_LOG_ERROR, + "failed to " + "unserialize rsp buffer to dictionary"); + goto out; + } + } + + if (!strcmp(replace_op, "GF_REPLACE_OP_COMMIT_FORCE")) { + if (rsp.op_ret || ret) + rb_operation_str = gf_strdup( + "replace-brick commit " + "force operation failed"); else - cli_out ("volume replace-brick: success: %s", msg); - ret = rsp.op_ret; + rb_operation_str = gf_strdup( + "replace-brick commit " + "force operation " + "successful"); + } else { + gf_log(frame->this->name, GF_LOG_DEBUG, "Unknown operation"); + } + + if (rsp.op_ret && (strcmp(rsp.op_errstr, ""))) { + rb_operation_str = gf_strdup(rsp.op_errstr); + } + + gf_log("cli", GF_LOG_INFO, "Received resp to replace brick"); + snprintf(msg, sizeof(msg), "%s", + rb_operation_str ? rb_operation_str : "Unknown operation"); + + if (global_state->mode & GLUSTER_MODE_XML) { + ret = cli_xml_output_vol_replace_brick(rsp_dict, rsp.op_ret, + rsp.op_errno, msg); + if (ret) + gf_log("cli", GF_LOG_ERROR, "Error outputting to xml"); + goto out; + } + + if (rsp.op_ret) + cli_err("volume replace-brick: failed: %s", msg); + else + cli_out("volume replace-brick: success: %s", msg); + ret = rsp.op_ret; out: - if (frame) - frame->local = NULL; + if (frame) + frame->local = NULL; - if (local) - cli_local_wipe (local); + if (local) + cli_local_wipe(local); - if (rb_operation_str) - GF_FREE (rb_operation_str); + if (rb_operation_str) + GF_FREE(rb_operation_str); - cli_cmd_broadcast_response (ret); - free (rsp.dict.dict_val); - if (rsp_dict) - dict_unref (rsp_dict); + cli_cmd_broadcast_response(ret); + free(rsp.dict.dict_val); + if (rsp_dict) + dict_unref(rsp_dict); - return ret; + return ret; } - static int -gf_cli_log_rotate_cbk (struct rpc_req *req, struct iovec *iov, - int count, void *myframe) -{ - gf_cli_rsp rsp = {0,}; - int ret = -1; - char msg[1024] = {0,}; - - GF_ASSERT (myframe); - - if (-1 == req->rpc_status) { - goto out; - } - - ret = xdr_to_generic (*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); - if (ret < 0) { - gf_log (((call_frame_t *) myframe)->this->name, GF_LOG_ERROR, - "Failed to decode xdr response"); - goto out; - } - - gf_log ("cli", GF_LOG_DEBUG, "Received resp to log rotate"); - - if (rsp.op_ret && strcmp (rsp.op_errstr, "")) - snprintf (msg, sizeof (msg), "%s", rsp.op_errstr); - else - snprintf (msg, sizeof (msg), "log rotate %s", - (rsp.op_ret) ? "unsuccessful": "successful"); - - if (global_state->mode & GLUSTER_MODE_XML) { - ret = cli_xml_output_str ("volLogRotate", msg, rsp.op_ret, - rsp.op_errno, rsp.op_errstr); - if (ret) - gf_log ("cli", GF_LOG_ERROR, - "Error outputting to xml"); - goto out; - } +gf_cli_log_rotate_cbk(struct rpc_req *req, struct iovec *iov, int count, + void *myframe) +{ + gf_cli_rsp rsp = { + 0, + }; + int ret = -1; + char msg[1024] = { + 0, + }; + + GF_ASSERT(myframe); + + if (-1 == req->rpc_status) { + goto out; + } + + ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); + if (ret < 0) { + gf_log(((call_frame_t *)myframe)->this->name, GF_LOG_ERROR, + "Failed to decode xdr response"); + goto out; + } + + gf_log("cli", GF_LOG_DEBUG, "Received resp to log rotate"); + + if (rsp.op_ret && strcmp(rsp.op_errstr, "")) + snprintf(msg, sizeof(msg), "%s", rsp.op_errstr); + else + snprintf(msg, sizeof(msg), "log rotate %s", + (rsp.op_ret) ? "unsuccessful" : "successful"); + + if (global_state->mode & GLUSTER_MODE_XML) { + ret = cli_xml_output_str("volLogRotate", msg, rsp.op_ret, rsp.op_errno, + rsp.op_errstr); + if (ret) + gf_log("cli", GF_LOG_ERROR, "Error outputting to xml"); + goto out; + } - if (rsp.op_ret) - cli_err ("volume log-rotate: failed: %s", msg); - else - cli_out ("volume log-rotate: success"); - ret = rsp.op_ret; + if (rsp.op_ret) + cli_err("volume log-rotate: failed: %s", msg); + else + cli_out("volume log-rotate: success"); + ret = rsp.op_ret; out: - cli_cmd_broadcast_response (ret); - free (rsp.dict.dict_val); + cli_cmd_broadcast_response(ret); + free(rsp.dict.dict_val); - return ret; + return ret; } static int -gf_cli_sync_volume_cbk (struct rpc_req *req, struct iovec *iov, - int count, void *myframe) -{ - gf_cli_rsp rsp = {0,}; - int ret = -1; - char msg[1024] = {0,}; - - GF_ASSERT (myframe); - - if (-1 == req->rpc_status) { - goto out; - } - - ret = xdr_to_generic (*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); - if (ret < 0) { - gf_log (((call_frame_t *) myframe)->this->name, GF_LOG_ERROR, - "Failed to decode xdr response"); - goto out; - } - - gf_log ("cli", GF_LOG_DEBUG, "Received resp to sync"); - - if (rsp.op_ret && strcmp (rsp.op_errstr, "")) - snprintf (msg, sizeof (msg), "volume sync: failed: %s", - rsp.op_errstr); - else - snprintf (msg, sizeof (msg), "volume sync: %s", - (rsp.op_ret) ? "failed": "success"); - - if (global_state->mode & GLUSTER_MODE_XML) { - ret = cli_xml_output_str ("volSync", msg, rsp.op_ret, - rsp.op_errno, rsp.op_errstr); - if (ret) - gf_log ("cli", GF_LOG_ERROR, - "Error outputting to xml"); - goto out; - } +gf_cli_sync_volume_cbk(struct rpc_req *req, struct iovec *iov, int count, + void *myframe) +{ + gf_cli_rsp rsp = { + 0, + }; + int ret = -1; + char msg[1024] = { + 0, + }; + + GF_ASSERT(myframe); + + if (-1 == req->rpc_status) { + goto out; + } + + ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); + if (ret < 0) { + gf_log(((call_frame_t *)myframe)->this->name, GF_LOG_ERROR, + "Failed to decode xdr response"); + goto out; + } + + gf_log("cli", GF_LOG_DEBUG, "Received resp to sync"); + + if (rsp.op_ret && strcmp(rsp.op_errstr, "")) + snprintf(msg, sizeof(msg), "volume sync: failed: %s", rsp.op_errstr); + else + snprintf(msg, sizeof(msg), "volume sync: %s", + (rsp.op_ret) ? "failed" : "success"); + + if (global_state->mode & GLUSTER_MODE_XML) { + ret = cli_xml_output_str("volSync", msg, rsp.op_ret, rsp.op_errno, + rsp.op_errstr); + if (ret) + gf_log("cli", GF_LOG_ERROR, "Error outputting to xml"); + goto out; + } - if (rsp.op_ret) - cli_err ("%s", msg); - else - cli_out ("%s", msg); - ret = rsp.op_ret; + if (rsp.op_ret) + cli_err("%s", msg); + else + cli_out("%s", msg); + ret = rsp.op_ret; out: - cli_cmd_broadcast_response (ret); - return ret; + cli_cmd_broadcast_response(ret); + return ret; } static int -print_quota_list_usage_output (cli_local_t *local, char *path, int64_t avail, - char *sl_str, quota_limits_t *limits, - quota_meta_t *used_space, gf_boolean_t sl, - gf_boolean_t hl, double sl_num, - gf_boolean_t limit_set) -{ - int32_t ret = -1; - char *used_str = NULL; - char *avail_str = NULL; - char *hl_str = NULL; - char *sl_val = NULL; - - used_str = gf_uint64_2human_readable (used_space->size); - - if (limit_set) { - hl_str = gf_uint64_2human_readable (limits->hl); - avail_str = gf_uint64_2human_readable (avail); - - sl_val = gf_uint64_2human_readable (sl_num); - } - - if (global_state->mode & GLUSTER_MODE_XML) { - ret = cli_quota_xml_output (local, path, limits->hl, - sl_str, sl_num, used_space->size, - avail, sl ? "Yes" : "No", - hl ? "Yes" : "No", limit_set); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Failed to " - "output in xml format for quota " - "list command"); - } - goto out; - } - - if (limit_set) { - if (!used_str) { - cli_out ("%-40s %7s %7s(%s) %8"PRIu64 "%9"PRIu64"" - "%15s %18s", path, hl_str, sl_str, sl_val, - used_space->size, avail, - sl ? "Yes" : "No", hl ? "Yes" : "No"); - } else { - cli_out ("%-40s %7s %7s(%s) %8s %7s %15s %20s", - path, hl_str, sl_str, sl_val, used_str, - avail_str, sl ? "Yes" : "No", - hl ? "Yes" : "No"); - } +print_quota_list_usage_output(cli_local_t *local, char *path, int64_t avail, + char *sl_str, quota_limits_t *limits, + quota_meta_t *used_space, gf_boolean_t sl, + gf_boolean_t hl, double sl_num, + gf_boolean_t limit_set) +{ + int32_t ret = -1; + char *used_str = NULL; + char *avail_str = NULL; + char *hl_str = NULL; + char *sl_val = NULL; + + used_str = gf_uint64_2human_readable(used_space->size); + + if (limit_set) { + hl_str = gf_uint64_2human_readable(limits->hl); + avail_str = gf_uint64_2human_readable(avail); + + sl_val = gf_uint64_2human_readable(sl_num); + } + + if (global_state->mode & GLUSTER_MODE_XML) { + ret = cli_quota_xml_output(local, path, limits->hl, sl_str, sl_num, + used_space->size, avail, sl ? "Yes" : "No", + hl ? "Yes" : "No", limit_set); + if (ret) { + gf_log("cli", GF_LOG_ERROR, + "Failed to " + "output in xml format for quota " + "list command"); + } + goto out; + } + + if (limit_set) { + if (!used_str) { + cli_out("%-40s %7s %7s(%s) %8" PRIu64 "%9" PRIu64 + "" + "%15s %18s", + path, hl_str, sl_str, sl_val, used_space->size, avail, + sl ? "Yes" : "No", hl ? "Yes" : "No"); } else { - cli_out ("%-36s %10s %10s %14s %9s %15s %18s", - path, "N/A", "N/A", used_str, "N/A", - "N/A", "N/A"); + cli_out("%-40s %7s %7s(%s) %8s %7s %15s %20s", path, hl_str, sl_str, + sl_val, used_str, avail_str, sl ? "Yes" : "No", + hl ? "Yes" : "No"); } + } else { + cli_out("%-36s %10s %10s %14s %9s %15s %18s", path, "N/A", "N/A", + used_str, "N/A", "N/A", "N/A"); + } - ret = 0; + ret = 0; out: - GF_FREE (hl_str); - GF_FREE (used_str); - GF_FREE (avail_str); - GF_FREE (sl_val); + GF_FREE(hl_str); + GF_FREE(used_str); + GF_FREE(avail_str); + GF_FREE(sl_val); - return ret; + return ret; } static int -print_quota_list_object_output (cli_local_t *local, char *path, int64_t avail, +print_quota_list_object_output(cli_local_t *local, char *path, int64_t avail, char *sl_str, quota_limits_t *limits, quota_meta_t *used_space, gf_boolean_t sl, gf_boolean_t hl, double sl_num, gf_boolean_t limit_set) { - int32_t ret = -1; - int64_t sl_val = sl_num; + int32_t ret = -1; + int64_t sl_val = sl_num; - if (global_state->mode & GLUSTER_MODE_XML) { - ret = cli_quota_object_xml_output (local, path, sl_str, sl_val, - limits, used_space, avail, - sl ? "Yes" : "No", - hl ? "Yes" : "No", - limit_set); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Failed to " - "output in xml format for quota " - "list command"); - } - goto out; + if (global_state->mode & GLUSTER_MODE_XML) { + ret = cli_quota_object_xml_output(local, path, sl_str, sl_val, limits, + used_space, avail, sl ? "Yes" : "No", + hl ? "Yes" : "No", limit_set); + if (ret) { + gf_log("cli", GF_LOG_ERROR, + "Failed to " + "output in xml format for quota " + "list command"); } + goto out; + } - if (limit_set) { - cli_out ("%-40s %9"PRIu64" %9s(%"PRId64") %10"PRIu64"" - "%10"PRIu64" %11"PRIu64" %15s %20s", - path, limits->hl, sl_str, sl_val, - used_space->file_count, used_space->dir_count, - avail, sl ? "Yes" : "No", hl ? "Yes" : "No"); - } else { - cli_out ("%-40s %9s %9s %10"PRIu64" %10"PRIu64" %11s %15s %20s", - path, "N/A", "N/A", used_space->file_count, - used_space->dir_count, "N/A", "N/A", "N/A"); - } - ret = 0; + if (limit_set) { + cli_out("%-40s %9" PRIu64 " %9s(%" PRId64 ") %10" PRIu64 + "" + "%10" PRIu64 " %11" PRIu64 " %15s %20s", + path, limits->hl, sl_str, sl_val, used_space->file_count, + used_space->dir_count, avail, sl ? "Yes" : "No", + hl ? "Yes" : "No"); + } else { + cli_out("%-40s %9s %9s %10" PRIu64 " %10" PRIu64 " %11s %15s %20s", + path, "N/A", "N/A", used_space->file_count, + used_space->dir_count, "N/A", "N/A", "N/A"); + } + ret = 0; out: - return ret; + return ret; } static int -print_quota_list_output (cli_local_t *local, char *path, char *default_sl, - quota_limits_t *limits, quota_meta_t *used_space, - int type, gf_boolean_t limit_set) -{ - int64_t avail = 0; - char percent_str[20] = {0}; - char *sl_final = NULL; - int ret = -1; - double sl_num = 0; - gf_boolean_t sl = _gf_false; - gf_boolean_t hl = _gf_false; - int64_t used_size = 0; - - GF_ASSERT (local); - GF_ASSERT (path); - - if (limit_set) { - if (limits->sl < 0) { - ret = gf_string2percent (default_sl, &sl_num); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, - "could not convert default soft limit" - " to percent"); - goto out; - } - sl_num = (sl_num * limits->hl) / 100; - sl_final = default_sl; - } else { - sl_num = (limits->sl * limits->hl) / 100; - snprintf (percent_str, sizeof (percent_str), "%"PRIu64"%%", - limits->sl); - sl_final = percent_str; - } - if (type == GF_QUOTA_OPTION_TYPE_LIST) - used_size = used_space->size; - else - used_size = used_space->file_count + used_space->dir_count; - - if (limits->hl > used_size) { - avail = limits->hl - used_size; - hl = _gf_false; - if (used_size > sl_num) - sl = _gf_true; - else - sl = _gf_false; - } else { - avail = 0; - hl = sl = _gf_true; - } +print_quota_list_output(cli_local_t *local, char *path, char *default_sl, + quota_limits_t *limits, quota_meta_t *used_space, + int type, gf_boolean_t limit_set) +{ + int64_t avail = 0; + char percent_str[20] = {0}; + char *sl_final = NULL; + int ret = -1; + double sl_num = 0; + gf_boolean_t sl = _gf_false; + gf_boolean_t hl = _gf_false; + int64_t used_size = 0; + + GF_ASSERT(local); + GF_ASSERT(path); + + if (limit_set) { + if (limits->sl < 0) { + ret = gf_string2percent(default_sl, &sl_num); + if (ret) { + gf_log("cli", GF_LOG_ERROR, + "could not convert default soft limit" + " to percent"); + goto out; + } + sl_num = (sl_num * limits->hl) / 100; + sl_final = default_sl; + } else { + sl_num = (limits->sl * limits->hl) / 100; + snprintf(percent_str, sizeof(percent_str), "%" PRIu64 "%%", + limits->sl); + sl_final = percent_str; } - if (type == GF_QUOTA_OPTION_TYPE_LIST) - ret = print_quota_list_usage_output (local, path, avail, - sl_final, limits, - used_space, sl, hl, - sl_num, limit_set); + used_size = used_space->size; else - ret = print_quota_list_object_output (local, path, avail, - sl_final, limits, - used_space, sl, hl, - sl_num, limit_set); + used_size = used_space->file_count + used_space->dir_count; + + if (limits->hl > used_size) { + avail = limits->hl - used_size; + hl = _gf_false; + if (used_size > sl_num) + sl = _gf_true; + else + sl = _gf_false; + } else { + avail = 0; + hl = sl = _gf_true; + } + } + + if (type == GF_QUOTA_OPTION_TYPE_LIST) + ret = print_quota_list_usage_output(local, path, avail, sl_final, + limits, used_space, sl, hl, sl_num, + limit_set); + else + ret = print_quota_list_object_output(local, path, avail, sl_final, + limits, used_space, sl, hl, sl_num, + limit_set); out: - return ret; + return ret; } static int -print_quota_list_from_mountdir (cli_local_t *local, char *mountdir, - char *default_sl, char *path, int type) -{ - int ret = -1; - ssize_t xattr_size = 0; - quota_limits_t limits = {0,}; - quota_meta_t used_space = {0,}; - char *key = NULL; - gf_boolean_t limit_set = _gf_true; - - GF_ASSERT (local); - GF_ASSERT (mountdir); - GF_ASSERT (path); - - if (type == GF_QUOTA_OPTION_TYPE_LIST) - key = QUOTA_LIMIT_KEY; - else - key = QUOTA_LIMIT_OBJECTS_KEY; - - - ret = sys_lgetxattr (mountdir, key, (void *)&limits, sizeof (limits)); - if (ret < 0) { - gf_log ("cli", GF_LOG_ERROR, "Failed to get the xattr %s " - "on %s. Reason : %s", key, mountdir, strerror (errno)); - - switch (errno) { +print_quota_list_from_mountdir(cli_local_t *local, char *mountdir, + char *default_sl, char *path, int type) +{ + int ret = -1; + ssize_t xattr_size = 0; + quota_limits_t limits = { + 0, + }; + quota_meta_t used_space = { + 0, + }; + char *key = NULL; + gf_boolean_t limit_set = _gf_true; + + GF_ASSERT(local); + GF_ASSERT(mountdir); + GF_ASSERT(path); + + if (type == GF_QUOTA_OPTION_TYPE_LIST) + key = QUOTA_LIMIT_KEY; + else + key = QUOTA_LIMIT_OBJECTS_KEY; + + ret = sys_lgetxattr(mountdir, key, (void *)&limits, sizeof(limits)); + if (ret < 0) { + gf_log("cli", GF_LOG_ERROR, + "Failed to get the xattr %s " + "on %s. Reason : %s", + key, mountdir, strerror(errno)); + + switch (errno) { #if defined(ENODATA) - case ENODATA: + case ENODATA: #endif #if defined(ENOATTR) && (ENOATTR != ENODATA) - case ENOATTR: + case ENOATTR: #endif - /* If it's an ENOATTR, quota/inode-quota is - * configured(limit is set at least for one directory). - * The user is trying to issue 'list/list-objects' - * command for a directory on which quota limit is - * not set and we are showing the used-space in case - * of list-usage and showing (dir_count, file_count) - * in case of list-objects. Other labels are - * shown "N/A". - */ - - limit_set = _gf_false; - goto enoattr; - break; - - default: - cli_err ("%-40s %s", path, strerror (errno)); - break; - } + /* If it's an ENOATTR, quota/inode-quota is + * configured(limit is set at least for one directory). + * The user is trying to issue 'list/list-objects' + * command for a directory on which quota limit is + * not set and we are showing the used-space in case + * of list-usage and showing (dir_count, file_count) + * in case of list-objects. Other labels are + * shown "N/A". + */ - goto out; + limit_set = _gf_false; + goto enoattr; + break; + + default: + cli_err("%-40s %s", path, strerror(errno)); + break; } - limits.hl = ntoh64 (limits.hl); - limits.sl = ntoh64 (limits.sl); + goto out; + } + + limits.hl = ntoh64(limits.hl); + limits.sl = ntoh64(limits.sl); enoattr: - xattr_size = sys_lgetxattr (mountdir, QUOTA_SIZE_KEY, NULL, 0); - if (xattr_size < (sizeof (int64_t) * 2) && - type == GF_QUOTA_OPTION_TYPE_LIST_OBJECTS) { - ret = -1; + xattr_size = sys_lgetxattr(mountdir, QUOTA_SIZE_KEY, NULL, 0); + if (xattr_size < (sizeof(int64_t) * 2) && + type == GF_QUOTA_OPTION_TYPE_LIST_OBJECTS) { + ret = -1; - /* This can happen when glusterfs is upgraded from 3.6 to 3.7 - * and the xattr healing is not completed. - */ - } else if (xattr_size > (sizeof (int64_t) * 2)) { - ret = sys_lgetxattr (mountdir, QUOTA_SIZE_KEY, - &used_space, sizeof (used_space)); - } else if (xattr_size > 0) { - /* This is for compatibility. - * Older version had only file usage - */ - ret = sys_lgetxattr (mountdir, QUOTA_SIZE_KEY, - &(used_space.size), sizeof (used_space.size)); - used_space.file_count = 0; - used_space.dir_count = 0; - } else { - ret = -1; - } + /* This can happen when glusterfs is upgraded from 3.6 to 3.7 + * and the xattr healing is not completed. + */ + } else if (xattr_size > (sizeof(int64_t) * 2)) { + ret = sys_lgetxattr(mountdir, QUOTA_SIZE_KEY, &used_space, + sizeof(used_space)); + } else if (xattr_size > 0) { + /* This is for compatibility. + * Older version had only file usage + */ + ret = sys_lgetxattr(mountdir, QUOTA_SIZE_KEY, &(used_space.size), + sizeof(used_space.size)); + used_space.file_count = 0; + used_space.dir_count = 0; + } else { + ret = -1; + } - if (ret < 0) { - gf_log ("cli", GF_LOG_ERROR, "Failed to get quota size " - "on path %s: %s", mountdir, strerror (errno)); - print_quota_list_empty (path, type); - goto out; - } + if (ret < 0) { + gf_log("cli", GF_LOG_ERROR, + "Failed to get quota size " + "on path %s: %s", + mountdir, strerror(errno)); + print_quota_list_empty(path, type); + goto out; + } - used_space.size = ntoh64 (used_space.size); - used_space.file_count = ntoh64 (used_space.file_count); - used_space.dir_count = ntoh64 (used_space.dir_count); + used_space.size = ntoh64(used_space.size); + used_space.file_count = ntoh64(used_space.file_count); + used_space.dir_count = ntoh64(used_space.dir_count); - ret = print_quota_list_output (local, path, default_sl, &limits, - &used_space, type, limit_set); + ret = print_quota_list_output(local, path, default_sl, &limits, &used_space, + type, limit_set); out: - return ret; + return ret; } int -gluster_remove_auxiliary_mount (char *volname) +gluster_remove_auxiliary_mount(char *volname) { - int ret = -1; - char mountdir[PATH_MAX] = {0,}; - xlator_t *this = NULL; + int ret = -1; + char mountdir[PATH_MAX] = { + 0, + }; + xlator_t *this = NULL; - this = THIS; - GF_ASSERT (this); + this = THIS; + GF_ASSERT(this); - GLUSTERD_GET_QUOTA_LIST_MOUNT_PATH (mountdir, volname, "/"); - ret = gf_umount_lazy (this->name, mountdir, 1); - if (ret) { - gf_log("cli", GF_LOG_ERROR, "umount on %s failed, " - "reason : %s", mountdir, strerror (errno)); - } + GLUSTERD_GET_QUOTA_LIST_MOUNT_PATH(mountdir, volname, "/"); + ret = gf_umount_lazy(this->name, mountdir, 1); + if (ret) { + gf_log("cli", GF_LOG_ERROR, + "umount on %s failed, " + "reason : %s", + mountdir, strerror(errno)); + } - return ret; + return ret; } int -gf_cli_print_limit_list_from_dict (cli_local_t *local, char *volname, - dict_t *dict, char *default_sl, int count, - int op_ret, int op_errno, char *op_errstr) -{ - int ret = -1; - int i = 0; - char key[1024] = {0,}; - char mountdir[PATH_MAX] = {0,}; - char *path = NULL; - int type = -1; - - if (!dict|| count <= 0) - goto out; - - ret = dict_get_int32 (dict, "type", &type); +gf_cli_print_limit_list_from_dict(cli_local_t *local, char *volname, + dict_t *dict, char *default_sl, int count, + int op_ret, int op_errno, char *op_errstr) +{ + int ret = -1; + int i = 0; + char key[1024] = { + 0, + }; + char mountdir[PATH_MAX] = { + 0, + }; + char *path = NULL; + int type = -1; + + if (!dict || count <= 0) + goto out; + + ret = dict_get_int32(dict, "type", &type); + if (ret) { + gf_log("cli", GF_LOG_ERROR, "Failed to get quota type"); + goto out; + } + + if (global_state->mode & GLUSTER_MODE_XML) { + ret = cli_xml_output_vol_quota_limit_list_begin(local, op_ret, op_errno, + op_errstr); if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Failed to get quota type"); - goto out; + gf_log("cli", GF_LOG_ERROR, "Error outputting xml begin"); + goto out; } + } else { + print_quota_list_header(type); + } - if (global_state->mode & GLUSTER_MODE_XML) { - ret = cli_xml_output_vol_quota_limit_list_begin - (local, op_ret, op_errno, op_errstr); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, - "Error outputting xml begin"); - goto out; - } - } else { - print_quota_list_header (type); - } - - while (count--) { - snprintf (key, sizeof (key), "path%d", i++); - - ret = dict_get_str (dict, key, &path); - if (ret < 0) { - gf_log ("cli", GF_LOG_DEBUG, "Path not present in limit" - " list"); - continue; - } + while (count--) { + snprintf(key, sizeof(key), "path%d", i++); - ret = gf_canonicalize_path (path); - if (ret) - goto out; - GLUSTERD_GET_QUOTA_LIST_MOUNT_PATH (mountdir, volname, path); - ret = print_quota_list_from_mountdir (local, mountdir, - default_sl, path, type); + ret = dict_get_str(dict, key, &path); + if (ret < 0) { + gf_log("cli", GF_LOG_DEBUG, + "Path not present in limit" + " list"); + continue; } + ret = gf_canonicalize_path(path); + if (ret) + goto out; + GLUSTERD_GET_QUOTA_LIST_MOUNT_PATH(mountdir, volname, path); + ret = print_quota_list_from_mountdir(local, mountdir, default_sl, path, + type); + } + out: - return ret; + return ret; } int -print_quota_list_from_quotad (call_frame_t *frame, dict_t *rsp_dict) -{ - char *path = NULL; - char *default_sl = NULL; - int ret = -1; - cli_local_t *local = NULL; - dict_t *gd_rsp_dict = NULL; - quota_meta_t used_space = {0, }; - quota_limits_t limits = {0, }; - quota_limits_t *size_limits = NULL; - int32_t type = 0; - int32_t success_count = 0; - - GF_ASSERT (frame); - - local = frame->local; - gd_rsp_dict = local->dict; - - ret = dict_get_int32 (rsp_dict, "type", &type); +print_quota_list_from_quotad(call_frame_t *frame, dict_t *rsp_dict) +{ + char *path = NULL; + char *default_sl = NULL; + int ret = -1; + cli_local_t *local = NULL; + dict_t *gd_rsp_dict = NULL; + quota_meta_t used_space = { + 0, + }; + quota_limits_t limits = { + 0, + }; + quota_limits_t *size_limits = NULL; + int32_t type = 0; + int32_t success_count = 0; + + GF_ASSERT(frame); + + local = frame->local; + gd_rsp_dict = local->dict; + + ret = dict_get_int32(rsp_dict, "type", &type); + if (ret) { + gf_log("cli", GF_LOG_ERROR, "Failed to get type"); + goto out; + } + + ret = dict_get_str(rsp_dict, GET_ANCESTRY_PATH_KEY, &path); + if (ret) { + gf_log("cli", GF_LOG_WARNING, + "path key is not present " + "in dict"); + goto out; + } + + ret = dict_get_str(gd_rsp_dict, "default-soft-limit", &default_sl); + if (ret) { + gf_log(frame->this->name, GF_LOG_ERROR, + "failed to " + "get default soft limit"); + goto out; + } + + if (type == GF_QUOTA_OPTION_TYPE_LIST) { + ret = dict_get_bin(rsp_dict, QUOTA_LIMIT_KEY, (void **)&size_limits); if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Failed to get type"); - goto out; - } - - ret = dict_get_str (rsp_dict, GET_ANCESTRY_PATH_KEY, &path); - if (ret) { - gf_log ("cli", GF_LOG_WARNING, "path key is not present " - "in dict"); - goto out; + gf_log("cli", GF_LOG_WARNING, "limit key not present in dict on %s", + path); + goto out; } - - ret = dict_get_str (gd_rsp_dict, "default-soft-limit", &default_sl); + } else { + ret = dict_get_bin(rsp_dict, QUOTA_LIMIT_OBJECTS_KEY, + (void **)&size_limits); if (ret) { - gf_log (frame->this->name, GF_LOG_ERROR, "failed to " - "get default soft limit"); - goto out; - } - - if (type == GF_QUOTA_OPTION_TYPE_LIST) { - ret = dict_get_bin (rsp_dict, QUOTA_LIMIT_KEY, - (void **)&size_limits); - if (ret) { - gf_log ("cli", GF_LOG_WARNING, - "limit key not present in dict on %s", - path); - goto out; - } + gf_log("cli", GF_LOG_WARNING, + "object limit key not present in dict on %s", path); + goto out; + } + } + + limits.hl = ntoh64(size_limits->hl); + limits.sl = ntoh64(size_limits->sl); + + if (type == GF_QUOTA_OPTION_TYPE_LIST) + ret = quota_dict_get_meta(rsp_dict, QUOTA_SIZE_KEY, &used_space); + else + ret = quota_dict_get_inode_meta(rsp_dict, QUOTA_SIZE_KEY, &used_space); + + if (ret < 0) { + gf_log("cli", GF_LOG_WARNING, "size key not present in dict"); + print_quota_list_empty(path, type); + goto out; + } + + LOCK(&local->lock); + { + ret = dict_get_int32(gd_rsp_dict, "quota-list-success-count", + &success_count); + if (ret) + success_count = 0; + + ret = dict_set_int32(gd_rsp_dict, "quota-list-success-count", + success_count + 1); + } + UNLOCK(&local->lock); + if (ret) { + gf_log("cli", GF_LOG_ERROR, + "Failed to set " + "quota-list-success-count in dict"); + goto out; + } + + if (success_count == 0) { + if (!(global_state->mode & GLUSTER_MODE_XML)) { + print_quota_list_header(type); } else { - ret = dict_get_bin (rsp_dict, QUOTA_LIMIT_OBJECTS_KEY, - (void **)&size_limits); - if (ret) { - gf_log ("cli", GF_LOG_WARNING, - "object limit key not present in dict on %s", - path); - goto out; - } - } - - limits.hl = ntoh64 (size_limits->hl); - limits.sl = ntoh64 (size_limits->sl); - - if (type == GF_QUOTA_OPTION_TYPE_LIST) - ret = quota_dict_get_meta (rsp_dict, QUOTA_SIZE_KEY, - &used_space); - else - ret = quota_dict_get_inode_meta (rsp_dict, QUOTA_SIZE_KEY, - &used_space); - - if (ret < 0) { - gf_log ("cli", GF_LOG_WARNING, - "size key not present in dict"); - print_quota_list_empty (path, type); - goto out; - } - - LOCK (&local->lock); - { - ret = dict_get_int32 (gd_rsp_dict, "quota-list-success-count", - &success_count); - if (ret) - success_count = 0; - - ret = dict_set_int32 (gd_rsp_dict, - "quota-list-success-count", - success_count + 1); - } - UNLOCK (&local->lock); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Failed to set " - "quota-list-success-count in dict"); + ret = cli_xml_output_vol_quota_limit_list_begin(local, 0, 0, NULL); + if (ret) { + gf_log("cli", GF_LOG_ERROR, + "Error in " + "printing xml output"); goto out; + } } + } - if (success_count == 0) { - if (!(global_state->mode & GLUSTER_MODE_XML)) { - print_quota_list_header (type); - } else { - ret = cli_xml_output_vol_quota_limit_list_begin - (local, 0, 0, NULL); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Error in " - "printing xml output"); - goto out; - } - } - } - - ret = print_quota_list_output (local, path, default_sl, &limits, - &used_space, type, _gf_true); + ret = print_quota_list_output(local, path, default_sl, &limits, &used_space, + type, _gf_true); out: - return ret; + return ret; } -void* -cli_cmd_broadcast_response_detached (void *opaque) +void * +cli_cmd_broadcast_response_detached(void *opaque) { - int32_t ret = 0; + int32_t ret = 0; - ret = (intptr_t) opaque; - cli_cmd_broadcast_response (ret); + ret = (intptr_t)opaque; + cli_cmd_broadcast_response(ret); - return NULL; + return NULL; } int32_t -cli_quota_compare_path (struct list_head *list1, - struct list_head *list2) +cli_quota_compare_path(struct list_head *list1, struct list_head *list2) { - struct list_node *node1 = NULL; - struct list_node *node2 = NULL; - dict_t *dict1 = NULL; - dict_t *dict2 = NULL; - char *path1 = NULL; - char *path2 = NULL; - int ret = 0; + struct list_node *node1 = NULL; + struct list_node *node2 = NULL; + dict_t *dict1 = NULL; + dict_t *dict2 = NULL; + char *path1 = NULL; + char *path2 = NULL; + int ret = 0; - node1 = list_entry (list1, struct list_node, list); - node2 = list_entry (list2, struct list_node, list); + node1 = list_entry(list1, struct list_node, list); + node2 = list_entry(list2, struct list_node, list); - dict1 = node1->ptr; - dict2 = node2->ptr; + dict1 = node1->ptr; + dict2 = node2->ptr; - ret = dict_get_str (dict1, GET_ANCESTRY_PATH_KEY, &path1); - if (ret < 0) - return 0; + ret = dict_get_str(dict1, GET_ANCESTRY_PATH_KEY, &path1); + if (ret < 0) + return 0; - ret = dict_get_str (dict2, GET_ANCESTRY_PATH_KEY, &path2); - if (ret < 0) - return 0; + ret = dict_get_str(dict2, GET_ANCESTRY_PATH_KEY, &path2); + if (ret < 0) + return 0; - return strcmp (path1, path2); + return strcmp(path1, path2); } int -cli_quotad_getlimit_cbk (struct rpc_req *req, struct iovec *iov, - int count, void *myframe) +cli_quotad_getlimit_cbk(struct rpc_req *req, struct iovec *iov, int count, + void *myframe) { /*TODO: we need to gather the path, hard-limit, soft-limit and used space*/ - gf_cli_rsp rsp = {0,}; - int ret = -1; - dict_t *dict = NULL; - struct list_node *node = NULL; - struct list_node *tmpnode = NULL; - call_frame_t *frame = NULL; - cli_local_t *local = NULL; - int32_t list_count = 0; - pthread_t th_id = {0, }; - int32_t max_count = 0; - - GF_ASSERT (myframe); - - frame = myframe; - - GF_ASSERT (frame->local); - - local = frame->local; - - LOCK (&local->lock); - { - ret = dict_get_int32 (local->dict, "quota-list-count", - &list_count); - if (ret) - list_count = 0; - - list_count++; - ret = dict_set_int32 (local->dict, "quota-list-count", - list_count); - } - UNLOCK (&local->lock); - - if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Failed to set " - "quota-list-count in dict"); - goto out; - } - - if (-1 == req->rpc_status) { - if (list_count == 0) - cli_err ("Connection failed. Please check if quota " - "daemon is operational."); - ret = -1; - goto out; - } - - ret = xdr_to_generic (*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); - if (ret < 0) { - gf_log (frame->this->name, GF_LOG_ERROR, - "Failed to decode xdr response"); - goto out; - } - - if (rsp.op_ret) { - ret = -1; - if (strcmp (rsp.op_errstr, "")) - cli_err ("quota command failed : %s", rsp.op_errstr); - else - cli_err ("quota command : failed"); - goto out; - } - - if (rsp.dict.dict_len) { - /* Unserialize the dictionary */ - dict = dict_new (); - - ret = dict_unserialize (rsp.dict.dict_val, - rsp.dict.dict_len, - &dict); - if (ret < 0) { - gf_log ("cli", GF_LOG_ERROR, - "failed to " - "unserialize req-buffer to dictionary"); - goto out; - } - - node = list_node_add_order (dict, &local->dict_list, - cli_quota_compare_path); - if (node == NULL) { - gf_log ("cli", GF_LOG_ERROR, - "failed to add node to the list"); - dict_unref (dict); - goto out; - } + gf_cli_rsp rsp = { + 0, + }; + int ret = -1; + dict_t *dict = NULL; + struct list_node *node = NULL; + struct list_node *tmpnode = NULL; + call_frame_t *frame = NULL; + cli_local_t *local = NULL; + int32_t list_count = 0; + pthread_t th_id = { + 0, + }; + int32_t max_count = 0; + + GF_ASSERT(myframe); + + frame = myframe; + + GF_ASSERT(frame->local); + + local = frame->local; + + LOCK(&local->lock); + { + ret = dict_get_int32(local->dict, "quota-list-count", &list_count); + if (ret) + list_count = 0; + + list_count++; + ret = dict_set_int32(local->dict, "quota-list-count", list_count); + } + UNLOCK(&local->lock); + + if (ret) { + gf_log("cli", GF_LOG_ERROR, + "Failed to set " + "quota-list-count in dict"); + goto out; + } + + if (-1 == req->rpc_status) { + if (list_count == 0) + cli_err( + "Connection failed. Please check if quota " + "daemon is operational."); + ret = -1; + goto out; + } + + ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); + if (ret < 0) { + gf_log(frame->this->name, GF_LOG_ERROR, + "Failed to decode xdr response"); + goto out; + } + + if (rsp.op_ret) { + ret = -1; + if (strcmp(rsp.op_errstr, "")) + cli_err("quota command failed : %s", rsp.op_errstr); + else + cli_err("quota command : failed"); + goto out; + } - } + if (rsp.dict.dict_len) { + /* Unserialize the dictionary */ + dict = dict_new(); - ret = dict_get_int32 (local->dict, "max_count", - &max_count); + ret = dict_unserialize(rsp.dict.dict_val, rsp.dict.dict_len, &dict); if (ret < 0) { - gf_log ("cli", GF_LOG_ERROR, - "failed to get max_count"); - goto out; + gf_log("cli", GF_LOG_ERROR, + "failed to " + "unserialize req-buffer to dictionary"); + goto out; } - if (list_count == max_count) { - list_for_each_entry_safe (node, tmpnode, - &local->dict_list, list) { - dict = node->ptr; - print_quota_list_from_quotad (frame, dict); - list_node_del (node); - dict_unref (dict); - } + node = list_node_add_order(dict, &local->dict_list, + cli_quota_compare_path); + if (node == NULL) { + gf_log("cli", GF_LOG_ERROR, "failed to add node to the list"); + dict_unref(dict); + goto out; } + } + ret = dict_get_int32(local->dict, "max_count", &max_count); + if (ret < 0) { + gf_log("cli", GF_LOG_ERROR, "failed to get max_count"); + goto out; + } + if (list_count == max_count) { + list_for_each_entry_safe(node, tmpnode, &local->dict_list, list) + { + dict = node->ptr; + print_quota_list_from_quotad(frame, dict); + list_node_del(node); + dict_unref(dict); + } + } + +out: + /* Bad Fix: CLI holds the lock to process a command. + * When processing quota list command, below sequence of steps executed + * in the same thread and causing deadlock + * + * 1) CLI holds the lock + * 2) Send rpc_clnt_submit request to quotad for quota usage + * 3) If quotad is down, rpc_clnt_submit invokes cbk function with error + * 4) cbk function cli_quotad_getlimit_cbk invokes + * cli_cmd_broadcast_response which tries to hold lock to broadcast + * the results and hangs, because same thread has already holding + * the lock + * + * Broadcasting response in a separate thread which is not a + * good fix. This needs to be re-visted with better solution + */ + if (ret == -1) { + ret = pthread_create(&th_id, NULL, cli_cmd_broadcast_response_detached, + (void *)-1); + if (ret) + gf_log("cli", GF_LOG_ERROR, + "pthread_create failed: " + "%s", + strerror(errno)); + } else { + cli_cmd_broadcast_response(ret); + } -out: - /* Bad Fix: CLI holds the lock to process a command. - * When processing quota list command, below sequence of steps executed - * in the same thread and causing deadlock - * - * 1) CLI holds the lock - * 2) Send rpc_clnt_submit request to quotad for quota usage - * 3) If quotad is down, rpc_clnt_submit invokes cbk function with error - * 4) cbk function cli_quotad_getlimit_cbk invokes - * cli_cmd_broadcast_response which tries to hold lock to broadcast - * the results and hangs, because same thread has already holding - * the lock - * - * Broadcasting response in a separate thread which is not a - * good fix. This needs to be re-visted with better solution - */ - if (ret == -1) { - ret = pthread_create (&th_id, NULL, - cli_cmd_broadcast_response_detached, - (void *)-1); - if (ret) - gf_log ("cli", GF_LOG_ERROR, "pthread_create failed: " - "%s", strerror (errno)); - } else { - cli_cmd_broadcast_response (ret); - } - - free (rsp.dict.dict_val); - return ret; + free(rsp.dict.dict_val); + return ret; } int -cli_quotad_getlimit (call_frame_t *frame, xlator_t *this, void *data) +cli_quotad_getlimit(call_frame_t *frame, xlator_t *this, void *data) { - gf_cli_req req = {{0,}}; - int ret = 0; - dict_t *dict = NULL; + gf_cli_req req = {{ + 0, + }}; + int ret = 0; + dict_t *dict = NULL; - if (!frame || !this || !data) { - ret = -1; - goto out; - } + if (!frame || !this || !data) { + ret = -1; + goto out; + } - dict = data; - ret = add_cli_cmd_timeout_to_dict (dict); + dict = data; + ret = add_cli_cmd_timeout_to_dict(dict); - ret = dict_allocate_and_serialize (dict, &req.dict.dict_val, - &req.dict.dict_len); - if (ret < 0) { - gf_log (this->name, GF_LOG_ERROR, - "failed to serialize the data"); + ret = dict_allocate_and_serialize(dict, &req.dict.dict_val, + &req.dict.dict_len); + if (ret < 0) { + gf_log(this->name, GF_LOG_ERROR, "failed to serialize the data"); - goto out; - } + goto out; + } - ret = cli_cmd_submit (global_quotad_rpc, &req, frame, &cli_quotad_clnt, - GF_AGGREGATOR_GETLIMIT, NULL, - this, cli_quotad_getlimit_cbk, - (xdrproc_t) xdr_gf_cli_req); + ret = cli_cmd_submit(global_quotad_rpc, &req, frame, &cli_quotad_clnt, + GF_AGGREGATOR_GETLIMIT, NULL, this, + cli_quotad_getlimit_cbk, (xdrproc_t)xdr_gf_cli_req); out: - gf_log ("cli", GF_LOG_DEBUG, "Returning %d", ret); - return ret; + gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret); + return ret; } void -gf_cli_quota_list (cli_local_t *local, char *volname, dict_t *dict, - char *default_sl, int count, int op_ret, - int op_errno, char *op_errstr) +gf_cli_quota_list(cli_local_t *local, char *volname, dict_t *dict, + char *default_sl, int count, int op_ret, int op_errno, + char *op_errstr) { - GF_VALIDATE_OR_GOTO ("cli", volname, out); + GF_VALIDATE_OR_GOTO("cli", volname, out); - if (!connected) - goto out; + if (!connected) + goto out; - if (count > 0) - gf_cli_print_limit_list_from_dict (local, volname, dict, - default_sl, count, op_ret, - op_errno, op_errstr); + if (count > 0) + gf_cli_print_limit_list_from_dict(local, volname, dict, default_sl, + count, op_ret, op_errno, op_errstr); out: - return; + return; } int -gf_cli_quota_cbk (struct rpc_req *req, struct iovec *iov, - int count, void *myframe) -{ - gf_cli_rsp rsp = {0,}; - int ret = -1; - dict_t *dict = NULL; - char *volname = NULL; - int32_t type = 0; - call_frame_t *frame = NULL; - char *default_sl = NULL; - cli_local_t *local = NULL; - char *default_sl_dup = NULL; - int32_t entry_count = 0; - - GF_ASSERT (myframe); - - if (-1 == req->rpc_status) { - goto out; +gf_cli_quota_cbk(struct rpc_req *req, struct iovec *iov, int count, + void *myframe) +{ + gf_cli_rsp rsp = { + 0, + }; + int ret = -1; + dict_t *dict = NULL; + char *volname = NULL; + int32_t type = 0; + call_frame_t *frame = NULL; + char *default_sl = NULL; + cli_local_t *local = NULL; + char *default_sl_dup = NULL; + int32_t entry_count = 0; + + GF_ASSERT(myframe); + + if (-1 == req->rpc_status) { + goto out; + } + + frame = myframe; + + GF_ASSERT(frame->local); + + local = frame->local; + + ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); + if (ret < 0) { + gf_log(frame->this->name, GF_LOG_ERROR, + "Failed to decode xdr response"); + goto out; + } + + if (rsp.op_ret) { + ret = -1; + if (global_state->mode & GLUSTER_MODE_XML) + goto xml_output; + + if (strcmp(rsp.op_errstr, "")) { + cli_err("quota command failed : %s", rsp.op_errstr); + if (rsp.op_ret == -ENOENT) + cli_err( + "please enter the path relative to " + "the volume"); + } else { + cli_err("quota command : failed"); } - frame = myframe; - - GF_ASSERT (frame->local); + goto out; + } - local = frame->local; + if (rsp.dict.dict_len) { + /* Unserialize the dictionary */ + dict = dict_new(); - ret = xdr_to_generic (*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); + ret = dict_unserialize(rsp.dict.dict_val, rsp.dict.dict_len, &dict); if (ret < 0) { - gf_log (frame->this->name, GF_LOG_ERROR, - "Failed to decode xdr response"); - goto out; - } - - if (rsp.op_ret) { - ret = -1; - if (global_state->mode & GLUSTER_MODE_XML) - goto xml_output; - - if (strcmp (rsp.op_errstr, "")) { - cli_err ("quota command failed : %s", rsp.op_errstr); - if (rsp.op_ret == -ENOENT) - cli_err ("please enter the path relative to " - "the volume"); - } else { - cli_err ("quota command : failed"); - } - - goto out; - } - - if (rsp.dict.dict_len) { - /* Unserialize the dictionary */ - dict = dict_new (); - - ret = dict_unserialize (rsp.dict.dict_val, - rsp.dict.dict_len, - &dict); - if (ret < 0) { - gf_log ("cli", GF_LOG_ERROR, - "failed to " - "unserialize req-buffer to dictionary"); - goto out; - } + gf_log("cli", GF_LOG_ERROR, + "failed to " + "unserialize req-buffer to dictionary"); + goto out; + } + } + + gf_log("cli", GF_LOG_DEBUG, "Received resp to quota command"); + + ret = dict_get_str(dict, "volname", &volname); + if (ret) + gf_log(frame->this->name, GF_LOG_ERROR, "failed to get volname"); + + ret = dict_get_str(dict, "default-soft-limit", &default_sl); + if (ret) + gf_log(frame->this->name, GF_LOG_TRACE, + "failed to get " + "default soft limit"); + + // default-soft-limit is part of rsp_dict only iff we sent + // GLUSTER_CLI_QUOTA with type being GF_QUOTA_OPTION_TYPE_LIST + if (default_sl) { + default_sl_dup = gf_strdup(default_sl); + if (!default_sl_dup) { + ret = -1; + goto out; + } + ret = dict_set_dynstr(local->dict, "default-soft-limit", + default_sl_dup); + if (ret) { + gf_log(frame->this->name, GF_LOG_TRACE, + "failed to set default soft limit"); + GF_FREE(default_sl_dup); } + } - gf_log ("cli", GF_LOG_DEBUG, "Received resp to quota command"); + ret = dict_get_int32(dict, "type", &type); + if (ret) + gf_log(frame->this->name, GF_LOG_TRACE, "failed to get type"); - ret = dict_get_str (dict, "volname", &volname); - if (ret) - gf_log (frame->this->name, GF_LOG_ERROR, - "failed to get volname"); + ret = dict_get_int32(dict, "count", &entry_count); + if (ret) + gf_log(frame->this->name, GF_LOG_TRACE, "failed to get count"); - ret = dict_get_str (dict, "default-soft-limit", &default_sl); - if (ret) - gf_log (frame->this->name, GF_LOG_TRACE, "failed to get " - "default soft limit"); - - // default-soft-limit is part of rsp_dict only iff we sent - // GLUSTER_CLI_QUOTA with type being GF_QUOTA_OPTION_TYPE_LIST - if (default_sl) { - default_sl_dup = gf_strdup (default_sl); - if (!default_sl_dup) { - ret = -1; - goto out; - } - ret = dict_set_dynstr (local->dict, "default-soft-limit", - default_sl_dup); - if (ret) { - gf_log (frame->this->name, GF_LOG_TRACE, - "failed to set default soft limit"); - GF_FREE (default_sl_dup); - } - } - - ret = dict_get_int32 (dict, "type", &type); - if (ret) - gf_log (frame->this->name, GF_LOG_TRACE, - "failed to get type"); - - ret = dict_get_int32 (dict, "count", &entry_count); - if (ret) - gf_log (frame->this->name, GF_LOG_TRACE, "failed to get count"); + if ((type == GF_QUOTA_OPTION_TYPE_LIST) || + (type == GF_QUOTA_OPTION_TYPE_LIST_OBJECTS)) { + gf_cli_quota_list(local, volname, dict, default_sl, entry_count, + rsp.op_ret, rsp.op_errno, rsp.op_errstr); - if ((type == GF_QUOTA_OPTION_TYPE_LIST) - || (type == GF_QUOTA_OPTION_TYPE_LIST_OBJECTS)) { - gf_cli_quota_list (local, volname, dict, default_sl, - entry_count, rsp.op_ret, - rsp.op_errno, rsp.op_errstr); - - if (global_state->mode & GLUSTER_MODE_XML) { - ret = cli_xml_output_vol_quota_limit_list_end (local); - if (ret < 0) { - ret = -1; - gf_log ("cli", GF_LOG_ERROR, "Error in printing" - " xml output"); - } - goto out; - } + if (global_state->mode & GLUSTER_MODE_XML) { + ret = cli_xml_output_vol_quota_limit_list_end(local); + if (ret < 0) { + ret = -1; + gf_log("cli", GF_LOG_ERROR, + "Error in printing" + " xml output"); + } + goto out; } + } xml_output: - if (global_state->mode & GLUSTER_MODE_XML) { - ret = cli_xml_output_str ("volQuota", NULL, rsp.op_ret, - rsp.op_errno, rsp.op_errstr); - if (ret) - gf_log ("cli", GF_LOG_ERROR, - "Error outputting to xml"); - goto out; - } + if (global_state->mode & GLUSTER_MODE_XML) { + ret = cli_xml_output_str("volQuota", NULL, rsp.op_ret, rsp.op_errno, + rsp.op_errstr); + if (ret) + gf_log("cli", GF_LOG_ERROR, "Error outputting to xml"); + goto out; + } - if (!rsp.op_ret && type != GF_QUOTA_OPTION_TYPE_LIST - && type != GF_QUOTA_OPTION_TYPE_LIST_OBJECTS) - cli_out ("volume quota : success"); + if (!rsp.op_ret && type != GF_QUOTA_OPTION_TYPE_LIST && + type != GF_QUOTA_OPTION_TYPE_LIST_OBJECTS) + cli_out("volume quota : success"); - ret = rsp.op_ret; + ret = rsp.op_ret; out: - if ((type == GF_QUOTA_OPTION_TYPE_LIST) - || (type == GF_QUOTA_OPTION_TYPE_LIST_OBJECTS)) { - gluster_remove_auxiliary_mount (volname); - } + if ((type == GF_QUOTA_OPTION_TYPE_LIST) || + (type == GF_QUOTA_OPTION_TYPE_LIST_OBJECTS)) { + gluster_remove_auxiliary_mount(volname); + } - cli_cmd_broadcast_response (ret); - if (dict) - dict_unref (dict); + cli_cmd_broadcast_response(ret); + if (dict) + dict_unref(dict); - free (rsp.dict.dict_val); + free(rsp.dict.dict_val); - return ret; + return ret; } int -gf_cli_getspec_cbk (struct rpc_req *req, struct iovec *iov, - int count, void *myframe) +gf_cli_getspec_cbk(struct rpc_req *req, struct iovec *iov, int count, + void *myframe) { - gf_getspec_rsp rsp = {0,}; - int ret = -1; - char *spec = NULL; + gf_getspec_rsp rsp = { + 0, + }; + int ret = -1; + char *spec = NULL; - GF_ASSERT (myframe); + GF_ASSERT(myframe); - if (-1 == req->rpc_status) { - goto out; - } + if (-1 == req->rpc_status) { + goto out; + } - ret = xdr_to_generic (*iov, &rsp, (xdrproc_t)xdr_gf_getspec_rsp); - if (ret < 0) { - gf_log (((call_frame_t *) myframe)->this->name, GF_LOG_ERROR, - "Failed to decode xdr response"); - goto out; - } + ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_getspec_rsp); + if (ret < 0) { + gf_log(((call_frame_t *)myframe)->this->name, GF_LOG_ERROR, + "Failed to decode xdr response"); + goto out; + } - if (rsp.op_ret == -1) { - gf_log (((call_frame_t *) myframe)->this->name, GF_LOG_ERROR, - "getspec failed"); - goto out; - } + if (rsp.op_ret == -1) { + gf_log(((call_frame_t *)myframe)->this->name, GF_LOG_ERROR, + "getspec failed"); + goto out; + } - gf_log ("cli", GF_LOG_INFO, "Received resp to getspec"); + gf_log("cli", GF_LOG_INFO, "Received resp to getspec"); - spec = GF_MALLOC (rsp.op_ret + 1, cli_mt_char); - if (!spec) { - gf_log("", GF_LOG_ERROR, "out of memory"); - goto out; - } - memcpy (spec, rsp.spec, rsp.op_ret); - spec[rsp.op_ret] = '\0'; - cli_out ("%s", spec); - GF_FREE (spec); + spec = GF_MALLOC(rsp.op_ret + 1, cli_mt_char); + if (!spec) { + gf_log("", GF_LOG_ERROR, "out of memory"); + goto out; + } + memcpy(spec, rsp.spec, rsp.op_ret); + spec[rsp.op_ret] = '\0'; + cli_out("%s", spec); + GF_FREE(spec); - ret = 0; + ret = 0; out: - cli_cmd_broadcast_response (ret); - return ret; + cli_cmd_broadcast_response(ret); + return ret; } int -gf_cli_pmap_b2p_cbk (struct rpc_req *req, struct iovec *iov, - int count, void *myframe) +gf_cli_pmap_b2p_cbk(struct rpc_req *req, struct iovec *iov, int count, + void *myframe) { - pmap_port_by_brick_rsp rsp = {0,}; - int ret = -1; - char *spec = NULL; + pmap_port_by_brick_rsp rsp = { + 0, + }; + int ret = -1; + char *spec = NULL; - GF_ASSERT (myframe); + GF_ASSERT(myframe); - if (-1 == req->rpc_status) { - goto out; - } + if (-1 == req->rpc_status) { + goto out; + } - ret = xdr_to_generic (*iov, &rsp, (xdrproc_t)xdr_pmap_port_by_brick_rsp); - if (ret < 0) { - gf_log (((call_frame_t *) myframe)->this->name, GF_LOG_ERROR, - "Failed to decode xdr response"); - goto out; - } + ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_pmap_port_by_brick_rsp); + if (ret < 0) { + gf_log(((call_frame_t *)myframe)->this->name, GF_LOG_ERROR, + "Failed to decode xdr response"); + goto out; + } - if (rsp.op_ret == -1) { - gf_log (((call_frame_t *) myframe)->this->name, GF_LOG_ERROR, - "pump_b2p failed"); - goto out; - } + if (rsp.op_ret == -1) { + gf_log(((call_frame_t *)myframe)->this->name, GF_LOG_ERROR, + "pump_b2p failed"); + goto out; + } - gf_log ("cli", GF_LOG_INFO, "Received resp to pmap b2p"); + gf_log("cli", GF_LOG_INFO, "Received resp to pmap b2p"); - cli_out ("%d", rsp.port); - GF_FREE (spec); + cli_out("%d", rsp.port); + GF_FREE(spec); - ret = rsp.op_ret; + ret = rsp.op_ret; out: - cli_cmd_broadcast_response (ret); - return ret; + cli_cmd_broadcast_response(ret); + return ret; } - int32_t -gf_cli_probe (call_frame_t *frame, xlator_t *this, - void *data) +gf_cli_probe(call_frame_t *frame, xlator_t *this, void *data) { - gf_cli_req req = {{0,},}; - int ret = 0; - dict_t *dict = NULL; - int port = 0; - - if (!frame || !this || !data) { - ret = -1; - goto out; - } - - dict = data; - - ret = dict_get_int32 (dict, "port", &port); - if (ret) { - ret = dict_set_int32 (dict, "port", CLI_GLUSTERD_PORT); - if (ret) - goto out; - } + gf_cli_req req = { + { + 0, + }, + }; + int ret = 0; + dict_t *dict = NULL; + int port = 0; + + if (!frame || !this || !data) { + ret = -1; + goto out; + } + + dict = data; + + ret = dict_get_int32(dict, "port", &port); + if (ret) { + ret = dict_set_int32(dict, "port", CLI_GLUSTERD_PORT); + if (ret) + goto out; + } - ret = cli_to_glusterd (&req, frame, gf_cli_probe_cbk, - (xdrproc_t) xdr_gf_cli_req, dict, - GLUSTER_CLI_PROBE, this, cli_rpc_prog, NULL); + ret = cli_to_glusterd(&req, frame, gf_cli_probe_cbk, + (xdrproc_t)xdr_gf_cli_req, dict, GLUSTER_CLI_PROBE, + this, cli_rpc_prog, NULL); out: - GF_FREE (req.dict.dict_val); - gf_log ("cli", GF_LOG_DEBUG, "Returning %d", ret); + GF_FREE(req.dict.dict_val); + gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret); - return ret; + return ret; } int32_t -gf_cli_deprobe (call_frame_t *frame, xlator_t *this, - void *data) +gf_cli_deprobe(call_frame_t *frame, xlator_t *this, void *data) { - gf_cli_req req = {{0,},}; - int ret = 0; - dict_t *dict = NULL; - int port = 0; - int flags = 0; - - if (!frame || !this || !data) { - ret = -1; - goto out; - } - - dict = data; - ret = dict_get_int32 (dict, "port", &port); - if (ret) { - ret = dict_set_int32 (dict, "port", CLI_GLUSTERD_PORT); - if (ret) - goto out; - } + gf_cli_req req = { + { + 0, + }, + }; + int ret = 0; + dict_t *dict = NULL; + int port = 0; + int flags = 0; + + if (!frame || !this || !data) { + ret = -1; + goto out; + } + + dict = data; + ret = dict_get_int32(dict, "port", &port); + if (ret) { + ret = dict_set_int32(dict, "port", CLI_GLUSTERD_PORT); + if (ret) + goto out; + } - ret = dict_get_int32 (dict, "flags", &flags); - if (ret) { - ret = dict_set_int32 (dict, "flags", 0); - if (ret) - goto out; - } + ret = dict_get_int32(dict, "flags", &flags); + if (ret) { + ret = dict_set_int32(dict, "flags", 0); + if (ret) + goto out; + } - ret = cli_to_glusterd (&req, frame, gf_cli_deprobe_cbk, - (xdrproc_t)xdr_gf_cli_req, dict, - GLUSTER_CLI_DEPROBE, this, cli_rpc_prog, NULL); + ret = cli_to_glusterd(&req, frame, gf_cli_deprobe_cbk, + (xdrproc_t)xdr_gf_cli_req, dict, GLUSTER_CLI_DEPROBE, + this, cli_rpc_prog, NULL); out: - GF_FREE (req.dict.dict_val); - gf_log ("cli", GF_LOG_DEBUG, "Returning %d", ret); + GF_FREE(req.dict.dict_val); + gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret); - return ret; + return ret; } int32_t -gf_cli_list_friends (call_frame_t *frame, xlator_t *this, - void *data) +gf_cli_list_friends(call_frame_t *frame, xlator_t *this, void *data) { - gf1_cli_peer_list_req req = {0,}; - int ret = 0; - unsigned long flags = 0; + gf1_cli_peer_list_req req = { + 0, + }; + int ret = 0; + unsigned long flags = 0; - if (!frame || !this) { - ret = -1; - goto out; - } + if (!frame || !this) { + ret = -1; + goto out; + } - GF_ASSERT (frame->local == NULL); + GF_ASSERT(frame->local == NULL); - flags = (long)data; - req.flags = flags; - frame->local = (void*)flags; - ret = cli_cmd_submit (NULL, &req, frame, cli_rpc_prog, - GLUSTER_CLI_LIST_FRIENDS, NULL, - this, gf_cli_list_friends_cbk, - (xdrproc_t) xdr_gf1_cli_peer_list_req); + flags = (long)data; + req.flags = flags; + frame->local = (void *)flags; + ret = cli_cmd_submit( + NULL, &req, frame, cli_rpc_prog, GLUSTER_CLI_LIST_FRIENDS, NULL, this, + gf_cli_list_friends_cbk, (xdrproc_t)xdr_gf1_cli_peer_list_req); out: - if (ret && frame) { - /* - * If everything goes fine, gf_cli_list_friends_cbk() - * [invoked through cli_cmd_submit()]resets the - * frame->local to NULL. In case cli_cmd_submit() - * fails in between, RESET frame->local here. - */ - frame->local = NULL; - } - gf_log ("cli", GF_LOG_DEBUG, "Returning %d", ret); - return ret; + if (ret && frame) { + /* + * If everything goes fine, gf_cli_list_friends_cbk() + * [invoked through cli_cmd_submit()]resets the + * frame->local to NULL. In case cli_cmd_submit() + * fails in between, RESET frame->local here. + */ + frame->local = NULL; + } + gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret); + return ret; } int32_t -gf_cli_get_state (call_frame_t *frame, xlator_t *this, void *data) +gf_cli_get_state(call_frame_t *frame, xlator_t *this, void *data) { - gf_cli_req req = {{0,},}; - int ret = 0; - dict_t *dict = NULL; + gf_cli_req req = { + { + 0, + }, + }; + int ret = 0; + dict_t *dict = NULL; - if (!frame || !this || !data) { - ret = -1; - goto out; - } + if (!frame || !this || !data) { + ret = -1; + goto out; + } - dict = data; + dict = data; - ret = cli_to_glusterd (&req, frame, gf_cli_get_state_cbk, - (xdrproc_t) xdr_gf_cli_req, dict, - GLUSTER_CLI_GET_STATE, this, cli_rpc_prog, - NULL); + ret = cli_to_glusterd(&req, frame, gf_cli_get_state_cbk, + (xdrproc_t)xdr_gf_cli_req, dict, + GLUSTER_CLI_GET_STATE, this, cli_rpc_prog, NULL); out: - gf_log ("cli", GF_LOG_DEBUG, "Returning %d", ret); + gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret); - return ret; + return ret; } int32_t -gf_cli_get_next_volume (call_frame_t *frame, xlator_t *this, - void *data) +gf_cli_get_next_volume(call_frame_t *frame, xlator_t *this, void *data) { + int ret = 0; + cli_cmd_volume_get_ctx_t *ctx = NULL; + cli_local_t *local = NULL; - int ret = 0; - cli_cmd_volume_get_ctx_t *ctx = NULL; - cli_local_t *local = NULL; + if (!frame || !this || !data) { + ret = -1; + goto out; + } - if (!frame || !this || !data) { - ret = -1; - goto out; - } + ctx = data; + local = frame->local; - ctx = data; - local = frame->local; - - if (global_state->mode & GLUSTER_MODE_XML) { - ret = cli_xml_output_vol_info_begin (local, 0, 0, ""); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Error outputting to xml"); - goto out; - } + if (global_state->mode & GLUSTER_MODE_XML) { + ret = cli_xml_output_vol_info_begin(local, 0, 0, ""); + if (ret) { + gf_log("cli", GF_LOG_ERROR, "Error outputting to xml"); + goto out; } + } - ret = gf_cli_get_volume (frame, this, data); - + ret = gf_cli_get_volume(frame, this, data); - if (!local || !local->get_vol.volname) { - if ((global_state->mode & GLUSTER_MODE_XML)) - goto end_xml; + if (!local || !local->get_vol.volname) { + if ((global_state->mode & GLUSTER_MODE_XML)) + goto end_xml; - cli_err ("No volumes present"); - goto out; - } + cli_err("No volumes present"); + goto out; + } + ctx->volname = local->get_vol.volname; + while (ctx->volname) { + ret = gf_cli_get_volume(frame, this, ctx); + if (ret) + goto out; ctx->volname = local->get_vol.volname; - - while (ctx->volname) { - ret = gf_cli_get_volume (frame, this, ctx); - if (ret) - goto out; - ctx->volname = local->get_vol.volname; - } + } end_xml: - if (global_state->mode & GLUSTER_MODE_XML) { - ret = cli_xml_output_vol_info_end (local); - if (ret) - gf_log ("cli", GF_LOG_ERROR, "Error outputting to xml"); - } + if (global_state->mode & GLUSTER_MODE_XML) { + ret = cli_xml_output_vol_info_end(local); + if (ret) + gf_log("cli", GF_LOG_ERROR, "Error outputting to xml"); + } out: - gf_log ("cli", GF_LOG_DEBUG, "Returning %d", ret); - return ret; + gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret); + return ret; } int32_t -gf_cli_get_volume (call_frame_t *frame, xlator_t *this, - void *data) +gf_cli_get_volume(call_frame_t *frame, xlator_t *this, void *data) { - gf_cli_req req = {{0,}}; - int ret = 0; - cli_cmd_volume_get_ctx_t *ctx = NULL; - dict_t *dict = NULL; - int32_t flags = 0; + gf_cli_req req = {{ + 0, + }}; + int ret = 0; + cli_cmd_volume_get_ctx_t *ctx = NULL; + dict_t *dict = NULL; + int32_t flags = 0; - if (!frame || !this || !data) { - ret = -1; - goto out; - } + if (!frame || !this || !data) { + ret = -1; + goto out; + } - ctx = data; + ctx = data; - dict = dict_new (); - if (!dict) - goto out; + dict = dict_new(); + if (!dict) + goto out; - if (ctx->volname) { - ret = dict_set_str (dict, "volname", ctx->volname); - if (ret) - goto out; - } + if (ctx->volname) { + ret = dict_set_str(dict, "volname", ctx->volname); + if (ret) + goto out; + } - flags = ctx->flags; - ret = dict_set_int32 (dict, "flags", flags); - if (ret) { - gf_log (frame->this->name, GF_LOG_ERROR, "failed to set flags"); - goto out; - } + flags = ctx->flags; + ret = dict_set_int32(dict, "flags", flags); + if (ret) { + gf_log(frame->this->name, GF_LOG_ERROR, "failed to set flags"); + goto out; + } - ret = dict_allocate_and_serialize (dict, &req.dict.dict_val, - &req.dict.dict_len); - if (ret) { - gf_log (frame->this->name, GF_LOG_ERROR, - "failed to serialize dict"); - goto out; - } + ret = dict_allocate_and_serialize(dict, &req.dict.dict_val, + &req.dict.dict_len); + if (ret) { + gf_log(frame->this->name, GF_LOG_ERROR, "failed to serialize dict"); + goto out; + } - ret = cli_cmd_submit (NULL, &req, frame, cli_rpc_prog, - GLUSTER_CLI_GET_VOLUME, NULL, - this, gf_cli_get_volume_cbk, - (xdrproc_t) xdr_gf_cli_req); + ret = cli_cmd_submit(NULL, &req, frame, cli_rpc_prog, + GLUSTER_CLI_GET_VOLUME, NULL, this, + gf_cli_get_volume_cbk, (xdrproc_t)xdr_gf_cli_req); out: - if (dict) - dict_unref (dict); + if (dict) + dict_unref(dict); - GF_FREE (req.dict.dict_val); + GF_FREE(req.dict.dict_val); - gf_log ("cli", GF_LOG_DEBUG, "Returning %d", ret); - return ret; + gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret); + return ret; } int32_t -gf_cli3_1_uuid_get (call_frame_t *frame, xlator_t *this, - void *data) +gf_cli3_1_uuid_get(call_frame_t *frame, xlator_t *this, void *data) { - gf_cli_req req = {{0,}}; - int ret = 0; - dict_t *dict = NULL; + gf_cli_req req = {{ + 0, + }}; + int ret = 0; + dict_t *dict = NULL; - if (!frame || !this || !data) { - ret = -1; - goto out; - } + if (!frame || !this || !data) { + ret = -1; + goto out; + } - dict = data; - ret = cli_to_glusterd (&req, frame, gf_cli3_1_uuid_get_cbk, - (xdrproc_t)xdr_gf_cli_req, dict, - GLUSTER_CLI_UUID_GET, this, cli_rpc_prog, - NULL); + dict = data; + ret = cli_to_glusterd(&req, frame, gf_cli3_1_uuid_get_cbk, + (xdrproc_t)xdr_gf_cli_req, dict, GLUSTER_CLI_UUID_GET, + this, cli_rpc_prog, NULL); out: - gf_log ("cli", GF_LOG_DEBUG, "Returning %d", ret); - return ret; + gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret); + return ret; } int32_t -gf_cli3_1_uuid_reset (call_frame_t *frame, xlator_t *this, - void *data) +gf_cli3_1_uuid_reset(call_frame_t *frame, xlator_t *this, void *data) { - gf_cli_req req = {{0,}}; - int ret = 0; - dict_t *dict = NULL; + gf_cli_req req = {{ + 0, + }}; + int ret = 0; + dict_t *dict = NULL; - if (!frame || !this || !data) { - ret = -1; - goto out; - } + if (!frame || !this || !data) { + ret = -1; + goto out; + } - dict = data; - ret = cli_to_glusterd (&req, frame, gf_cli3_1_uuid_reset_cbk, - (xdrproc_t)xdr_gf_cli_req, dict, - GLUSTER_CLI_UUID_RESET, this, cli_rpc_prog, - NULL); + dict = data; + ret = cli_to_glusterd(&req, frame, gf_cli3_1_uuid_reset_cbk, + (xdrproc_t)xdr_gf_cli_req, dict, + GLUSTER_CLI_UUID_RESET, this, cli_rpc_prog, NULL); out: - gf_log ("cli", GF_LOG_DEBUG, "Returning %d", ret); - return ret; + gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret); + return ret; } int32_t -gf_cli_create_volume (call_frame_t *frame, xlator_t *this, - void *data) +gf_cli_create_volume(call_frame_t *frame, xlator_t *this, void *data) { - gf_cli_req req = {{0,}}; - int ret = 0; - dict_t *dict = NULL; + gf_cli_req req = {{ + 0, + }}; + int ret = 0; + dict_t *dict = NULL; - if (!frame || !this || !data) { - ret = -1; - goto out; - } + if (!frame || !this || !data) { + ret = -1; + goto out; + } - dict = data; + dict = data; - ret = cli_to_glusterd (&req, frame, gf_cli_create_volume_cbk, - (xdrproc_t) xdr_gf_cli_req, dict, - GLUSTER_CLI_CREATE_VOLUME, this, cli_rpc_prog, - NULL); + ret = cli_to_glusterd(&req, frame, gf_cli_create_volume_cbk, + (xdrproc_t)xdr_gf_cli_req, dict, + GLUSTER_CLI_CREATE_VOLUME, this, cli_rpc_prog, NULL); out: - gf_log ("cli", GF_LOG_DEBUG, "Returning %d", ret); + gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret); - GF_FREE (req.dict.dict_val); + GF_FREE(req.dict.dict_val); - return ret; + return ret; } int32_t -gf_cli_delete_volume (call_frame_t *frame, xlator_t *this, - void *data) +gf_cli_delete_volume(call_frame_t *frame, xlator_t *this, void *data) { - gf_cli_req req = {{0,}}; - int ret = 0; - dict_t *dict = NULL; + gf_cli_req req = {{ + 0, + }}; + int ret = 0; + dict_t *dict = NULL; - if (!frame || !this || !data) { - ret = -1; - goto out; - } + if (!frame || !this || !data) { + ret = -1; + goto out; + } - dict = data; + dict = data; - ret = cli_to_glusterd (&req, frame, gf_cli_delete_volume_cbk, - (xdrproc_t) xdr_gf_cli_req, dict, - GLUSTER_CLI_DELETE_VOLUME, this, cli_rpc_prog, - NULL); + ret = cli_to_glusterd(&req, frame, gf_cli_delete_volume_cbk, + (xdrproc_t)xdr_gf_cli_req, dict, + GLUSTER_CLI_DELETE_VOLUME, this, cli_rpc_prog, NULL); out: - GF_FREE (req.dict.dict_val); - gf_log ("cli", GF_LOG_DEBUG, "Returning %d", ret); + GF_FREE(req.dict.dict_val); + gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret); - return ret; + return ret; } int32_t -gf_cli_start_volume (call_frame_t *frame, xlator_t *this, - void *data) +gf_cli_start_volume(call_frame_t *frame, xlator_t *this, void *data) { - gf_cli_req req = {{0,}}; - int ret = 0; - dict_t *dict = NULL; + gf_cli_req req = {{ + 0, + }}; + int ret = 0; + dict_t *dict = NULL; - if (!frame || !this || !data) { - ret = -1; - goto out; - } + if (!frame || !this || !data) { + ret = -1; + goto out; + } - dict = data; + dict = data; - ret = cli_to_glusterd (&req, frame, gf_cli_start_volume_cbk, - (xdrproc_t) xdr_gf_cli_req, dict, - GLUSTER_CLI_START_VOLUME, this, cli_rpc_prog, - NULL); + ret = cli_to_glusterd(&req, frame, gf_cli_start_volume_cbk, + (xdrproc_t)xdr_gf_cli_req, dict, + GLUSTER_CLI_START_VOLUME, this, cli_rpc_prog, NULL); out: - gf_log ("cli", GF_LOG_DEBUG, "Returning %d", ret); + gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret); - return ret; + return ret; } int32_t -gf_cli_stop_volume (call_frame_t *frame, xlator_t *this, - void *data) +gf_cli_stop_volume(call_frame_t *frame, xlator_t *this, void *data) { - gf_cli_req req = {{0,}}; - int ret = 0; - dict_t *dict = data; + gf_cli_req req = {{ + 0, + }}; + int ret = 0; + dict_t *dict = data; - if (!frame || !this || !data) { - ret = -1; - goto out; - } + if (!frame || !this || !data) { + ret = -1; + goto out; + } - dict = data; + dict = data; - ret = cli_to_glusterd (&req, frame, gf_cli_stop_volume_cbk, - (xdrproc_t) xdr_gf_cli_req, dict, - GLUSTER_CLI_STOP_VOLUME, this, cli_rpc_prog, - NULL); + ret = cli_to_glusterd(&req, frame, gf_cli_stop_volume_cbk, + (xdrproc_t)xdr_gf_cli_req, dict, + GLUSTER_CLI_STOP_VOLUME, this, cli_rpc_prog, NULL); out: - gf_log ("cli", GF_LOG_DEBUG, "Returning %d", ret); + gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret); - return ret; + return ret; } int32_t -gf_cli_defrag_volume (call_frame_t *frame, xlator_t *this, - void *data) +gf_cli_defrag_volume(call_frame_t *frame, xlator_t *this, void *data) { - gf_cli_req req = {{0,}}; - int ret = 0; - dict_t *dict = NULL; + gf_cli_req req = {{ + 0, + }}; + int ret = 0; + dict_t *dict = NULL; - if (!frame || !this || !data) { - ret = -1; - goto out; - } + if (!frame || !this || !data) { + ret = -1; + goto out; + } - dict = data; + dict = data; - ret = cli_to_glusterd (&req, frame, gf_cli_defrag_volume_cbk, - (xdrproc_t) xdr_gf_cli_req, dict, - GLUSTER_CLI_DEFRAG_VOLUME, this, cli_rpc_prog, - NULL); + ret = cli_to_glusterd(&req, frame, gf_cli_defrag_volume_cbk, + (xdrproc_t)xdr_gf_cli_req, dict, + GLUSTER_CLI_DEFRAG_VOLUME, this, cli_rpc_prog, NULL); out: - gf_log ("cli", GF_LOG_DEBUG, "Returning %d", ret); + gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret); - return ret; + return ret; } int32_t -gf_cli_rename_volume (call_frame_t *frame, xlator_t *this, - void *data) +gf_cli_rename_volume(call_frame_t *frame, xlator_t *this, void *data) { - gf_cli_req req = {{0,}}; - int ret = 0; - dict_t *dict = NULL; + gf_cli_req req = {{ + 0, + }}; + int ret = 0; + dict_t *dict = NULL; - if (!frame || !this || !data) { - ret = -1; - goto out; - } - - dict = data; + if (!frame || !this || !data) { + ret = -1; + goto out; + } - ret = dict_allocate_and_serialize (dict, &req.dict.dict_val, - &req.dict.dict_len); - if (ret < 0) { - gf_log (this->name, GF_LOG_ERROR, - "failed to serialize the data"); + dict = data; - goto out; - } + ret = dict_allocate_and_serialize(dict, &req.dict.dict_val, + &req.dict.dict_len); + if (ret < 0) { + gf_log(this->name, GF_LOG_ERROR, "failed to serialize the data"); + goto out; + } - ret = cli_cmd_submit (NULL, &req, frame, cli_rpc_prog, - GLUSTER_CLI_RENAME_VOLUME, NULL, - this, gf_cli_rename_volume_cbk, - (xdrproc_t) xdr_gf_cli_req); + ret = cli_cmd_submit(NULL, &req, frame, cli_rpc_prog, + GLUSTER_CLI_RENAME_VOLUME, NULL, this, + gf_cli_rename_volume_cbk, (xdrproc_t)xdr_gf_cli_req); out: - gf_log ("cli", GF_LOG_DEBUG, "Returning %d", ret); + gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret); - return ret; + return ret; } int32_t -gf_cli_reset_volume (call_frame_t *frame, xlator_t *this, - void *data) +gf_cli_reset_volume(call_frame_t *frame, xlator_t *this, void *data) { - gf_cli_req req = {{0,} }; - int ret = 0; - dict_t *dict = NULL; + gf_cli_req req = {{ + 0, + }}; + int ret = 0; + dict_t *dict = NULL; - if (!frame || !this || !data) { - ret = -1; - goto out; - } + if (!frame || !this || !data) { + ret = -1; + goto out; + } - dict = data; + dict = data; - ret = cli_to_glusterd (&req, frame, gf_cli_reset_volume_cbk, - (xdrproc_t) xdr_gf_cli_req, dict, - GLUSTER_CLI_RESET_VOLUME, this, cli_rpc_prog, - NULL); + ret = cli_to_glusterd(&req, frame, gf_cli_reset_volume_cbk, + (xdrproc_t)xdr_gf_cli_req, dict, + GLUSTER_CLI_RESET_VOLUME, this, cli_rpc_prog, NULL); out: - gf_log ("cli", GF_LOG_DEBUG, "Returning %d", ret); - return ret; + gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret); + return ret; } int32_t -gf_cli_set_volume (call_frame_t *frame, xlator_t *this, - void *data) +gf_cli_set_volume(call_frame_t *frame, xlator_t *this, void *data) { - gf_cli_req req = {{0,} }; - int ret = 0; - dict_t *dict = NULL; + gf_cli_req req = {{ + 0, + }}; + int ret = 0; + dict_t *dict = NULL; - if (!frame || !this || !data) { - ret = -1; - goto out; - } + if (!frame || !this || !data) { + ret = -1; + goto out; + } - dict = data; + dict = data; - ret = cli_to_glusterd (&req, frame, gf_cli_set_volume_cbk, - (xdrproc_t) xdr_gf_cli_req, dict, - GLUSTER_CLI_SET_VOLUME, this, cli_rpc_prog, - NULL); + ret = cli_to_glusterd(&req, frame, gf_cli_set_volume_cbk, + (xdrproc_t)xdr_gf_cli_req, dict, + GLUSTER_CLI_SET_VOLUME, this, cli_rpc_prog, NULL); out: - gf_log ("cli", GF_LOG_DEBUG, "Returning %d", ret); + gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret); - return ret; + return ret; } int32_t -gf_cli_add_brick (call_frame_t *frame, xlator_t *this, - void *data) +gf_cli_add_brick(call_frame_t *frame, xlator_t *this, void *data) { - gf_cli_req req = {{0,} }; - int ret = 0; - dict_t *dict = NULL; - char *volname = NULL; - int32_t count = 0; + gf_cli_req req = {{ + 0, + }}; + int ret = 0; + dict_t *dict = NULL; + char *volname = NULL; + int32_t count = 0; - if (!frame || !this || !data) { - ret = -1; - goto out; - } + if (!frame || !this || !data) { + ret = -1; + goto out; + } - dict = data; + dict = data; - ret = dict_get_str (dict, "volname", &volname); + ret = dict_get_str(dict, "volname", &volname); - if (ret) - goto out; + if (ret) + goto out; - ret = dict_get_int32 (dict, "count", &count); - if (ret) - goto out; + ret = dict_get_int32(dict, "count", &count); + if (ret) + goto out; - ret = cli_to_glusterd (&req, frame, gf_cli_add_brick_cbk, - (xdrproc_t) xdr_gf_cli_req, dict, - GLUSTER_CLI_ADD_BRICK, this, cli_rpc_prog, NULL); + ret = cli_to_glusterd(&req, frame, gf_cli_add_brick_cbk, + (xdrproc_t)xdr_gf_cli_req, dict, + GLUSTER_CLI_ADD_BRICK, this, cli_rpc_prog, NULL); out: - gf_log ("cli", GF_LOG_DEBUG, "Returning %d", ret); + gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret); - GF_FREE (req.dict.dict_val); + GF_FREE(req.dict.dict_val); - return ret; + return ret; } int32_t -gf_cli_tier (call_frame_t *frame, xlator_t *this, - void *data) +gf_cli_tier(call_frame_t *frame, xlator_t *this, void *data) { - int ret = 0; - int32_t command = 0; - gf_cli_req req = { {0,} }; - dict_t *dict = NULL; + int ret = 0; + int32_t command = 0; + gf_cli_req req = {{ + 0, + }}; + dict_t *dict = NULL; - if (!frame || !this || !data) { - ret = -1; - goto out; - } - dict = data; + if (!frame || !this || !data) { + ret = -1; + goto out; + } + dict = data; - ret = dict_get_int32 (dict, "rebalance-command", &command); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Failed to get rebalance-command"); - goto out; - } + ret = dict_get_int32(dict, "rebalance-command", &command); + if (ret) { + gf_log("cli", GF_LOG_ERROR, "Failed to get rebalance-command"); + goto out; + } - ret = cli_to_glusterd (&req, frame, gf_cli_defrag_volume_cbk, - (xdrproc_t) xdr_gf_cli_req, dict, - GLUSTER_CLI_TIER, this, cli_rpc_prog, - NULL); + ret = cli_to_glusterd(&req, frame, gf_cli_defrag_volume_cbk, + (xdrproc_t)xdr_gf_cli_req, dict, GLUSTER_CLI_TIER, + this, cli_rpc_prog, NULL); out: - gf_log ("cli", GF_LOG_DEBUG, "Returning %d", ret); + gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret); - GF_FREE (req.dict.dict_val); + GF_FREE(req.dict.dict_val); - return ret; + return ret; } int32_t -gf_cli_add_tier_brick (call_frame_t *frame, xlator_t *this, - void *data) +gf_cli_add_tier_brick(call_frame_t *frame, xlator_t *this, void *data) { - gf_cli_req req = { {0,} }; - int ret = 0; - dict_t *dict = NULL; + gf_cli_req req = {{ + 0, + }}; + int ret = 0; + dict_t *dict = NULL; - if (!frame || !this || !data) { - ret = -1; - goto out; - } + if (!frame || !this || !data) { + ret = -1; + goto out; + } - dict = data; + dict = data; - ret = cli_to_glusterd (&req, frame, gf_cli_add_tier_brick_cbk, - (xdrproc_t) xdr_gf_cli_req, dict, - GLUSTER_CLI_ADD_TIER_BRICK, this, - cli_rpc_prog, NULL); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Failed to send request to " - "glusterd"); - goto out; - } + ret = cli_to_glusterd(&req, frame, gf_cli_add_tier_brick_cbk, + (xdrproc_t)xdr_gf_cli_req, dict, + GLUSTER_CLI_ADD_TIER_BRICK, this, cli_rpc_prog, NULL); + if (ret) { + gf_log("cli", GF_LOG_ERROR, + "Failed to send request to " + "glusterd"); + goto out; + } out: - gf_log ("cli", GF_LOG_DEBUG, "Returning %d", ret); + gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret); - GF_FREE (req.dict.dict_val); - return ret; + GF_FREE(req.dict.dict_val); + return ret; } - int32_t -gf_cli_attach_tier (call_frame_t *frame, xlator_t *this, - void *data) -{ - gf_cli_req req = {{0,} }; - int ret = 0; - dict_t *dict = NULL; - dict_t *newdict = NULL; - char *tierwords[] = {"volume", "tier", "", - "start", NULL}; - const char **words = (const char **)tierwords; - char *volname = NULL; - cli_local_t *local = NULL; - cli_local_t *oldlocal = NULL; - - if (!frame || !this || !data) { - ret = -1; - goto out; - } - - dict = data; - - ret = cli_to_glusterd (&req, frame, gf_cli_attach_tier_cbk, - (xdrproc_t) xdr_gf_cli_req, dict, - GLUSTER_CLI_ATTACH_TIER, this, - cli_rpc_prog, NULL); - if (ret) - goto out; - ret = dict_get_str (dict, "volname", &volname); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Failed to get volume name"); - goto notify_cli; - } - - words[2] = volname; - ret = cli_cmd_volume_old_tier_parse ((const char **)words, - 4, &newdict); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Failed to parse tier start " - "command"); - goto notify_cli; - } - - gf_log ("cli", GF_LOG_DEBUG, "Sending tier start"); - - oldlocal = frame->local; - CLI_LOCAL_INIT (local, words, frame, newdict); - ret = gf_cli_tier (frame, this, newdict); - frame->local = oldlocal; - cli_local_wipe (local); +gf_cli_attach_tier(call_frame_t *frame, xlator_t *this, void *data) +{ + gf_cli_req req = {{ + 0, + }}; + int ret = 0; + dict_t *dict = NULL; + dict_t *newdict = NULL; + char *tierwords[] = {"volume", "tier", "", "start", NULL}; + const char **words = (const char **)tierwords; + char *volname = NULL; + cli_local_t *local = NULL; + cli_local_t *oldlocal = NULL; + + if (!frame || !this || !data) { + ret = -1; + goto out; + } + + dict = data; + + ret = cli_to_glusterd(&req, frame, gf_cli_attach_tier_cbk, + (xdrproc_t)xdr_gf_cli_req, dict, + GLUSTER_CLI_ATTACH_TIER, this, cli_rpc_prog, NULL); + if (ret) + goto out; + ret = dict_get_str(dict, "volname", &volname); + if (ret) { + gf_log("cli", GF_LOG_ERROR, "Failed to get volume name"); + goto notify_cli; + } + + words[2] = volname; + ret = cli_cmd_volume_old_tier_parse((const char **)words, 4, &newdict); + if (ret) { + gf_log("cli", GF_LOG_ERROR, + "Failed to parse tier start " + "command"); + goto notify_cli; + } + + gf_log("cli", GF_LOG_DEBUG, "Sending tier start"); + + oldlocal = frame->local; + CLI_LOCAL_INIT(local, words, frame, newdict); + ret = gf_cli_tier(frame, this, newdict); + frame->local = oldlocal; + cli_local_wipe(local); notify_cli: - if (ret) { - cli_out ("Failed to run tier start. Please execute tier start " - "command explicitly"); - cli_out ("Usage : gluster volume tier <volname> start"); - } + if (ret) { + cli_out( + "Failed to run tier start. Please execute tier start " + "command explicitly"); + cli_out("Usage : gluster volume tier <volname> start"); + } out: - gf_log ("cli", GF_LOG_DEBUG, "Returning %d", ret); + gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret); - GF_FREE (req.dict.dict_val); - return ret; + GF_FREE(req.dict.dict_val); + return ret; } int32_t -gf_cli_remove_tier_brick (call_frame_t *frame, xlator_t *this, - void *data) +gf_cli_remove_tier_brick(call_frame_t *frame, xlator_t *this, void *data) { - gf_cli_req status_req = { {0,} }; - int ret = 0; - dict_t *dict = NULL; - int32_t command = 0; - char *volname = NULL; + gf_cli_req status_req = {{ + 0, + }}; + int ret = 0; + dict_t *dict = NULL; + int32_t command = 0; + char *volname = NULL; - if (!frame || !this || !data) { - ret = -1; - goto out; - } + if (!frame || !this || !data) { + ret = -1; + goto out; + } - dict = data; + dict = data; - ret = dict_get_str (dict, "volname", &volname); - if (ret) - goto out; + ret = dict_get_str(dict, "volname", &volname); + if (ret) + goto out; - ret = dict_get_int32 (dict, "command", &command); - if (ret) - goto out; + ret = dict_get_int32(dict, "command", &command); + if (ret) + goto out; - ret = dict_set_int32 (dict, "rebalance-command", (int32_t) command); - if (ret) { - gf_log (this->name, GF_LOG_ERROR, - "Failed to set dict"); - goto out; - } + ret = dict_set_int32(dict, "rebalance-command", (int32_t)command); + if (ret) { + gf_log(this->name, GF_LOG_ERROR, "Failed to set dict"); + goto out; + } - ret = cli_to_glusterd (&status_req, frame, - gf_cli_remove_tier_brick_cbk, - (xdrproc_t) xdr_gf_cli_req, dict, - GLUSTER_CLI_TIER, this, - cli_rpc_prog, NULL); + ret = cli_to_glusterd(&status_req, frame, gf_cli_remove_tier_brick_cbk, + (xdrproc_t)xdr_gf_cli_req, dict, GLUSTER_CLI_TIER, + this, cli_rpc_prog, NULL); out: - gf_log ("cli", GF_LOG_DEBUG, "Returning %d", ret); + gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret); - GF_FREE (status_req.dict.dict_val); + GF_FREE(status_req.dict.dict_val); - return ret; + return ret; } int32_t -gf_cli_remove_brick (call_frame_t *frame, xlator_t *this, - void *data) -{ - gf_cli_req req = {{0,}};; - gf_cli_req status_req = {{0,}};; - int ret = 0; - dict_t *dict = NULL; - int32_t command = 0; - char *volname = NULL; - int32_t cmd = 0; - - if (!frame || !this || !data) { - ret = -1; - goto out; - } - - dict = data; - - ret = dict_get_str (dict, "volname", &volname); - if (ret) - goto out; - - ret = dict_get_int32 (dict, "command", &command); - if (ret) - goto out; - - if ((command != GF_OP_CMD_STATUS) && - (command != GF_OP_CMD_STOP)) { - - - ret = cli_to_glusterd (&req, frame, gf_cli_remove_brick_cbk, - (xdrproc_t) xdr_gf_cli_req, dict, - GLUSTER_CLI_REMOVE_BRICK, this, - cli_rpc_prog, NULL); - } else { - /* Need rebalance status to be sent :-) */ - if (command == GF_OP_CMD_STATUS) - cmd |= GF_DEFRAG_CMD_STATUS; - else - cmd |= GF_DEFRAG_CMD_STOP; - - ret = dict_set_int32 (dict, "rebalance-command", (int32_t) cmd); - if (ret) { - gf_log (this->name, GF_LOG_ERROR, - "Failed to set dict"); - goto out; - } +gf_cli_remove_brick(call_frame_t *frame, xlator_t *this, void *data) +{ + gf_cli_req req = {{ + 0, + }}; + ; + gf_cli_req status_req = {{ + 0, + }}; + ; + int ret = 0; + dict_t *dict = NULL; + int32_t command = 0; + char *volname = NULL; + int32_t cmd = 0; + + if (!frame || !this || !data) { + ret = -1; + goto out; + } + + dict = data; + + ret = dict_get_str(dict, "volname", &volname); + if (ret) + goto out; + + ret = dict_get_int32(dict, "command", &command); + if (ret) + goto out; + + if ((command != GF_OP_CMD_STATUS) && (command != GF_OP_CMD_STOP)) { + ret = cli_to_glusterd( + &req, frame, gf_cli_remove_brick_cbk, (xdrproc_t)xdr_gf_cli_req, + dict, GLUSTER_CLI_REMOVE_BRICK, this, cli_rpc_prog, NULL); + } else { + /* Need rebalance status to be sent :-) */ + if (command == GF_OP_CMD_STATUS) + cmd |= GF_DEFRAG_CMD_STATUS; + else + cmd |= GF_DEFRAG_CMD_STOP; - ret = cli_to_glusterd (&status_req, frame, - gf_cli3_remove_brick_status_cbk, - (xdrproc_t) xdr_gf_cli_req, dict, - GLUSTER_CLI_DEFRAG_VOLUME, this, - cli_rpc_prog, NULL); + ret = dict_set_int32(dict, "rebalance-command", (int32_t)cmd); + if (ret) { + gf_log(this->name, GF_LOG_ERROR, "Failed to set dict"); + goto out; + } - } + ret = cli_to_glusterd( + &status_req, frame, gf_cli3_remove_brick_status_cbk, + (xdrproc_t)xdr_gf_cli_req, dict, GLUSTER_CLI_DEFRAG_VOLUME, this, + cli_rpc_prog, NULL); + } out: - gf_log ("cli", GF_LOG_DEBUG, "Returning %d", ret); + gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret); - GF_FREE (req.dict.dict_val); + GF_FREE(req.dict.dict_val); - GF_FREE (status_req.dict.dict_val); + GF_FREE(status_req.dict.dict_val); - return ret; + return ret; } int32_t -gf_cli_reset_brick (call_frame_t *frame, xlator_t *this, void *data) -{ - gf_cli_req req = { {0,} }; - int ret = 0; - dict_t *dict = NULL; - char *dst_brick = NULL; - char *src_brick = NULL; - char *volname = NULL; - char *op = NULL; - - if (!frame || !this || !data) { - ret = -1; - goto out; - } - - dict = data; - - ret = dict_get_str (dict, "operation", &op); - if (ret) { - gf_log (this->name, GF_LOG_DEBUG, - "dict_get on operation failed"); - goto out; - } - - ret = dict_get_str (dict, "volname", &volname); +gf_cli_reset_brick(call_frame_t *frame, xlator_t *this, void *data) +{ + gf_cli_req req = {{ + 0, + }}; + int ret = 0; + dict_t *dict = NULL; + char *dst_brick = NULL; + char *src_brick = NULL; + char *volname = NULL; + char *op = NULL; + + if (!frame || !this || !data) { + ret = -1; + goto out; + } + + dict = data; + + ret = dict_get_str(dict, "operation", &op); + if (ret) { + gf_log(this->name, GF_LOG_DEBUG, "dict_get on operation failed"); + goto out; + } + + ret = dict_get_str(dict, "volname", &volname); + if (ret) { + gf_log(this->name, GF_LOG_DEBUG, "dict_get on volname failed"); + goto out; + } + + ret = dict_get_str(dict, "src-brick", &src_brick); + if (ret) { + gf_log(this->name, GF_LOG_DEBUG, "dict_get on src-brick failed"); + goto out; + } + + if (!strcmp(op, "GF_RESET_OP_COMMIT") || + !strcmp(op, "GF_RESET_OP_COMMIT_FORCE")) { + ret = dict_get_str(dict, "dst-brick", &dst_brick); if (ret) { - gf_log (this->name, GF_LOG_DEBUG, - "dict_get on volname failed"); - goto out; + gf_log(this->name, GF_LOG_DEBUG, "dict_get on dst-brick failed"); + goto out; } + } - ret = dict_get_str (dict, "src-brick", &src_brick); - if (ret) { - gf_log (this->name, GF_LOG_DEBUG, - "dict_get on src-brick failed"); - goto out; - } + gf_log(this->name, GF_LOG_DEBUG, "Received command reset-brick %s on %s.", + op, src_brick); - if (!strcmp (op, "GF_RESET_OP_COMMIT") || - !strcmp (op, "GF_RESET_OP_COMMIT_FORCE")) { - ret = dict_get_str (dict, "dst-brick", &dst_brick); - if (ret) { - gf_log (this->name, GF_LOG_DEBUG, - "dict_get on dst-brick failed"); - goto out; - } - } - - gf_log (this->name, GF_LOG_DEBUG, - "Received command reset-brick %s on %s.", op, src_brick); - - ret = cli_to_glusterd (&req, frame, gf_cli_reset_brick_cbk, - (xdrproc_t) xdr_gf_cli_req, dict, - GLUSTER_CLI_RESET_BRICK, this, cli_rpc_prog, - NULL); + ret = cli_to_glusterd(&req, frame, gf_cli_reset_brick_cbk, + (xdrproc_t)xdr_gf_cli_req, dict, + GLUSTER_CLI_RESET_BRICK, this, cli_rpc_prog, NULL); out: - gf_log ("cli", GF_LOG_DEBUG, "Returning %d", ret); + gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret); - GF_FREE (req.dict.dict_val); + GF_FREE(req.dict.dict_val); - return ret; + return ret; } int32_t -gf_cli_replace_brick (call_frame_t *frame, xlator_t *this, - void *data) -{ - gf_cli_req req = {{0,}}; - int ret = 0; - dict_t *dict = NULL; - char *src_brick = NULL; - char *dst_brick = NULL; - char *volname = NULL; - int32_t op = 0; - - if (!frame || !this || !data) { - ret = -1; - goto out; - } +gf_cli_replace_brick(call_frame_t *frame, xlator_t *this, void *data) +{ + gf_cli_req req = {{ + 0, + }}; + int ret = 0; + dict_t *dict = NULL; + char *src_brick = NULL; + char *dst_brick = NULL; + char *volname = NULL; + int32_t op = 0; - dict = data; + if (!frame || !this || !data) { + ret = -1; + goto out; + } - ret = dict_get_int32 (dict, "operation", &op); - if (ret) { - gf_log (this->name, GF_LOG_DEBUG, - "dict_get on operation failed"); - goto out; - } - ret = dict_get_str (dict, "volname", &volname); - if (ret) { - gf_log (this->name, GF_LOG_DEBUG, - "dict_get on volname failed"); - goto out; - } + dict = data; - ret = dict_get_str (dict, "src-brick", &src_brick); - if (ret) { - gf_log (this->name, GF_LOG_DEBUG, - "dict_get on src-brick failed"); - goto out; - } + ret = dict_get_int32(dict, "operation", &op); + if (ret) { + gf_log(this->name, GF_LOG_DEBUG, "dict_get on operation failed"); + goto out; + } + ret = dict_get_str(dict, "volname", &volname); + if (ret) { + gf_log(this->name, GF_LOG_DEBUG, "dict_get on volname failed"); + goto out; + } - ret = dict_get_str (dict, "dst-brick", &dst_brick); - if (ret) { - gf_log (this->name, GF_LOG_DEBUG, - "dict_get on dst-brick failed"); - goto out; - } + ret = dict_get_str(dict, "src-brick", &src_brick); + if (ret) { + gf_log(this->name, GF_LOG_DEBUG, "dict_get on src-brick failed"); + goto out; + } + + ret = dict_get_str(dict, "dst-brick", &dst_brick); + if (ret) { + gf_log(this->name, GF_LOG_DEBUG, "dict_get on dst-brick failed"); + goto out; + } - gf_log (this->name, GF_LOG_DEBUG, - "Received command replace-brick %s with " - "%s with operation=%d", src_brick, - dst_brick, op); + gf_log(this->name, GF_LOG_DEBUG, + "Received command replace-brick %s with " + "%s with operation=%d", + src_brick, dst_brick, op); - ret = cli_to_glusterd (&req, frame, gf_cli_replace_brick_cbk, - (xdrproc_t) xdr_gf_cli_req, dict, - GLUSTER_CLI_REPLACE_BRICK, this, cli_rpc_prog, - NULL); + ret = cli_to_glusterd(&req, frame, gf_cli_replace_brick_cbk, + (xdrproc_t)xdr_gf_cli_req, dict, + GLUSTER_CLI_REPLACE_BRICK, this, cli_rpc_prog, NULL); out: - gf_log ("cli", GF_LOG_DEBUG, "Returning %d", ret); + gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret); - GF_FREE (req.dict.dict_val); + GF_FREE(req.dict.dict_val); - return ret; + return ret; } - int32_t -gf_cli_log_rotate (call_frame_t *frame, xlator_t *this, - void *data) +gf_cli_log_rotate(call_frame_t *frame, xlator_t *this, void *data) { - gf_cli_req req = {{0,}}; - int ret = 0; - dict_t *dict = NULL; + gf_cli_req req = {{ + 0, + }}; + int ret = 0; + dict_t *dict = NULL; - if (!frame || !this || !data) { - ret = -1; - goto out; - } + if (!frame || !this || !data) { + ret = -1; + goto out; + } - dict = data; + dict = data; - ret = cli_to_glusterd (&req, frame, gf_cli_log_rotate_cbk, - (xdrproc_t) xdr_gf_cli_req, dict, - GLUSTER_CLI_LOG_ROTATE, this, cli_rpc_prog, - NULL); + ret = cli_to_glusterd(&req, frame, gf_cli_log_rotate_cbk, + (xdrproc_t)xdr_gf_cli_req, dict, + GLUSTER_CLI_LOG_ROTATE, this, cli_rpc_prog, NULL); out: - gf_log ("cli", GF_LOG_DEBUG, "Returning %d", ret); + gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret); - GF_FREE (req.dict.dict_val); - return ret; + GF_FREE(req.dict.dict_val); + return ret; } int32_t -gf_cli_sync_volume (call_frame_t *frame, xlator_t *this, - void *data) +gf_cli_sync_volume(call_frame_t *frame, xlator_t *this, void *data) { - int ret = 0; - gf_cli_req req = {{0,}}; - dict_t *dict = NULL; + int ret = 0; + gf_cli_req req = {{ + 0, + }}; + dict_t *dict = NULL; - if (!frame || !this || !data) { - ret = -1; - goto out; - } + if (!frame || !this || !data) { + ret = -1; + goto out; + } - dict = data; + dict = data; - ret = cli_to_glusterd (&req, frame, gf_cli_sync_volume_cbk, - (xdrproc_t) xdr_gf_cli_req, dict, - GLUSTER_CLI_SYNC_VOLUME, this, cli_rpc_prog, - NULL); + ret = cli_to_glusterd(&req, frame, gf_cli_sync_volume_cbk, + (xdrproc_t)xdr_gf_cli_req, dict, + GLUSTER_CLI_SYNC_VOLUME, this, cli_rpc_prog, NULL); out: - gf_log ("cli", GF_LOG_DEBUG, "Returning %d", ret); - GF_FREE (req.dict.dict_val); + gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret); + GF_FREE(req.dict.dict_val); - return ret; + return ret; } int32_t -gf_cli_getspec (call_frame_t *frame, xlator_t *this, - void *data) -{ - gf_getspec_req req = {0,}; - int ret = 0; - dict_t *dict = NULL; - dict_t *op_dict = NULL; - - if (!frame || !this || !data) { - ret = -1; - goto out; - } - - dict = data; - - ret = dict_get_str (dict, "volid", &req.key); - if (ret) - goto out; - - op_dict = dict_new (); - if (!op_dict) { - ret = -1; - goto out; - } - - // Set the supported min and max op-versions, so glusterd can make a - // decision - ret = dict_set_int32 (op_dict, "min-op-version", GD_OP_VERSION_MIN); - if (ret) { - gf_log (THIS->name, GF_LOG_ERROR, "Failed to set min-op-version" - " in request dict"); - goto out; - } - - ret = dict_set_int32 (op_dict, "max-op-version", GD_OP_VERSION_MAX); - if (ret) { - gf_log (THIS->name, GF_LOG_ERROR, "Failed to set max-op-version" - " in request dict"); - goto out; - } - - ret = dict_allocate_and_serialize (op_dict, &req.xdata.xdata_val, - &req.xdata.xdata_len); - if (ret < 0) { - gf_log (THIS->name, GF_LOG_ERROR, - "Failed to serialize dictionary"); - goto out; - } - - ret = cli_cmd_submit (NULL, &req, frame, &cli_handshake_prog, - GF_HNDSK_GETSPEC, NULL, - this, gf_cli_getspec_cbk, - (xdrproc_t) xdr_gf_getspec_req); - -out: - if (op_dict) { - dict_unref(op_dict); - } - gf_log ("cli", GF_LOG_DEBUG, "Returning %d", ret); - - return ret; +gf_cli_getspec(call_frame_t *frame, xlator_t *this, void *data) +{ + gf_getspec_req req = { + 0, + }; + int ret = 0; + dict_t *dict = NULL; + dict_t *op_dict = NULL; + + if (!frame || !this || !data) { + ret = -1; + goto out; + } + + dict = data; + + ret = dict_get_str(dict, "volid", &req.key); + if (ret) + goto out; + + op_dict = dict_new(); + if (!op_dict) { + ret = -1; + goto out; + } + + // Set the supported min and max op-versions, so glusterd can make a + // decision + ret = dict_set_int32(op_dict, "min-op-version", GD_OP_VERSION_MIN); + if (ret) { + gf_log(THIS->name, GF_LOG_ERROR, + "Failed to set min-op-version" + " in request dict"); + goto out; + } + + ret = dict_set_int32(op_dict, "max-op-version", GD_OP_VERSION_MAX); + if (ret) { + gf_log(THIS->name, GF_LOG_ERROR, + "Failed to set max-op-version" + " in request dict"); + goto out; + } + + ret = dict_allocate_and_serialize(op_dict, &req.xdata.xdata_val, + &req.xdata.xdata_len); + if (ret < 0) { + gf_log(THIS->name, GF_LOG_ERROR, "Failed to serialize dictionary"); + goto out; + } + + ret = cli_cmd_submit(NULL, &req, frame, &cli_handshake_prog, + GF_HNDSK_GETSPEC, NULL, this, gf_cli_getspec_cbk, + (xdrproc_t)xdr_gf_getspec_req); + +out: + if (op_dict) { + dict_unref(op_dict); + } + gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret); + + return ret; } int32_t -gf_cli_quota (call_frame_t *frame, xlator_t *this, - void *data) +gf_cli_quota(call_frame_t *frame, xlator_t *this, void *data) { - gf_cli_req req = {{0,}}; - int ret = 0; - dict_t *dict = NULL; + gf_cli_req req = {{ + 0, + }}; + int ret = 0; + dict_t *dict = NULL; - if (!frame || !this || !data) { - ret = -1; - goto out; - } + if (!frame || !this || !data) { + ret = -1; + goto out; + } - dict = data; + dict = data; - ret = cli_to_glusterd (&req, frame, gf_cli_quota_cbk, - (xdrproc_t) xdr_gf_cli_req, dict, - GLUSTER_CLI_QUOTA, this, cli_rpc_prog, NULL); + ret = cli_to_glusterd(&req, frame, gf_cli_quota_cbk, + (xdrproc_t)xdr_gf_cli_req, dict, GLUSTER_CLI_QUOTA, + this, cli_rpc_prog, NULL); out: - GF_FREE (req.dict.dict_val); + GF_FREE(req.dict.dict_val); - return ret; + return ret; } int32_t -gf_cli_pmap_b2p (call_frame_t *frame, xlator_t *this, void *data) +gf_cli_pmap_b2p(call_frame_t *frame, xlator_t *this, void *data) { - pmap_port_by_brick_req req = {0,}; - int ret = 0; - dict_t *dict = NULL; + pmap_port_by_brick_req req = { + 0, + }; + int ret = 0; + dict_t *dict = NULL; - if (!frame || !this || !data) { - ret = -1; - goto out; - } + if (!frame || !this || !data) { + ret = -1; + goto out; + } - dict = data; + dict = data; - ret = dict_get_str (dict, "brick", &req.brick); - if (ret) - goto out; + ret = dict_get_str(dict, "brick", &req.brick); + if (ret) + goto out; - ret = cli_cmd_submit (NULL, &req, frame, &cli_pmap_prog, - GF_PMAP_PORTBYBRICK, NULL, - this, gf_cli_pmap_b2p_cbk, - (xdrproc_t) xdr_pmap_port_by_brick_req); + ret = cli_cmd_submit(NULL, &req, frame, &cli_pmap_prog, GF_PMAP_PORTBYBRICK, + NULL, this, gf_cli_pmap_b2p_cbk, + (xdrproc_t)xdr_pmap_port_by_brick_req); out: - gf_log ("cli", GF_LOG_DEBUG, "Returning %d", ret); + gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret); - return ret; + return ret; } static int -gf_cli_fsm_log_cbk (struct rpc_req *req, struct iovec *iov, - int count, void *myframe) -{ - gf1_cli_fsm_log_rsp rsp = {0,}; - int ret = -1; - dict_t *dict = NULL; - int tr_count = 0; - char key[256] = {0}; - int i = 0; - char *old_state = NULL; - char *new_state = NULL; - char *event = NULL; - char *time = NULL; - - GF_ASSERT (myframe); - - if (-1 == req->rpc_status) { - goto out; - } - - ret = xdr_to_generic (*iov, &rsp, (xdrproc_t)xdr_gf1_cli_fsm_log_rsp); - if (ret < 0) { - gf_log (((call_frame_t *) myframe)->this->name, GF_LOG_ERROR, - "Failed to decode xdr response"); - goto out; - } - - if (rsp.op_ret) { - if (strcmp (rsp.op_errstr, "")) - cli_err ("%s", rsp.op_errstr); - cli_err ("fsm log unsuccessful"); - ret = rsp.op_ret; - goto out; - } - - dict = dict_new (); - if (!dict) { - ret = -1; - goto out; - } - - ret = dict_unserialize (rsp.fsm_log.fsm_log_val, - rsp.fsm_log.fsm_log_len, - &dict); - - if (ret) { - cli_err ("bad response"); - goto out; - } - - ret = dict_get_int32 (dict, "count", &tr_count); - if (tr_count) - cli_out("number of transitions: %d", tr_count); - else - cli_err("No transitions"); - for (i = 0; i < tr_count; i++) { - snprintf (key, sizeof (key), "log%d-old-state", i); - ret = dict_get_str (dict, key, &old_state); - if (ret) - goto out; +gf_cli_fsm_log_cbk(struct rpc_req *req, struct iovec *iov, int count, + void *myframe) +{ + gf1_cli_fsm_log_rsp rsp = { + 0, + }; + int ret = -1; + dict_t *dict = NULL; + int tr_count = 0; + char key[256] = {0}; + int i = 0; + char *old_state = NULL; + char *new_state = NULL; + char *event = NULL; + char *time = NULL; + + GF_ASSERT(myframe); + + if (-1 == req->rpc_status) { + goto out; + } + + ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf1_cli_fsm_log_rsp); + if (ret < 0) { + gf_log(((call_frame_t *)myframe)->this->name, GF_LOG_ERROR, + "Failed to decode xdr response"); + goto out; + } + + if (rsp.op_ret) { + if (strcmp(rsp.op_errstr, "")) + cli_err("%s", rsp.op_errstr); + cli_err("fsm log unsuccessful"); + ret = rsp.op_ret; + goto out; + } + + dict = dict_new(); + if (!dict) { + ret = -1; + goto out; + } + + ret = dict_unserialize(rsp.fsm_log.fsm_log_val, rsp.fsm_log.fsm_log_len, + &dict); + + if (ret) { + cli_err("bad response"); + goto out; + } + + ret = dict_get_int32(dict, "count", &tr_count); + if (tr_count) + cli_out("number of transitions: %d", tr_count); + else + cli_err("No transitions"); + for (i = 0; i < tr_count; i++) { + snprintf(key, sizeof(key), "log%d-old-state", i); + ret = dict_get_str(dict, key, &old_state); + if (ret) + goto out; - snprintf (key, sizeof (key), "log%d-event", i); - ret = dict_get_str (dict, key, &event); - if (ret) - goto out; + snprintf(key, sizeof(key), "log%d-event", i); + ret = dict_get_str(dict, key, &event); + if (ret) + goto out; - snprintf (key, sizeof (key), "log%d-new-state", i); - ret = dict_get_str (dict, key, &new_state); - if (ret) - goto out; + snprintf(key, sizeof(key), "log%d-new-state", i); + ret = dict_get_str(dict, key, &new_state); + if (ret) + goto out; - snprintf (key, sizeof (key), "log%d-time", i); - ret = dict_get_str (dict, key, &time); - if (ret) - goto out; - cli_out ("Old State: [%s]\n" - "New State: [%s]\n" - "Event : [%s]\n" - "timestamp: [%s]\n", old_state, new_state, event, time); - } + snprintf(key, sizeof(key), "log%d-time", i); + ret = dict_get_str(dict, key, &time); + if (ret) + goto out; + cli_out( + "Old State: [%s]\n" + "New State: [%s]\n" + "Event : [%s]\n" + "timestamp: [%s]\n", + old_state, new_state, event, time); + } - ret = rsp.op_ret; + ret = rsp.op_ret; out: - cli_cmd_broadcast_response (ret); - return ret; + cli_cmd_broadcast_response(ret); + return ret; } int32_t -gf_cli_fsm_log (call_frame_t *frame, xlator_t *this, void *data) +gf_cli_fsm_log(call_frame_t *frame, xlator_t *this, void *data) { - int ret = -1; - gf1_cli_fsm_log_req req = {0,}; + int ret = -1; + gf1_cli_fsm_log_req req = { + 0, + }; - GF_ASSERT (frame); - GF_ASSERT (this); - GF_ASSERT (data); + GF_ASSERT(frame); + GF_ASSERT(this); + GF_ASSERT(data); - if (!frame || !this || !data) - goto out; - req.name = data; - ret = cli_cmd_submit (NULL, &req, frame, cli_rpc_prog, - GLUSTER_CLI_FSM_LOG, NULL, - this, gf_cli_fsm_log_cbk, - (xdrproc_t) xdr_gf1_cli_fsm_log_req); + if (!frame || !this || !data) + goto out; + req.name = data; + ret = cli_cmd_submit(NULL, &req, frame, cli_rpc_prog, GLUSTER_CLI_FSM_LOG, + NULL, this, gf_cli_fsm_log_cbk, + (xdrproc_t)xdr_gf1_cli_fsm_log_req); out: - gf_log ("cli", GF_LOG_DEBUG, "Returning %d", ret); + gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret); - return ret; + return ret; } int -gf_cli_gsync_config_command (dict_t *dict) -{ - runner_t runner = {0,}; - char *subop = NULL; - char *gwd = NULL; - char *slave = NULL; - char *confpath = NULL; - char *master = NULL; - char *op_name = NULL; - int ret = -1; - char conf_path[PATH_MAX] = ""; - - if (dict_get_str (dict, "subop", &subop) != 0) - return -1; - - if (strcmp (subop, "get") != 0 && strcmp (subop, "get-all") != 0) { - cli_out (GEOREP" config updated successfully"); - return 0; - } - - if (dict_get_str (dict, "glusterd_workdir", &gwd) != 0 || - dict_get_str (dict, "slave", &slave) != 0) - return -1; - - if (dict_get_str (dict, "master", &master) != 0) - master = NULL; - if (dict_get_str (dict, "op_name", &op_name) != 0) - op_name = NULL; - - ret = dict_get_str (dict, "conf_path", &confpath); - if (ret || !confpath) { - ret = snprintf (conf_path, sizeof(conf_path) - 1, - "%s/"GEOREP"/gsyncd_template.conf", gwd); - conf_path[ret] = '\0'; - confpath = conf_path; - } - - runinit (&runner); - runner_add_args (&runner, GSYNCD_PREFIX"/gsyncd", "-c", NULL); - runner_argprintf (&runner, "%s", confpath); - runner_argprintf (&runner, "--iprefix=%s", DATADIR); - if (master) - runner_argprintf (&runner, ":%s", master); - runner_add_arg (&runner, slave); - runner_argprintf (&runner, "--config-%s", subop); - if (op_name) - runner_add_arg (&runner, op_name); - - return runner_run (&runner); +gf_cli_gsync_config_command(dict_t *dict) +{ + runner_t runner = { + 0, + }; + char *subop = NULL; + char *gwd = NULL; + char *slave = NULL; + char *confpath = NULL; + char *master = NULL; + char *op_name = NULL; + int ret = -1; + char conf_path[PATH_MAX] = ""; + + if (dict_get_str(dict, "subop", &subop) != 0) + return -1; + + if (strcmp(subop, "get") != 0 && strcmp(subop, "get-all") != 0) { + cli_out(GEOREP " config updated successfully"); + return 0; + } + + if (dict_get_str(dict, "glusterd_workdir", &gwd) != 0 || + dict_get_str(dict, "slave", &slave) != 0) + return -1; + + if (dict_get_str(dict, "master", &master) != 0) + master = NULL; + if (dict_get_str(dict, "op_name", &op_name) != 0) + op_name = NULL; + + ret = dict_get_str(dict, "conf_path", &confpath); + if (ret || !confpath) { + ret = snprintf(conf_path, sizeof(conf_path) - 1, + "%s/" GEOREP "/gsyncd_template.conf", gwd); + conf_path[ret] = '\0'; + confpath = conf_path; + } + + runinit(&runner); + runner_add_args(&runner, GSYNCD_PREFIX "/gsyncd", "-c", NULL); + runner_argprintf(&runner, "%s", confpath); + runner_argprintf(&runner, "--iprefix=%s", DATADIR); + if (master) + runner_argprintf(&runner, ":%s", master); + runner_add_arg(&runner, slave); + runner_argprintf(&runner, "--config-%s", subop); + if (op_name) + runner_add_arg(&runner, op_name); + + return runner_run(&runner); } - int -gf_cli_print_status (char **title_values, - gf_gsync_status_t **sts_vals, - int *spacing, int gsync_count, - int number_of_fields, int is_detail) -{ - int i = 0; - int j = 0; - int ret = 0; - int status_fields = 8; /* Indexed at 0 */ - int total_spacing = 0; - char **output_values = NULL; - char *tmp = NULL; - char *hyphens = NULL; - - /* calculating spacing for hyphens */ - for (i = 0; i < number_of_fields; i++) { - /* Suppressing detail output for status */ - if ((!is_detail) && (i > status_fields)) { - /* Suppressing detailed output for - * status */ - continue; - } - spacing[i] += 3; /* Adding extra space to - distinguish between fields */ - total_spacing += spacing[i]; - } - total_spacing += 4; /* For the spacing between the fields */ - - /* char pointers for each field */ - output_values = GF_CALLOC (number_of_fields, sizeof (char *), - gf_common_mt_char); - if (!output_values) { - ret = -1; - goto out; - } - for (i = 0; i < number_of_fields; i++) { - output_values[i] = GF_CALLOC (spacing[i] + 1, sizeof (char), - gf_common_mt_char); - if (!output_values[i]) { - ret = -1; - goto out; - } - } - - hyphens = GF_CALLOC (total_spacing + 1, sizeof (char), - gf_common_mt_char); - if (!hyphens) { +gf_cli_print_status(char **title_values, gf_gsync_status_t **sts_vals, + int *spacing, int gsync_count, int number_of_fields, + int is_detail) +{ + int i = 0; + int j = 0; + int ret = 0; + int status_fields = 8; /* Indexed at 0 */ + int total_spacing = 0; + char **output_values = NULL; + char *tmp = NULL; + char *hyphens = NULL; + + /* calculating spacing for hyphens */ + for (i = 0; i < number_of_fields; i++) { + /* Suppressing detail output for status */ + if ((!is_detail) && (i > status_fields)) { + /* Suppressing detailed output for + * status */ + continue; + } + spacing[i] += 3; /* Adding extra space to + distinguish between fields */ + total_spacing += spacing[i]; + } + total_spacing += 4; /* For the spacing between the fields */ + + /* char pointers for each field */ + output_values = GF_CALLOC(number_of_fields, sizeof(char *), + gf_common_mt_char); + if (!output_values) { + ret = -1; + goto out; + } + for (i = 0; i < number_of_fields; i++) { + output_values[i] = GF_CALLOC(spacing[i] + 1, sizeof(char), + gf_common_mt_char); + if (!output_values[i]) { + ret = -1; + goto out; + } + } + + hyphens = GF_CALLOC(total_spacing + 1, sizeof(char), gf_common_mt_char); + if (!hyphens) { + ret = -1; + goto out; + } + + cli_out(" "); + + /* setting the title "NODE", "MASTER", etc. from title_values[] + and printing the same */ + for (j = 0; j < number_of_fields; j++) { + if ((!is_detail) && (j > status_fields)) { + /* Suppressing detailed output for + * status */ + output_values[j][0] = '\0'; + continue; + } + memset(output_values[j], ' ', spacing[j]); + memcpy(output_values[j], title_values[j], strlen(title_values[j])); + output_values[j][spacing[j]] = '\0'; + } + cli_out("%s %s %s %s %s %s %s %s %s %s %s %s %s %s %s %s", output_values[0], + output_values[1], output_values[2], output_values[3], + output_values[4], output_values[5], output_values[6], + output_values[7], output_values[8], output_values[9], + output_values[10], output_values[11], output_values[12], + output_values[13], output_values[14], output_values[15]); + + /* setting and printing the hyphens */ + memset(hyphens, '-', total_spacing); + hyphens[total_spacing] = '\0'; + cli_out("%s", hyphens); + + for (i = 0; i < gsync_count; i++) { + for (j = 0; j < number_of_fields; j++) { + if ((!is_detail) && (j > status_fields)) { + /* Suppressing detailed output for + * status */ + output_values[j][0] = '\0'; + continue; + } + tmp = get_struct_variable(j, sts_vals[i]); + if (!tmp) { + gf_log("", GF_LOG_ERROR, "struct member empty."); ret = -1; goto out; + } + memset(output_values[j], ' ', spacing[j]); + memcpy(output_values[j], tmp, strlen(tmp)); + output_values[j][spacing[j]] = '\0'; } - cli_out (" "); - - /* setting the title "NODE", "MASTER", etc. from title_values[] - and printing the same */ - for (j = 0; j < number_of_fields; j++) { - if ((!is_detail) && (j > status_fields)) { - /* Suppressing detailed output for - * status */ - output_values[j][0] = '\0'; - continue; - } - memset (output_values[j], ' ', spacing[j]); - memcpy (output_values[j], title_values[j], - strlen(title_values[j])); - output_values[j][spacing[j]] = '\0'; - } - cli_out ("%s %s %s %s %s %s %s %s %s %s %s %s %s %s %s %s", - output_values[0], output_values[1], - output_values[2], output_values[3], - output_values[4], output_values[5], - output_values[6], output_values[7], - output_values[8], output_values[9], - output_values[10], output_values[11], - output_values[12], output_values[13], - output_values[14], output_values[15]); - - /* setting and printing the hyphens */ - memset (hyphens, '-', total_spacing); - hyphens[total_spacing] = '\0'; - cli_out ("%s", hyphens); - - for (i = 0; i < gsync_count; i++) { - for (j = 0; j < number_of_fields; j++) { - if ((!is_detail) && (j > status_fields)) { - /* Suppressing detailed output for - * status */ - output_values[j][0] = '\0'; - continue; - } - tmp = get_struct_variable(j, sts_vals[i]); - if (!tmp) { - gf_log ("", GF_LOG_ERROR, - "struct member empty."); - ret = -1; - goto out; - } - memset (output_values[j], ' ', spacing[j]); - memcpy (output_values[j], tmp, strlen (tmp)); - output_values[j][spacing[j]] = '\0'; - } - - cli_out ("%s %s %s %s %s %s %s %s %s %s %s %s %s %s %s %s", - output_values[0], output_values[1], - output_values[2], output_values[3], - output_values[4], output_values[5], - output_values[6], output_values[7], - output_values[8], output_values[9], - output_values[10], output_values[11], - output_values[12], output_values[13], - output_values[14], output_values[15]); - } + cli_out("%s %s %s %s %s %s %s %s %s %s %s %s %s %s %s %s", + output_values[0], output_values[1], output_values[2], + output_values[3], output_values[4], output_values[5], + output_values[6], output_values[7], output_values[8], + output_values[9], output_values[10], output_values[11], + output_values[12], output_values[13], output_values[14], + output_values[15]); + } out: - if (output_values) { - for (i = 0; i < number_of_fields; i++) { - if (output_values[i]) - GF_FREE (output_values[i]); - } - GF_FREE (output_values); + if (output_values) { + for (i = 0; i < number_of_fields; i++) { + if (output_values[i]) + GF_FREE(output_values[i]); } + GF_FREE(output_values); + } - if (hyphens) - GF_FREE (hyphens); + if (hyphens) + GF_FREE(hyphens); - return ret; + return ret; } int -gf_gsync_status_t_comparator (const void *p, const void *q) +gf_gsync_status_t_comparator(const void *p, const void *q) { - char *slavekey1 = NULL; - char *slavekey2 = NULL; - - slavekey1 = get_struct_variable (20, (*(gf_gsync_status_t **)p)); - slavekey2 = get_struct_variable (20, (*(gf_gsync_status_t **)q)); - if (!slavekey1 || !slavekey2) { - gf_log ("cli", GF_LOG_ERROR, - "struct member empty."); - return 0; - } - - return strcmp (slavekey1, slavekey2); -} - -int -gf_cli_read_status_data (dict_t *dict, - gf_gsync_status_t **sts_vals, - int *spacing, int gsync_count, - int number_of_fields) -{ - char *tmp = NULL; - char sts_val_name[PATH_MAX] = ""; - int ret = 0; - int i = 0; - int j = 0; - - /* Storing per node status info in each object */ - for (i = 0; i < gsync_count; i++) { - snprintf (sts_val_name, sizeof(sts_val_name), "status_value%d", i); - - /* Fetching the values from dict, and calculating - the max length for each field */ - ret = dict_get_bin (dict, sts_val_name, (void **)&(sts_vals[i])); - if (ret) - goto out; - - for (j = 0; j < number_of_fields; j++) { - tmp = get_struct_variable(j, sts_vals[i]); - if (!tmp) { - gf_log ("", GF_LOG_ERROR, - "struct member empty."); - ret = -1; - goto out; - } - if (strlen (tmp) > spacing[j]) - spacing[j] = strlen (tmp); - } - } + char *slavekey1 = NULL; + char *slavekey2 = NULL; - /* Sort based on Session Slave */ - qsort(sts_vals, gsync_count, - sizeof(gf_gsync_status_t *), - gf_gsync_status_t_comparator); + slavekey1 = get_struct_variable(20, (*(gf_gsync_status_t **)p)); + slavekey2 = get_struct_variable(20, (*(gf_gsync_status_t **)q)); + if (!slavekey1 || !slavekey2) { + gf_log("cli", GF_LOG_ERROR, "struct member empty."); + return 0; + } -out: - return ret; + return strcmp(slavekey1, slavekey2); } int -gf_cli_gsync_status_output (dict_t *dict, gf_boolean_t is_detail) -{ - int gsync_count = 0; - int i = 0; - int ret = 0; - int spacing[16] = {0}; - int num_of_fields = 16; - char errmsg[1024] = ""; - char *master = NULL; - char *slave = NULL; - char *title_values[] = {"MASTER NODE", "MASTER VOL", - "MASTER BRICK", "SLAVE USER", - "SLAVE", "SLAVE NODE", - "STATUS", "CRAWL STATUS", - "LAST_SYNCED", "ENTRY", - "DATA", "META", "FAILURES", - "CHECKPOINT TIME", - "CHECKPOINT COMPLETED", - "CHECKPOINT COMPLETION TIME"}; - gf_gsync_status_t **sts_vals = NULL; - - /* Checks if any session is active or not */ - ret = dict_get_int32 (dict, "gsync-count", &gsync_count); - if (ret) { - ret = dict_get_str (dict, "master", &master); - - ret = dict_get_str (dict, "slave", &slave); - - if (master) { - if (slave) - snprintf (errmsg, sizeof(errmsg), "No active " - "geo-replication sessions between %s" - " and %s", master, slave); - else - snprintf (errmsg, sizeof(errmsg), "No active " - "geo-replication sessions for %s", - master); - } else - snprintf (errmsg, sizeof(errmsg), "No active " - "geo-replication sessions"); - - gf_log ("cli", GF_LOG_INFO, "%s", errmsg); - cli_out ("%s", errmsg); - ret = -1; - goto out; - } - - for (i = 0; i < num_of_fields; i++) - spacing[i] = strlen(title_values[i]); +gf_cli_read_status_data(dict_t *dict, gf_gsync_status_t **sts_vals, + int *spacing, int gsync_count, int number_of_fields) +{ + char *tmp = NULL; + char sts_val_name[PATH_MAX] = ""; + int ret = 0; + int i = 0; + int j = 0; + + /* Storing per node status info in each object */ + for (i = 0; i < gsync_count; i++) { + snprintf(sts_val_name, sizeof(sts_val_name), "status_value%d", i); + + /* Fetching the values from dict, and calculating + the max length for each field */ + ret = dict_get_bin(dict, sts_val_name, (void **)&(sts_vals[i])); + if (ret) + goto out; - /* gsync_count = number of nodes reporting output. - each sts_val object will store output of each - node */ - sts_vals = GF_MALLOC (gsync_count * sizeof (gf_gsync_status_t *), - gf_common_mt_char); - if (!sts_vals) { + for (j = 0; j < number_of_fields; j++) { + tmp = get_struct_variable(j, sts_vals[i]); + if (!tmp) { + gf_log("", GF_LOG_ERROR, "struct member empty."); ret = -1; goto out; + } + if (strlen(tmp) > spacing[j]) + spacing[j] = strlen(tmp); } - for (i = 0; i < gsync_count; i++) { - sts_vals[i] = GF_CALLOC (1, sizeof (gf_gsync_status_t), - gf_common_mt_char); - if (!sts_vals[i]) { - ret = -1; - goto out; - } - } + } - ret = gf_cli_read_status_data (dict, sts_vals, spacing, - gsync_count, num_of_fields); - if (ret) { - gf_log ("", GF_LOG_ERROR, "Unable to read status data"); - goto out; - } - - ret = gf_cli_print_status (title_values, sts_vals, spacing, gsync_count, - num_of_fields, is_detail); - if (ret) { - gf_log ("", GF_LOG_ERROR, "Unable to print status output"); - goto out; - } + /* Sort based on Session Slave */ + qsort(sts_vals, gsync_count, sizeof(gf_gsync_status_t *), + gf_gsync_status_t_comparator); out: - if (sts_vals) - GF_FREE (sts_vals); + return ret; +} - return ret; +int +gf_cli_gsync_status_output(dict_t *dict, gf_boolean_t is_detail) +{ + int gsync_count = 0; + int i = 0; + int ret = 0; + int spacing[16] = {0}; + int num_of_fields = 16; + char errmsg[1024] = ""; + char *master = NULL; + char *slave = NULL; + char *title_values[] = {"MASTER NODE", + "MASTER VOL", + "MASTER BRICK", + "SLAVE USER", + "SLAVE", + "SLAVE NODE", + "STATUS", + "CRAWL STATUS", + "LAST_SYNCED", + "ENTRY", + "DATA", + "META", + "FAILURES", + "CHECKPOINT TIME", + "CHECKPOINT COMPLETED", + "CHECKPOINT COMPLETION TIME"}; + gf_gsync_status_t **sts_vals = NULL; + + /* Checks if any session is active or not */ + ret = dict_get_int32(dict, "gsync-count", &gsync_count); + if (ret) { + ret = dict_get_str(dict, "master", &master); + + ret = dict_get_str(dict, "slave", &slave); + + if (master) { + if (slave) + snprintf(errmsg, sizeof(errmsg), + "No active " + "geo-replication sessions between %s" + " and %s", + master, slave); + else + snprintf(errmsg, sizeof(errmsg), + "No active " + "geo-replication sessions for %s", + master); + } else + snprintf(errmsg, sizeof(errmsg), + "No active " + "geo-replication sessions"); + + gf_log("cli", GF_LOG_INFO, "%s", errmsg); + cli_out("%s", errmsg); + ret = -1; + goto out; + } + + for (i = 0; i < num_of_fields; i++) + spacing[i] = strlen(title_values[i]); + + /* gsync_count = number of nodes reporting output. + each sts_val object will store output of each + node */ + sts_vals = GF_MALLOC(gsync_count * sizeof(gf_gsync_status_t *), + gf_common_mt_char); + if (!sts_vals) { + ret = -1; + goto out; + } + for (i = 0; i < gsync_count; i++) { + sts_vals[i] = GF_CALLOC(1, sizeof(gf_gsync_status_t), + gf_common_mt_char); + if (!sts_vals[i]) { + ret = -1; + goto out; + } + } + + ret = gf_cli_read_status_data(dict, sts_vals, spacing, gsync_count, + num_of_fields); + if (ret) { + gf_log("", GF_LOG_ERROR, "Unable to read status data"); + goto out; + } + + ret = gf_cli_print_status(title_values, sts_vals, spacing, gsync_count, + num_of_fields, is_detail); + if (ret) { + gf_log("", GF_LOG_ERROR, "Unable to print status output"); + goto out; + } + +out: + if (sts_vals) + GF_FREE(sts_vals); + + return ret; } static int32_t -write_contents_to_common_pem_file (dict_t *dict, int output_count) -{ - char *workdir = NULL; - char common_pem_file[PATH_MAX] = ""; - char *output = NULL; - char output_name[PATH_MAX] = ""; - int bytes_written = 0; - int fd = -1; - int ret = -1; - int i = -1; - - ret = dict_get_str (dict, "glusterd_workdir", &workdir); - if (ret || !workdir) { - gf_log ("", GF_LOG_ERROR, "Unable to fetch workdir"); +write_contents_to_common_pem_file(dict_t *dict, int output_count) +{ + char *workdir = NULL; + char common_pem_file[PATH_MAX] = ""; + char *output = NULL; + char output_name[PATH_MAX] = ""; + int bytes_written = 0; + int fd = -1; + int ret = -1; + int i = -1; + + ret = dict_get_str(dict, "glusterd_workdir", &workdir); + if (ret || !workdir) { + gf_log("", GF_LOG_ERROR, "Unable to fetch workdir"); + ret = -1; + goto out; + } + + snprintf(common_pem_file, sizeof(common_pem_file), + "%s/geo-replication/common_secret.pem.pub", workdir); + + sys_unlink(common_pem_file); + + fd = open(common_pem_file, O_WRONLY | O_CREAT, 0600); + if (fd == -1) { + gf_log("", GF_LOG_ERROR, + "Failed to open %s" + " Error : %s", + common_pem_file, strerror(errno)); + ret = -1; + goto out; + } + + for (i = 1; i <= output_count; i++) { + snprintf(output_name, sizeof(output_name), "output_%d", i); + ret = dict_get_str(dict, output_name, &output); + if (ret) { + gf_log("", GF_LOG_ERROR, "Failed to get %s.", output_name); + cli_out("Unable to fetch output."); + } + if (output) { + bytes_written = sys_write(fd, output, strlen(output)); + if (bytes_written != strlen(output)) { + gf_log("", GF_LOG_ERROR, + "Failed to write " + "to %s", + common_pem_file); ret = -1; goto out; - } - - snprintf (common_pem_file, sizeof(common_pem_file), - "%s/geo-replication/common_secret.pem.pub", - workdir); - - sys_unlink (common_pem_file); - - fd = open (common_pem_file, O_WRONLY | O_CREAT, 0600); - if (fd == -1) { - gf_log ("", GF_LOG_ERROR, "Failed to open %s" - " Error : %s", common_pem_file, - strerror (errno)); + } + /* Adding the new line character */ + bytes_written = sys_write(fd, "\n", 1); + if (bytes_written != 1) { + gf_log("", GF_LOG_ERROR, "Failed to add new line char"); ret = -1; goto out; + } + output = NULL; } + } - for (i = 1; i <= output_count; i++) { - snprintf (output_name, sizeof (output_name), - "output_%d", i); - ret = dict_get_str (dict, output_name, &output); - if (ret) { - gf_log ("", GF_LOG_ERROR, "Failed to get %s.", - output_name); - cli_out ("Unable to fetch output."); - } - if (output) { - bytes_written = sys_write (fd, output, strlen(output)); - if (bytes_written != strlen(output)) { - gf_log ("", GF_LOG_ERROR, "Failed to write " - "to %s", common_pem_file); - ret = -1; - goto out; - } - /* Adding the new line character */ - bytes_written = sys_write (fd, "\n", 1); - if (bytes_written != 1) { - gf_log ("", GF_LOG_ERROR, - "Failed to add new line char"); - ret = -1; - goto out; - } - output = NULL; - } - } - - cli_out ("Common secret pub file present at %s", common_pem_file); - ret = 0; + cli_out("Common secret pub file present at %s", common_pem_file); + ret = 0; out: - if (fd >= 0) - sys_close (fd); + if (fd >= 0) + sys_close(fd); - gf_log ("", GF_LOG_DEBUG, "Returning %d", ret); - return ret; + gf_log("", GF_LOG_DEBUG, "Returning %d", ret); + return ret; } int -gf_cli_sys_exec_cbk (struct rpc_req *req, struct iovec *iov, - int count, void *myframe) -{ - int ret = -1; - int output_count = -1; - int i = -1; - char *output = NULL; - char *command = NULL; - char output_name[PATH_MAX] = ""; - gf_cli_rsp rsp = {0, }; - dict_t *dict = NULL; - - GF_ASSERT (myframe); - - if (req->rpc_status == -1) { - goto out; - } - - ret = xdr_to_generic (*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); - if (ret < 0) { - gf_log (((call_frame_t *) myframe)->this->name, GF_LOG_ERROR, - "Failed to decode xdr response"); - goto out; - } - - dict = dict_new (); - - if (!dict) { - ret = -1; - goto out; - } - - ret = dict_unserialize (rsp.dict.dict_val, rsp.dict.dict_len, &dict); +gf_cli_sys_exec_cbk(struct rpc_req *req, struct iovec *iov, int count, + void *myframe) +{ + int ret = -1; + int output_count = -1; + int i = -1; + char *output = NULL; + char *command = NULL; + char output_name[PATH_MAX] = ""; + gf_cli_rsp rsp = { + 0, + }; + dict_t *dict = NULL; + + GF_ASSERT(myframe); + + if (req->rpc_status == -1) { + goto out; + } + + ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); + if (ret < 0) { + gf_log(((call_frame_t *)myframe)->this->name, GF_LOG_ERROR, + "Failed to decode xdr response"); + goto out; + } + + dict = dict_new(); + + if (!dict) { + ret = -1; + goto out; + } + + ret = dict_unserialize(rsp.dict.dict_val, rsp.dict.dict_len, &dict); + + if (ret) + goto out; + + if (rsp.op_ret) { + cli_err("%s", rsp.op_errstr ? rsp.op_errstr : "Command failed."); + ret = rsp.op_ret; + goto out; + } - if (ret) - goto out; + ret = dict_get_int32(dict, "output_count", &output_count); + if (ret) { + cli_out("Command executed successfully."); + ret = 0; + goto out; + } - if (rsp.op_ret) { - cli_err ("%s", rsp.op_errstr ? rsp.op_errstr : - "Command failed."); - ret = rsp.op_ret; - goto out; - } + ret = dict_get_str(dict, "command", &command); + if (ret) { + gf_log("", GF_LOG_ERROR, "Unable to get command from dict"); + goto out; + } - ret = dict_get_int32 (dict, "output_count", &output_count); - if (ret) { - cli_out ("Command executed successfully."); - ret = 0; - goto out; - } + if (!strcmp(command, "gsec_create")) { + ret = write_contents_to_common_pem_file(dict, output_count); + if (!ret) + goto out; + } - ret = dict_get_str (dict, "command", &command); + for (i = 1; i <= output_count; i++) { + snprintf(output_name, sizeof(output_name), "output_%d", i); + ret = dict_get_str(dict, output_name, &output); if (ret) { - gf_log ("", GF_LOG_ERROR, - "Unable to get command from dict"); - goto out; + gf_log("", GF_LOG_ERROR, "Failed to get %s.", output_name); + cli_out("Unable to fetch output."); } - - if (!strcmp (command, "gsec_create")) { - ret = write_contents_to_common_pem_file (dict, output_count); - if (!ret) - goto out; + if (output) { + cli_out("%s", output); + output = NULL; } + } - for (i = 1; i <= output_count; i++) { - snprintf (output_name, sizeof (output_name), - "output_%d", i); - ret = dict_get_str (dict, output_name, &output); - if (ret) { - gf_log ("", GF_LOG_ERROR, "Failed to get %s.", - output_name); - cli_out ("Unable to fetch output."); - } - if (output) { - cli_out ("%s", output); - output = NULL; - } - } - - ret = 0; + ret = 0; out: - if (dict) - dict_unref (dict); - cli_cmd_broadcast_response (ret); + if (dict) + dict_unref(dict); + cli_cmd_broadcast_response(ret); - free (rsp.dict.dict_val); + free(rsp.dict.dict_val); - return ret; + return ret; } int -gf_cli_copy_file_cbk (struct rpc_req *req, struct iovec *iov, - int count, void *myframe) +gf_cli_copy_file_cbk(struct rpc_req *req, struct iovec *iov, int count, + void *myframe) { - int ret = -1; - gf_cli_rsp rsp = {0, }; - dict_t *dict = NULL; + int ret = -1; + gf_cli_rsp rsp = { + 0, + }; + dict_t *dict = NULL; - GF_ASSERT (myframe); + GF_ASSERT(myframe); - if (req->rpc_status == -1) { - goto out; - } + if (req->rpc_status == -1) { + goto out; + } - ret = xdr_to_generic (*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); - if (ret < 0) { - gf_log (((call_frame_t *) myframe)->this->name, GF_LOG_ERROR, - "Failed to decode xdr response"); - goto out; - } + ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); + if (ret < 0) { + gf_log(((call_frame_t *)myframe)->this->name, GF_LOG_ERROR, + "Failed to decode xdr response"); + goto out; + } - dict = dict_new (); + dict = dict_new(); - if (!dict) { - ret = -1; - goto out; - } + if (!dict) { + ret = -1; + goto out; + } - ret = dict_unserialize (rsp.dict.dict_val, rsp.dict.dict_len, &dict); + ret = dict_unserialize(rsp.dict.dict_val, rsp.dict.dict_len, &dict); - if (ret) - goto out; + if (ret) + goto out; - if (rsp.op_ret) { - cli_err ("%s", rsp.op_errstr ? rsp.op_errstr : - "Copy unsuccessful"); - ret = rsp.op_ret; - goto out; - } + if (rsp.op_ret) { + cli_err("%s", rsp.op_errstr ? rsp.op_errstr : "Copy unsuccessful"); + ret = rsp.op_ret; + goto out; + } - cli_out ("Successfully copied file."); + cli_out("Successfully copied file."); out: - if (dict) - dict_unref (dict); - cli_cmd_broadcast_response (ret); + if (dict) + dict_unref(dict); + cli_cmd_broadcast_response(ret); - free (rsp.dict.dict_val); + free(rsp.dict.dict_val); - return ret; + return ret; } int -gf_cli_gsync_set_cbk (struct rpc_req *req, struct iovec *iov, - int count, void *myframe) -{ - int ret = -1; - gf_cli_rsp rsp = {0, }; - dict_t *dict = NULL; - char *gsync_status = NULL; - char *master = NULL; - char *slave = NULL; - int32_t type = 0; - gf_boolean_t status_detail = _gf_false; - - GF_ASSERT (myframe); - - if (req->rpc_status == -1) { - goto out; - } - - ret = xdr_to_generic (*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); - if (ret < 0) { - gf_log (((call_frame_t *) myframe)->this->name, GF_LOG_ERROR, - "Failed to decode xdr response"); - goto out; - } - - dict = dict_new (); - - if (!dict) { - ret = -1; - goto out; - } - - ret = dict_unserialize (rsp.dict.dict_val, rsp.dict.dict_len, &dict); - +gf_cli_gsync_set_cbk(struct rpc_req *req, struct iovec *iov, int count, + void *myframe) +{ + int ret = -1; + gf_cli_rsp rsp = { + 0, + }; + dict_t *dict = NULL; + char *gsync_status = NULL; + char *master = NULL; + char *slave = NULL; + int32_t type = 0; + gf_boolean_t status_detail = _gf_false; + + GF_ASSERT(myframe); + + if (req->rpc_status == -1) { + goto out; + } + + ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); + if (ret < 0) { + gf_log(((call_frame_t *)myframe)->this->name, GF_LOG_ERROR, + "Failed to decode xdr response"); + goto out; + } + + dict = dict_new(); + + if (!dict) { + ret = -1; + goto out; + } + + ret = dict_unserialize(rsp.dict.dict_val, rsp.dict.dict_len, &dict); + + if (ret) + goto out; + + if (global_state->mode & GLUSTER_MODE_XML) { + ret = cli_xml_output_vol_gsync(dict, rsp.op_ret, rsp.op_errno, + rsp.op_errstr); if (ret) - goto out; + gf_log("cli", GF_LOG_ERROR, "Error outputting to xml"); + goto out; + } - if (global_state->mode & GLUSTER_MODE_XML) { - ret = cli_xml_output_vol_gsync (dict, rsp.op_ret, rsp.op_errno, - rsp.op_errstr); - if (ret) - gf_log ("cli", GF_LOG_ERROR, - "Error outputting to xml"); - goto out; - } - - if (rsp.op_ret) { - cli_err ("%s", rsp.op_errstr ? rsp.op_errstr : - GEOREP" command unsuccessful"); - ret = rsp.op_ret; - goto out; - } - - ret = dict_get_str (dict, "gsync-status", &gsync_status); - if (!ret) - cli_out ("%s", gsync_status); - - ret = dict_get_int32 (dict, "type", &type); - if (ret) { - gf_log (((call_frame_t *) myframe)->this->name, GF_LOG_ERROR, - "failed to get type"); - goto out; - } - - switch (type) { - case GF_GSYNC_OPTION_TYPE_START: - case GF_GSYNC_OPTION_TYPE_STOP: - if (dict_get_str (dict, "master", &master) != 0) - master = "???"; - if (dict_get_str (dict, "slave", &slave) != 0) - slave = "???"; - - cli_out ("%s " GEOREP " session between %s & %s" - " has been successful", - type == GF_GSYNC_OPTION_TYPE_START ? - "Starting" : "Stopping", - master, slave); - break; - - case GF_GSYNC_OPTION_TYPE_PAUSE: - case GF_GSYNC_OPTION_TYPE_RESUME: - if (dict_get_str (dict, "master", &master) != 0) - master = "???"; - if (dict_get_str (dict, "slave", &slave) != 0) - slave = "???"; - - cli_out ("%s " GEOREP " session between %s & %s" - " has been successful", - type == GF_GSYNC_OPTION_TYPE_PAUSE ? - "Pausing" : "Resuming", - master, slave); - break; - - case GF_GSYNC_OPTION_TYPE_CONFIG: - ret = gf_cli_gsync_config_command (dict); - break; - - case GF_GSYNC_OPTION_TYPE_STATUS: - status_detail = dict_get_str_boolean (dict, - "status-detail", - _gf_false); - ret = gf_cli_gsync_status_output (dict, status_detail); - break; - - case GF_GSYNC_OPTION_TYPE_DELETE: - if (dict_get_str (dict, "master", &master) != 0) - master = "???"; - if (dict_get_str (dict, "slave", &slave) != 0) - slave = "???"; - cli_out ("Deleting " GEOREP " session between %s & %s" - " has been successful", master, slave); - break; - - case GF_GSYNC_OPTION_TYPE_CREATE: - if (dict_get_str (dict, "master", &master) != 0) - master = "???"; - if (dict_get_str (dict, "slave", &slave) != 0) - slave = "???"; - cli_out ("Creating " GEOREP " session between %s & %s" - " has been successful", master, slave); - break; + if (rsp.op_ret) { + cli_err("%s", + rsp.op_errstr ? rsp.op_errstr : GEOREP " command unsuccessful"); + ret = rsp.op_ret; + goto out; + } + + ret = dict_get_str(dict, "gsync-status", &gsync_status); + if (!ret) + cli_out("%s", gsync_status); + + ret = dict_get_int32(dict, "type", &type); + if (ret) { + gf_log(((call_frame_t *)myframe)->this->name, GF_LOG_ERROR, + "failed to get type"); + goto out; + } + + switch (type) { + case GF_GSYNC_OPTION_TYPE_START: + case GF_GSYNC_OPTION_TYPE_STOP: + if (dict_get_str(dict, "master", &master) != 0) + master = "???"; + if (dict_get_str(dict, "slave", &slave) != 0) + slave = "???"; + + cli_out( + "%s " GEOREP + " session between %s & %s" + " has been successful", + type == GF_GSYNC_OPTION_TYPE_START ? "Starting" : "Stopping", + master, slave); + break; + + case GF_GSYNC_OPTION_TYPE_PAUSE: + case GF_GSYNC_OPTION_TYPE_RESUME: + if (dict_get_str(dict, "master", &master) != 0) + master = "???"; + if (dict_get_str(dict, "slave", &slave) != 0) + slave = "???"; + + cli_out("%s " GEOREP + " session between %s & %s" + " has been successful", + type == GF_GSYNC_OPTION_TYPE_PAUSE ? "Pausing" : "Resuming", + master, slave); + break; + + case GF_GSYNC_OPTION_TYPE_CONFIG: + ret = gf_cli_gsync_config_command(dict); + break; + + case GF_GSYNC_OPTION_TYPE_STATUS: + status_detail = dict_get_str_boolean(dict, "status-detail", + _gf_false); + ret = gf_cli_gsync_status_output(dict, status_detail); + break; + + case GF_GSYNC_OPTION_TYPE_DELETE: + if (dict_get_str(dict, "master", &master) != 0) + master = "???"; + if (dict_get_str(dict, "slave", &slave) != 0) + slave = "???"; + cli_out("Deleting " GEOREP + " session between %s & %s" + " has been successful", + master, slave); + break; + + case GF_GSYNC_OPTION_TYPE_CREATE: + if (dict_get_str(dict, "master", &master) != 0) + master = "???"; + if (dict_get_str(dict, "slave", &slave) != 0) + slave = "???"; + cli_out("Creating " GEOREP + " session between %s & %s" + " has been successful", + master, slave); + break; - default: - cli_out (GEOREP" command executed successfully"); - } + default: + cli_out(GEOREP " command executed successfully"); + } out: - if (dict) - dict_unref (dict); - cli_cmd_broadcast_response (ret); + if (dict) + dict_unref(dict); + cli_cmd_broadcast_response(ret); - free (rsp.dict.dict_val); + free(rsp.dict.dict_val); - return ret; + return ret; } int32_t -gf_cli_sys_exec (call_frame_t *frame, xlator_t *this, void *data) +gf_cli_sys_exec(call_frame_t *frame, xlator_t *this, void *data) { - int ret = 0; - dict_t *dict = NULL; - gf_cli_req req = {{0,}}; + int ret = 0; + dict_t *dict = NULL; + gf_cli_req req = {{ + 0, + }}; - if (!frame || !this || !data) { - ret = -1; - gf_log ("cli", GF_LOG_ERROR, "Invalid data"); - goto out; - } + if (!frame || !this || !data) { + ret = -1; + gf_log("cli", GF_LOG_ERROR, "Invalid data"); + goto out; + } - dict = data; + dict = data; - ret = cli_to_glusterd (&req, frame, gf_cli_sys_exec_cbk, - (xdrproc_t) xdr_gf_cli_req, dict, - GLUSTER_CLI_SYS_EXEC, this, cli_rpc_prog, - NULL); + ret = cli_to_glusterd(&req, frame, gf_cli_sys_exec_cbk, + (xdrproc_t)xdr_gf_cli_req, dict, GLUSTER_CLI_SYS_EXEC, + this, cli_rpc_prog, NULL); out: - GF_FREE (req.dict.dict_val); - return ret; + GF_FREE(req.dict.dict_val); + return ret; } int32_t -gf_cli_copy_file (call_frame_t *frame, xlator_t *this, void *data) +gf_cli_copy_file(call_frame_t *frame, xlator_t *this, void *data) { - int ret = 0; - dict_t *dict = NULL; - gf_cli_req req = {{0,}}; + int ret = 0; + dict_t *dict = NULL; + gf_cli_req req = {{ + 0, + }}; - if (!frame || !this || !data) { - ret = -1; - gf_log ("cli", GF_LOG_ERROR, "Invalid data"); - goto out; - } + if (!frame || !this || !data) { + ret = -1; + gf_log("cli", GF_LOG_ERROR, "Invalid data"); + goto out; + } - dict = data; + dict = data; - ret = cli_to_glusterd (&req, frame, gf_cli_copy_file_cbk, - (xdrproc_t) xdr_gf_cli_req, dict, - GLUSTER_CLI_COPY_FILE, this, cli_rpc_prog, - NULL); + ret = cli_to_glusterd(&req, frame, gf_cli_copy_file_cbk, + (xdrproc_t)xdr_gf_cli_req, dict, + GLUSTER_CLI_COPY_FILE, this, cli_rpc_prog, NULL); out: - GF_FREE (req.dict.dict_val); - return ret; + GF_FREE(req.dict.dict_val); + return ret; } int32_t -gf_cli_gsync_set (call_frame_t *frame, xlator_t *this, - void *data) +gf_cli_gsync_set(call_frame_t *frame, xlator_t *this, void *data) { - int ret = 0; - dict_t *dict = NULL; - gf_cli_req req = {{0,}}; + int ret = 0; + dict_t *dict = NULL; + gf_cli_req req = {{ + 0, + }}; - if (!frame || !this || !data) { - ret = -1; - goto out; - } + if (!frame || !this || !data) { + ret = -1; + goto out; + } - dict = data; + dict = data; - ret = cli_to_glusterd (&req, frame, gf_cli_gsync_set_cbk, - (xdrproc_t) xdr_gf_cli_req, dict, - GLUSTER_CLI_GSYNC_SET, this, cli_rpc_prog, - NULL); + ret = cli_to_glusterd(&req, frame, gf_cli_gsync_set_cbk, + (xdrproc_t)xdr_gf_cli_req, dict, + GLUSTER_CLI_GSYNC_SET, this, cli_rpc_prog, NULL); out: - GF_FREE (req.dict.dict_val); + GF_FREE(req.dict.dict_val); - return ret; + return ret; } - int -cli_profile_info_percentage_cmp (void *a, void *b) -{ - cli_profile_info_t *ia = NULL; - cli_profile_info_t *ib = NULL; - int ret = 0; - - ia = a; - ib = b; - if (ia->percentage_avg_latency < ib->percentage_avg_latency) - ret = -1; - else if (ia->percentage_avg_latency > ib->percentage_avg_latency) - ret = 1; - else - ret = 0; - return ret; +cli_profile_info_percentage_cmp(void *a, void *b) +{ + cli_profile_info_t *ia = NULL; + cli_profile_info_t *ib = NULL; + int ret = 0; + + ia = a; + ib = b; + if (ia->percentage_avg_latency < ib->percentage_avg_latency) + ret = -1; + else if (ia->percentage_avg_latency > ib->percentage_avg_latency) + ret = 1; + else + ret = 0; + return ret; } - void -cmd_profile_volume_brick_out (dict_t *dict, int count, int interval) -{ - char key[256] = {0}; - int i = 0; - uint64_t sec = 0; - uint64_t r_count = 0; - uint64_t w_count = 0; - uint64_t rb_counts[32] = {0}; - uint64_t wb_counts[32] = {0}; - cli_profile_info_t profile_info[GF_FOP_MAXVALUE] = {{0}}; - cli_profile_info_t upcall_info[GF_UPCALL_FLAGS_MAXVALUE] = {{0},}; - char output[128] = {0}; - int per_line = 0; - char read_blocks[128] = {0}; - char write_blocks[128] = {0}; - int index = 0; - int is_header_printed = 0; - int ret = 0; - double total_percentage_latency = 0; - - for (i = 0; i < 32; i++) { - snprintf (key, sizeof (key), "%d-%d-read-%d", count, - interval, (1 << i)); - ret = dict_get_uint64 (dict, key, &rb_counts[i]); - if (ret) { - gf_log ("cli", GF_LOG_DEBUG, - "failed to get %s from dict", key); - } +cmd_profile_volume_brick_out(dict_t *dict, int count, int interval) +{ + char key[256] = {0}; + int i = 0; + uint64_t sec = 0; + uint64_t r_count = 0; + uint64_t w_count = 0; + uint64_t rb_counts[32] = {0}; + uint64_t wb_counts[32] = {0}; + cli_profile_info_t profile_info[GF_FOP_MAXVALUE] = {{0}}; + cli_profile_info_t upcall_info[GF_UPCALL_FLAGS_MAXVALUE] = { + {0}, + }; + char output[128] = {0}; + int per_line = 0; + char read_blocks[128] = {0}; + char write_blocks[128] = {0}; + int index = 0; + int is_header_printed = 0; + int ret = 0; + double total_percentage_latency = 0; + + for (i = 0; i < 32; i++) { + snprintf(key, sizeof(key), "%d-%d-read-%d", count, interval, (1 << i)); + ret = dict_get_uint64(dict, key, &rb_counts[i]); + if (ret) { + gf_log("cli", GF_LOG_DEBUG, "failed to get %s from dict", key); } + } - for (i = 0; i < 32; i++) { - snprintf (key, sizeof (key), "%d-%d-write-%d", count, interval, - (1<<i)); - ret = dict_get_uint64 (dict, key, &wb_counts[i]); - if (ret) { - gf_log ("cli", GF_LOG_DEBUG, - "failed to get %s from dict", key); - } + for (i = 0; i < 32; i++) { + snprintf(key, sizeof(key), "%d-%d-write-%d", count, interval, (1 << i)); + ret = dict_get_uint64(dict, key, &wb_counts[i]); + if (ret) { + gf_log("cli", GF_LOG_DEBUG, "failed to get %s from dict", key); } + } - for (i = 0; i < GF_UPCALL_FLAGS_MAXVALUE; i++) { - snprintf (key, sizeof (key), "%d-%d-%d-upcall-hits", count, - interval, i); - ret = dict_get_uint64 (dict, key, &upcall_info[i].fop_hits); - if (ret) { - gf_log ("cli", GF_LOG_DEBUG, - "failed to get %s from dict", key); - } - upcall_info[i].fop_name = (char *)gf_upcall_list[i]; + for (i = 0; i < GF_UPCALL_FLAGS_MAXVALUE; i++) { + snprintf(key, sizeof(key), "%d-%d-%d-upcall-hits", count, interval, i); + ret = dict_get_uint64(dict, key, &upcall_info[i].fop_hits); + if (ret) { + gf_log("cli", GF_LOG_DEBUG, "failed to get %s from dict", key); } + upcall_info[i].fop_name = (char *)gf_upcall_list[i]; + } - for (i = 0; i < GF_FOP_MAXVALUE; i++) { - snprintf (key, sizeof (key), "%d-%d-%d-hits", count, - interval, i); - ret = dict_get_uint64 (dict, key, &profile_info[i].fop_hits); - if (ret) { - gf_log ("cli", GF_LOG_DEBUG, - "failed to get %s from dict", key); - } - - snprintf (key, sizeof (key), "%d-%d-%d-avglatency", count, - interval, i); - ret = dict_get_double (dict, key, &profile_info[i].avg_latency); - if (ret) { - gf_log ("cli", GF_LOG_DEBUG, - "failed to get %s from dict", key); - } - - snprintf (key, sizeof (key), "%d-%d-%d-minlatency", count, - interval, i); - ret = dict_get_double (dict, key, &profile_info[i].min_latency); - if (ret) { - gf_log ("cli", GF_LOG_DEBUG, - "failed to get %s from dict", key); - } - - snprintf (key, sizeof (key), "%d-%d-%d-maxlatency", count, - interval, i); - ret = dict_get_double (dict, key, &profile_info[i].max_latency); - if (ret) { - gf_log ("cli", GF_LOG_DEBUG, - "failed to get %s from dict", key); - } - profile_info[i].fop_name = (char *)gf_fop_list[i]; - - total_percentage_latency += - (profile_info[i].fop_hits * profile_info[i].avg_latency); - } - if (total_percentage_latency) { - for (i = 0; i < GF_FOP_MAXVALUE; i++) { - profile_info[i].percentage_avg_latency = 100 * ( - (profile_info[i].avg_latency* profile_info[i].fop_hits) / - total_percentage_latency); - } - gf_array_insertionsort (profile_info, 1, GF_FOP_MAXVALUE - 1, - sizeof (cli_profile_info_t), - cli_profile_info_percentage_cmp); - } - snprintf (key, sizeof (key), "%d-%d-duration", count, interval); - ret = dict_get_uint64 (dict, key, &sec); + for (i = 0; i < GF_FOP_MAXVALUE; i++) { + snprintf(key, sizeof(key), "%d-%d-%d-hits", count, interval, i); + ret = dict_get_uint64(dict, key, &profile_info[i].fop_hits); if (ret) { - gf_log ("cli", GF_LOG_DEBUG, "failed to get %s from dict", key); + gf_log("cli", GF_LOG_DEBUG, "failed to get %s from dict", key); } - snprintf (key, sizeof (key), "%d-%d-total-read", count, interval); - ret = dict_get_uint64 (dict, key, &r_count); + snprintf(key, sizeof(key), "%d-%d-%d-avglatency", count, interval, i); + ret = dict_get_double(dict, key, &profile_info[i].avg_latency); if (ret) { - gf_log ("cli", GF_LOG_DEBUG, "failed to get %s from dict", key); + gf_log("cli", GF_LOG_DEBUG, "failed to get %s from dict", key); } - snprintf (key, sizeof (key), "%d-%d-total-write", count, interval); - ret = dict_get_uint64 (dict, key, &w_count); + snprintf(key, sizeof(key), "%d-%d-%d-minlatency", count, interval, i); + ret = dict_get_double(dict, key, &profile_info[i].min_latency); if (ret) { - gf_log ("cli", GF_LOG_DEBUG, "failed to get %s from dict", key); + gf_log("cli", GF_LOG_DEBUG, "failed to get %s from dict", key); } - if (interval == -1) - cli_out ("Cumulative Stats:"); - else - cli_out ("Interval %d Stats:", interval); - snprintf (output, sizeof (output), "%14s", "Block Size:"); - snprintf (read_blocks, sizeof (read_blocks), "%14s", "No. of Reads:"); - snprintf (write_blocks, sizeof (write_blocks), "%14s", "No. of Writes:"); - index = 14; - for (i = 0; i < 32; i++) { - if ((rb_counts[i] == 0) && (wb_counts[i] == 0)) - continue; - per_line++; - snprintf (output+index, sizeof (output)-index, "%19db+ ", (1<<i)); - if (rb_counts[i]) { - snprintf (read_blocks+index, sizeof (read_blocks)-index, - "%21"PRId64" ", rb_counts[i]); - } else { - snprintf (read_blocks+index, sizeof (read_blocks)-index, - "%21s ", "0"); - } - if (wb_counts[i]) { - snprintf (write_blocks+index, sizeof (write_blocks)-index, - "%21"PRId64" ", wb_counts[i]); - } else { - snprintf (write_blocks+index, sizeof (write_blocks)-index, - "%21s ", "0"); - } - index += 22; - if (per_line == 3) { - cli_out ("%s", output); - cli_out ("%s", read_blocks); - cli_out ("%s", write_blocks); - cli_out (" "); - per_line = 0; - snprintf (output, sizeof (output), "%14s", "Block Size:"); - snprintf (read_blocks, sizeof (read_blocks), "%14s", - "No. of Reads:"); - snprintf (write_blocks, sizeof (write_blocks), "%14s", - "No. of Writes:"); - index = 14; - } + snprintf(key, sizeof(key), "%d-%d-%d-maxlatency", count, interval, i); + ret = dict_get_double(dict, key, &profile_info[i].max_latency); + if (ret) { + gf_log("cli", GF_LOG_DEBUG, "failed to get %s from dict", key); } + profile_info[i].fop_name = (char *)gf_fop_list[i]; - if (per_line != 0) { - cli_out ("%s", output); - cli_out ("%s", read_blocks); - cli_out ("%s", write_blocks); - } + total_percentage_latency += (profile_info[i].fop_hits * + profile_info[i].avg_latency); + } + if (total_percentage_latency) { for (i = 0; i < GF_FOP_MAXVALUE; i++) { - if (profile_info[i].fop_hits == 0) - continue; - if (is_header_printed == 0) { - cli_out ("%10s %13s %13s %13s %14s %11s", "%-latency", - "Avg-latency", "Min-Latency", "Max-Latency", - "No. of calls", "Fop"); - cli_out ("%10s %13s %13s %13s %14s %11s", "---------", - "-----------", "-----------", "-----------", - "------------", "----"); - is_header_printed = 1; - } - if (profile_info[i].fop_hits) { - cli_out ("%10.2lf %10.2lf us %10.2lf us %10.2lf us" - " %14"PRId64" %11s", - profile_info[i].percentage_avg_latency, - profile_info[i].avg_latency, - profile_info[i].min_latency, - profile_info[i].max_latency, - profile_info[i].fop_hits, - profile_info[i].fop_name); - } - } - - for (i = 0; i < GF_UPCALL_FLAGS_MAXVALUE; i++) { - if (upcall_info[i].fop_hits == 0) - continue; - if (upcall_info[i].fop_hits) { - cli_out ("%10.2lf %10.2lf us %10.2lf us %10.2lf us" - " %14"PRId64" %11s", - upcall_info[i].percentage_avg_latency, - upcall_info[i].avg_latency, - upcall_info[i].min_latency, - upcall_info[i].max_latency, - upcall_info[i].fop_hits, - upcall_info[i].fop_name); - } + profile_info[i] + .percentage_avg_latency = 100 * ((profile_info[i].avg_latency * + profile_info[i].fop_hits) / + total_percentage_latency); + } + gf_array_insertionsort(profile_info, 1, GF_FOP_MAXVALUE - 1, + sizeof(cli_profile_info_t), + cli_profile_info_percentage_cmp); + } + snprintf(key, sizeof(key), "%d-%d-duration", count, interval); + ret = dict_get_uint64(dict, key, &sec); + if (ret) { + gf_log("cli", GF_LOG_DEBUG, "failed to get %s from dict", key); + } + + snprintf(key, sizeof(key), "%d-%d-total-read", count, interval); + ret = dict_get_uint64(dict, key, &r_count); + if (ret) { + gf_log("cli", GF_LOG_DEBUG, "failed to get %s from dict", key); + } + + snprintf(key, sizeof(key), "%d-%d-total-write", count, interval); + ret = dict_get_uint64(dict, key, &w_count); + if (ret) { + gf_log("cli", GF_LOG_DEBUG, "failed to get %s from dict", key); + } + + if (interval == -1) + cli_out("Cumulative Stats:"); + else + cli_out("Interval %d Stats:", interval); + snprintf(output, sizeof(output), "%14s", "Block Size:"); + snprintf(read_blocks, sizeof(read_blocks), "%14s", "No. of Reads:"); + snprintf(write_blocks, sizeof(write_blocks), "%14s", "No. of Writes:"); + index = 14; + for (i = 0; i < 32; i++) { + if ((rb_counts[i] == 0) && (wb_counts[i] == 0)) + continue; + per_line++; + snprintf(output + index, sizeof(output) - index, "%19db+ ", (1 << i)); + if (rb_counts[i]) { + snprintf(read_blocks + index, sizeof(read_blocks) - index, + "%21" PRId64 " ", rb_counts[i]); + } else { + snprintf(read_blocks + index, sizeof(read_blocks) - index, "%21s ", + "0"); } - - cli_out (" "); - cli_out ("%12s: %"PRId64" seconds", "Duration", sec); - cli_out ("%12s: %"PRId64" bytes", "Data Read", r_count); - cli_out ("%12s: %"PRId64" bytes", "Data Written", w_count); - cli_out (" "); + if (wb_counts[i]) { + snprintf(write_blocks + index, sizeof(write_blocks) - index, + "%21" PRId64 " ", wb_counts[i]); + } else { + snprintf(write_blocks + index, sizeof(write_blocks) - index, + "%21s ", "0"); + } + index += 22; + if (per_line == 3) { + cli_out("%s", output); + cli_out("%s", read_blocks); + cli_out("%s", write_blocks); + cli_out(" "); + per_line = 0; + snprintf(output, sizeof(output), "%14s", "Block Size:"); + snprintf(read_blocks, sizeof(read_blocks), "%14s", "No. of Reads:"); + snprintf(write_blocks, sizeof(write_blocks), "%14s", + "No. of Writes:"); + index = 14; + } + } + + if (per_line != 0) { + cli_out("%s", output); + cli_out("%s", read_blocks); + cli_out("%s", write_blocks); + } + for (i = 0; i < GF_FOP_MAXVALUE; i++) { + if (profile_info[i].fop_hits == 0) + continue; + if (is_header_printed == 0) { + cli_out("%10s %13s %13s %13s %14s %11s", "%-latency", "Avg-latency", + "Min-Latency", "Max-Latency", "No. of calls", "Fop"); + cli_out("%10s %13s %13s %13s %14s %11s", "---------", "-----------", + "-----------", "-----------", "------------", "----"); + is_header_printed = 1; + } + if (profile_info[i].fop_hits) { + cli_out( + "%10.2lf %10.2lf us %10.2lf us %10.2lf us" + " %14" PRId64 " %11s", + profile_info[i].percentage_avg_latency, + profile_info[i].avg_latency, profile_info[i].min_latency, + profile_info[i].max_latency, profile_info[i].fop_hits, + profile_info[i].fop_name); + } + } + + for (i = 0; i < GF_UPCALL_FLAGS_MAXVALUE; i++) { + if (upcall_info[i].fop_hits == 0) + continue; + if (upcall_info[i].fop_hits) { + cli_out( + "%10.2lf %10.2lf us %10.2lf us %10.2lf us" + " %14" PRId64 " %11s", + upcall_info[i].percentage_avg_latency, + upcall_info[i].avg_latency, upcall_info[i].min_latency, + upcall_info[i].max_latency, upcall_info[i].fop_hits, + upcall_info[i].fop_name); + } + } + + cli_out(" "); + cli_out("%12s: %" PRId64 " seconds", "Duration", sec); + cli_out("%12s: %" PRId64 " bytes", "Data Read", r_count); + cli_out("%12s: %" PRId64 " bytes", "Data Written", w_count); + cli_out(" "); } int32_t -gf_cli_profile_volume_cbk (struct rpc_req *req, struct iovec *iov, - int count, void *myframe) -{ - gf_cli_rsp rsp = {0,}; - int ret = -1; - dict_t *dict = NULL; - gf1_cli_stats_op op = GF_CLI_STATS_NONE; - char key[256] = {0}; - int interval = 0; - int i = 1; - int32_t brick_count = 0; - char *volname = NULL; - char *brick = NULL; - char str[1024] = {0,}; - int stats_cleared = 0; - gf1_cli_info_op info_op = GF_CLI_INFO_NONE; - - GF_ASSERT (myframe); - - if (-1 == req->rpc_status) { - goto out; +gf_cli_profile_volume_cbk(struct rpc_req *req, struct iovec *iov, int count, + void *myframe) +{ + gf_cli_rsp rsp = { + 0, + }; + int ret = -1; + dict_t *dict = NULL; + gf1_cli_stats_op op = GF_CLI_STATS_NONE; + char key[256] = {0}; + int interval = 0; + int i = 1; + int32_t brick_count = 0; + char *volname = NULL; + char *brick = NULL; + char str[1024] = { + 0, + }; + int stats_cleared = 0; + gf1_cli_info_op info_op = GF_CLI_INFO_NONE; + + GF_ASSERT(myframe); + + if (-1 == req->rpc_status) { + goto out; + } + + gf_log("cli", GF_LOG_DEBUG, "Received resp to profile"); + ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); + if (ret < 0) { + gf_log(((call_frame_t *)myframe)->this->name, GF_LOG_ERROR, + "Failed to decode xdr response"); + goto out; + } + + dict = dict_new(); + + if (!dict) { + ret = -1; + goto out; + } + + ret = dict_unserialize(rsp.dict.dict_val, rsp.dict.dict_len, &dict); + + if (ret) { + gf_log("", GF_LOG_ERROR, "Unable to allocate memory"); + goto out; + } else { + dict->extra_stdfree = rsp.dict.dict_val; + } + + if (global_state->mode & GLUSTER_MODE_XML) { + ret = cli_xml_output_vol_profile(dict, rsp.op_ret, rsp.op_errno, + rsp.op_errstr); + if (ret) + gf_log("cli", GF_LOG_ERROR, "Error outputting to xml"); + goto out; + } + + ret = dict_get_str(dict, "volname", &volname); + if (ret) + goto out; + + ret = dict_get_int32(dict, "op", (int32_t *)&op); + if (ret) + goto out; + + if (rsp.op_ret && strcmp(rsp.op_errstr, "")) { + cli_err("%s", rsp.op_errstr); + } else { + switch (op) { + case GF_CLI_STATS_START: + cli_out("Starting volume profile on %s has been %s ", volname, + (rsp.op_ret) ? "unsuccessful" : "successful"); + break; + case GF_CLI_STATS_STOP: + cli_out("Stopping volume profile on %s has been %s ", volname, + (rsp.op_ret) ? "unsuccessful" : "successful"); + break; + case GF_CLI_STATS_INFO: + break; + default: + cli_out("volume profile on %s has been %s ", volname, + (rsp.op_ret) ? "unsuccessful" : "successful"); + break; } + } - gf_log ("cli", GF_LOG_DEBUG, "Received resp to profile"); - ret = xdr_to_generic (*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); - if (ret < 0) { - gf_log (((call_frame_t *) myframe)->this->name, GF_LOG_ERROR, - "Failed to decode xdr response"); - goto out; - } + if (rsp.op_ret) { + ret = rsp.op_ret; + goto out; + } - dict = dict_new (); + if (GF_CLI_STATS_INFO != op) { + ret = 0; + goto out; + } - if (!dict) { - ret = -1; - goto out; - } + ret = dict_get_int32(dict, "info-op", (int32_t *)&info_op); + if (ret) + goto out; + + ret = dict_get_int32(dict, "count", &brick_count); + if (ret) + goto out; - ret = dict_unserialize (rsp.dict.dict_val, - rsp.dict.dict_len, - &dict); + if (!brick_count) { + cli_err("All bricks of volume %s are down.", volname); + goto out; + } + while (i <= brick_count) { + snprintf(key, sizeof(key), "%d-brick", i); + ret = dict_get_str(dict, key, &brick); if (ret) { - gf_log ("", GF_LOG_ERROR, - "Unable to allocate memory"); - goto out; - } else { - dict->extra_stdfree = rsp.dict.dict_val; + gf_log("cli", GF_LOG_ERROR, "Couldn't get brick name"); + goto out; } - if (global_state->mode & GLUSTER_MODE_XML) { - ret = cli_xml_output_vol_profile (dict, rsp.op_ret, - rsp.op_errno, - rsp.op_errstr); - if (ret) - gf_log ("cli", GF_LOG_ERROR, - "Error outputting to xml"); - goto out; - } + ret = dict_get_str_boolean(dict, "nfs", _gf_false); - ret = dict_get_str (dict, "volname", &volname); if (ret) - goto out; - - ret = dict_get_int32 (dict, "op", (int32_t*)&op); - if (ret) - goto out; - - if (rsp.op_ret && strcmp (rsp.op_errstr, "")) { - cli_err ("%s", rsp.op_errstr); + snprintf(str, sizeof(str), "NFS Server : %s", brick); + else + snprintf(str, sizeof(str), "Brick: %s", brick); + cli_out("%s", str); + memset(str, '-', strlen(str)); + cli_out("%s", str); + + if (GF_CLI_INFO_CLEAR == info_op) { + snprintf(key, sizeof(key), "%d-stats-cleared", i); + ret = dict_get_int32(dict, key, &stats_cleared); + if (ret) + goto out; + cli_out(stats_cleared ? "Cleared stats." + : "Failed to clear stats."); } else { - switch (op) { - case GF_CLI_STATS_START: - cli_out ("Starting volume profile on %s has been %s ", - volname, - (rsp.op_ret) ? "unsuccessful": "successful"); - break; - case GF_CLI_STATS_STOP: - cli_out ("Stopping volume profile on %s has been %s ", - volname, - (rsp.op_ret) ? "unsuccessful": "successful"); - break; - case GF_CLI_STATS_INFO: - break; - default: - cli_out ("volume profile on %s has been %s ", - volname, - (rsp.op_ret) ? "unsuccessful": "successful"); - break; - } - } - - if (rsp.op_ret) { - ret = rsp.op_ret; - goto out; - } + snprintf(key, sizeof(key), "%d-cumulative", i); + ret = dict_get_int32(dict, key, &interval); + if (ret == 0) + cmd_profile_volume_brick_out(dict, i, interval); - if (GF_CLI_STATS_INFO != op) { - ret = 0; - goto out; + snprintf(key, sizeof(key), "%d-interval", i); + ret = dict_get_int32(dict, key, &interval); + if (ret == 0) + cmd_profile_volume_brick_out(dict, i, interval); } - - ret = dict_get_int32 (dict, "info-op", (int32_t*)&info_op); - if (ret) - goto out; - - ret = dict_get_int32 (dict, "count", &brick_count); - if (ret) - goto out; - - if (!brick_count) { - cli_err ("All bricks of volume %s are down.", volname); - goto out; - } - - while (i <= brick_count) { - snprintf (key, sizeof (key), "%d-brick", i); - ret = dict_get_str (dict, key, &brick); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Couldn't get brick name"); - goto out; - } - - ret = dict_get_str_boolean (dict, "nfs", _gf_false); - - if (ret) - snprintf (str, sizeof (str), "NFS Server : %s", brick); - else - snprintf (str, sizeof (str), "Brick: %s", brick); - cli_out ("%s", str); - memset (str, '-', strlen (str)); - cli_out ("%s", str); - - if (GF_CLI_INFO_CLEAR == info_op) { - snprintf (key, sizeof (key), "%d-stats-cleared", i); - ret = dict_get_int32 (dict, key, &stats_cleared); - if (ret) - goto out; - cli_out (stats_cleared ? "Cleared stats." : - "Failed to clear stats."); - } else { - snprintf (key, sizeof (key), "%d-cumulative", i); - ret = dict_get_int32 (dict, key, &interval); - if (ret == 0) - cmd_profile_volume_brick_out (dict, i, - interval); - - snprintf (key, sizeof (key), "%d-interval", i); - ret = dict_get_int32 (dict, key, &interval); - if (ret == 0) - cmd_profile_volume_brick_out (dict, i, - interval); - } - i++; - } - ret = rsp.op_ret; + i++; + } + ret = rsp.op_ret; out: - if (dict) - dict_unref (dict); - free (rsp.op_errstr); - cli_cmd_broadcast_response (ret); - return ret; + if (dict) + dict_unref(dict); + free(rsp.op_errstr); + cli_cmd_broadcast_response(ret); + return ret; } int32_t -gf_cli_profile_volume (call_frame_t *frame, xlator_t *this, void *data) +gf_cli_profile_volume(call_frame_t *frame, xlator_t *this, void *data) { - int ret = -1; - gf_cli_req req = {{0,}}; - dict_t *dict = NULL; + int ret = -1; + gf_cli_req req = {{ + 0, + }}; + dict_t *dict = NULL; - GF_ASSERT (frame); - GF_ASSERT (this); - GF_ASSERT (data); + GF_ASSERT(frame); + GF_ASSERT(this); + GF_ASSERT(data); - if (!frame || !this || !data) - goto out; - dict = data; + if (!frame || !this || !data) + goto out; + dict = data; - ret = cli_to_glusterd (&req, frame, gf_cli_profile_volume_cbk, - (xdrproc_t) xdr_gf_cli_req, dict, - GLUSTER_CLI_PROFILE_VOLUME, this, cli_rpc_prog, - NULL); + ret = cli_to_glusterd(&req, frame, gf_cli_profile_volume_cbk, + (xdrproc_t)xdr_gf_cli_req, dict, + GLUSTER_CLI_PROFILE_VOLUME, this, cli_rpc_prog, NULL); out: - gf_log ("cli", GF_LOG_DEBUG, "Returning %d", ret); + gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret); - GF_FREE (req.dict.dict_val); - return ret; + GF_FREE(req.dict.dict_val); + return ret; } int32_t -gf_cli_top_volume_cbk (struct rpc_req *req, struct iovec *iov, - int count, void *myframe) -{ - gf_cli_rsp rsp = {0,}; - int ret = -1; - dict_t *dict = NULL; - gf1_cli_stats_op op = GF_CLI_STATS_NONE; - char key[256] = {0}; - int i = 0; - int32_t brick_count = 0; - char brick[1024]; - int32_t members = 0; - char *filename; - char *bricks; - uint64_t value = 0; - int32_t j = 0; - gf1_cli_top_op top_op = GF_CLI_TOP_NONE; - uint64_t nr_open = 0; - uint64_t max_nr_open = 0; - double throughput = 0; - double time = 0; - int32_t time_sec = 0; - long int time_usec = 0; - char timestr[256] = {0, }; - char *openfd_str = NULL; - gf_boolean_t nfs = _gf_false; - gf_boolean_t clear_stats = _gf_false; - int stats_cleared = 0; - - GF_ASSERT (myframe); - - if (-1 == req->rpc_status) { - goto out; - } +gf_cli_top_volume_cbk(struct rpc_req *req, struct iovec *iov, int count, + void *myframe) +{ + gf_cli_rsp rsp = { + 0, + }; + int ret = -1; + dict_t *dict = NULL; + gf1_cli_stats_op op = GF_CLI_STATS_NONE; + char key[256] = {0}; + int i = 0; + int32_t brick_count = 0; + char brick[1024]; + int32_t members = 0; + char *filename; + char *bricks; + uint64_t value = 0; + int32_t j = 0; + gf1_cli_top_op top_op = GF_CLI_TOP_NONE; + uint64_t nr_open = 0; + uint64_t max_nr_open = 0; + double throughput = 0; + double time = 0; + int32_t time_sec = 0; + long int time_usec = 0; + char timestr[256] = { + 0, + }; + char *openfd_str = NULL; + gf_boolean_t nfs = _gf_false; + gf_boolean_t clear_stats = _gf_false; + int stats_cleared = 0; + + GF_ASSERT(myframe); + + if (-1 == req->rpc_status) { + goto out; + } + + gf_log("cli", GF_LOG_DEBUG, "Received resp to top"); + ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); + if (ret < 0) { + gf_log(((call_frame_t *)myframe)->this->name, GF_LOG_ERROR, + "Failed to decode xdr response"); + goto out; + } + + if (rsp.op_ret) { + if (strcmp(rsp.op_errstr, "")) + cli_err("%s", rsp.op_errstr); + cli_err("volume top unsuccessful"); + ret = rsp.op_ret; + goto out; + } - gf_log ("cli", GF_LOG_DEBUG, "Received resp to top"); - ret = xdr_to_generic (*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); - if (ret < 0) { - gf_log (((call_frame_t *) myframe)->this->name, GF_LOG_ERROR, - "Failed to decode xdr response"); - goto out; - } + dict = dict_new(); - if (rsp.op_ret) { - if (strcmp (rsp.op_errstr, "")) - cli_err ("%s", rsp.op_errstr); - cli_err ("volume top unsuccessful"); - ret = rsp.op_ret; - goto out; - } + if (!dict) { + ret = -1; + goto out; + } - dict = dict_new (); + ret = dict_unserialize(rsp.dict.dict_val, rsp.dict.dict_len, &dict); - if (!dict) { - ret = -1; - goto out; - } + if (ret) { + gf_log("", GF_LOG_ERROR, "Unable to allocate memory"); + goto out; + } - ret = dict_unserialize (rsp.dict.dict_val, - rsp.dict.dict_len, - &dict); + ret = dict_get_int32(dict, "op", (int32_t *)&op); + if (op != GF_CLI_STATS_TOP) { + ret = 0; + goto out; + } + + if (global_state->mode & GLUSTER_MODE_XML) { + ret = cli_xml_output_vol_top(dict, rsp.op_ret, rsp.op_errno, + rsp.op_errstr); if (ret) { - gf_log ("", GF_LOG_ERROR, - "Unable to allocate memory"); - goto out; + gf_log("cli", GF_LOG_ERROR, "Error outputting to xml"); } + goto out; + } - ret = dict_get_int32 (dict, "op", (int32_t*)&op); + ret = dict_get_int32(dict, "count", &brick_count); + if (ret) + goto out; + snprintf(key, sizeof(key), "%d-top-op", 1); + ret = dict_get_int32(dict, key, (int32_t *)&top_op); + if (ret) + goto out; - if (op != GF_CLI_STATS_TOP) { - ret = 0; + clear_stats = dict_get_str_boolean(dict, "clear-stats", _gf_false); + + while (i < brick_count) { + i++; + snprintf(brick, sizeof(brick), "%d-brick", i); + ret = dict_get_str(dict, brick, &bricks); + if (ret) + goto out; + + nfs = dict_get_str_boolean(dict, "nfs", _gf_false); + + if (clear_stats) { + snprintf(key, sizeof(key), "%d-stats-cleared", i); + ret = dict_get_int32(dict, key, &stats_cleared); + if (ret) goto out; + cli_out(stats_cleared ? "Cleared stats for %s %s" + : "Failed to clear stats for %s %s", + nfs ? "NFS server on" : "brick", bricks); + continue; } - if (global_state->mode & GLUSTER_MODE_XML) { - ret = cli_xml_output_vol_top (dict, rsp.op_ret, - rsp.op_errno, - rsp.op_errstr); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, - "Error outputting to xml"); + if (nfs) + cli_out("NFS Server : %s", bricks); + else + cli_out("Brick: %s", bricks); + + snprintf(key, sizeof(key), "%d-members", i); + ret = dict_get_int32(dict, key, &members); + + switch (top_op) { + case GF_CLI_TOP_OPEN: + snprintf(key, sizeof(key), "%d-current-open", i); + ret = dict_get_uint64(dict, key, &nr_open); + if (ret) + break; + snprintf(key, sizeof(key), "%d-max-open", i); + ret = dict_get_uint64(dict, key, &max_nr_open); + if (ret) + goto out; + snprintf(key, sizeof(key), "%d-max-openfd-time", i); + ret = dict_get_str(dict, key, &openfd_str); + if (ret) + goto out; + cli_out("Current open fds: %" PRIu64 + ", Max open" + " fds: %" PRIu64 ", Max openfd time: %s", + nr_open, max_nr_open, openfd_str); + case GF_CLI_TOP_READ: + case GF_CLI_TOP_WRITE: + case GF_CLI_TOP_OPENDIR: + case GF_CLI_TOP_READDIR: + if (!members) { + continue; + } + cli_out("Count\t\tfilename\n======================="); + break; + case GF_CLI_TOP_READ_PERF: + case GF_CLI_TOP_WRITE_PERF: + snprintf(key, sizeof(key), "%d-throughput", i); + ret = dict_get_double(dict, key, &throughput); + if (!ret) { + snprintf(key, sizeof(key), "%d-time", i); + ret = dict_get_double(dict, key, &time); } + if (!ret) + cli_out("Throughput %.2f MBps time %.4f secs", throughput, + time / 1e6); + + if (!members) { + continue; + } + cli_out("%*s %-*s %-*s", VOL_TOP_PERF_SPEED_WIDTH, "MBps", + VOL_TOP_PERF_FILENAME_DEF_WIDTH, "Filename", + VOL_TOP_PERF_TIME_WIDTH, "Time"); + cli_out("%*s %-*s %-*s", VOL_TOP_PERF_SPEED_WIDTH, + "====", VOL_TOP_PERF_FILENAME_DEF_WIDTH, + "========", VOL_TOP_PERF_TIME_WIDTH, "===="); + break; + default: goto out; } - ret = dict_get_int32 (dict, "count", &brick_count); - if (ret) - goto out; - snprintf (key, sizeof (key), "%d-top-op", 1); - ret = dict_get_int32 (dict, key, (int32_t*)&top_op); - if (ret) - goto out; - - clear_stats = dict_get_str_boolean (dict, "clear-stats", _gf_false); - - while (i < brick_count) { - i++; - snprintf (brick, sizeof (brick), "%d-brick", i); - ret = dict_get_str (dict, brick, &bricks); - if (ret) - goto out; - - nfs = dict_get_str_boolean (dict, "nfs", _gf_false); - - if (clear_stats) { - snprintf (key, sizeof (key), "%d-stats-cleared", i); - ret = dict_get_int32 (dict, key, &stats_cleared); - if (ret) - goto out; - cli_out (stats_cleared ? "Cleared stats for %s %s" : - "Failed to clear stats for %s %s", - nfs ? "NFS server on" : "brick", bricks); - continue; - } - - if (nfs) - cli_out ("NFS Server : %s", bricks); + for (j = 1; j <= members; j++) { + snprintf(key, sizeof(key), "%d-filename-%d", i, j); + ret = dict_get_str(dict, key, &filename); + if (ret) + break; + snprintf(key, sizeof(key), "%d-value-%d", i, j); + ret = dict_get_uint64(dict, key, &value); + if (ret) + goto out; + if (top_op == GF_CLI_TOP_READ_PERF || + top_op == GF_CLI_TOP_WRITE_PERF) { + snprintf(key, sizeof(key), "%d-time-sec-%d", i, j); + ret = dict_get_int32(dict, key, (int32_t *)&time_sec); + if (ret) + goto out; + snprintf(key, sizeof(key), "%d-time-usec-%d", i, j); + ret = dict_get_int32(dict, key, (int32_t *)&time_usec); + if (ret) + goto out; + gf_time_fmt(timestr, sizeof timestr, time_sec, gf_timefmt_FT); + snprintf(timestr + strlen(timestr), + sizeof timestr - strlen(timestr), ".%ld", time_usec); + if (strlen(filename) < VOL_TOP_PERF_FILENAME_DEF_WIDTH) + cli_out("%*" PRIu64 " %-*s %-*s", VOL_TOP_PERF_SPEED_WIDTH, + value, VOL_TOP_PERF_FILENAME_DEF_WIDTH, filename, + VOL_TOP_PERF_TIME_WIDTH, timestr); else - cli_out ("Brick: %s", bricks); - - snprintf(key, sizeof (key), "%d-members", i); - ret = dict_get_int32 (dict, key, &members); - - switch (top_op) { - case GF_CLI_TOP_OPEN: - snprintf (key, sizeof (key), "%d-current-open", i); - ret = dict_get_uint64 (dict, key, &nr_open); - if (ret) - break; - snprintf (key, sizeof (key), "%d-max-open", i); - ret = dict_get_uint64 (dict, key, &max_nr_open); - if (ret) - goto out; - snprintf (key, sizeof (key), "%d-max-openfd-time", i); - ret = dict_get_str (dict, key, &openfd_str); - if (ret) - goto out; - cli_out ("Current open fds: %"PRIu64", Max open" - " fds: %"PRIu64", Max openfd time: %s", nr_open, - max_nr_open, openfd_str); - case GF_CLI_TOP_READ: - case GF_CLI_TOP_WRITE: - case GF_CLI_TOP_OPENDIR: - case GF_CLI_TOP_READDIR: - if (!members) { - continue; - } - cli_out ("Count\t\tfilename\n======================="); - break; - case GF_CLI_TOP_READ_PERF: - case GF_CLI_TOP_WRITE_PERF: - snprintf (key, sizeof (key), "%d-throughput", i); - ret = dict_get_double (dict, key, &throughput); - if (!ret) { - snprintf (key, sizeof (key), "%d-time", i); - ret = dict_get_double (dict, key, &time); - } - if (!ret) - cli_out ("Throughput %.2f MBps time %.4f secs", throughput, - time / 1e6); - - if (!members) { - continue; - } - cli_out ("%*s %-*s %-*s", - VOL_TOP_PERF_SPEED_WIDTH, "MBps", - VOL_TOP_PERF_FILENAME_DEF_WIDTH, "Filename", - VOL_TOP_PERF_TIME_WIDTH, "Time"); - cli_out ("%*s %-*s %-*s", - VOL_TOP_PERF_SPEED_WIDTH, "====", - VOL_TOP_PERF_FILENAME_DEF_WIDTH, "========", - VOL_TOP_PERF_TIME_WIDTH, "===="); - break; - default: - goto out; - } - - for (j = 1; j <= members; j++) { - snprintf (key, sizeof (key), "%d-filename-%d", i, j); - ret = dict_get_str (dict, key, &filename); - if (ret) - break; - snprintf (key, sizeof (key), "%d-value-%d", i, j); - ret = dict_get_uint64 (dict, key, &value); - if (ret) - goto out; - if ( top_op == GF_CLI_TOP_READ_PERF || - top_op == GF_CLI_TOP_WRITE_PERF) { - snprintf (key, sizeof (key), "%d-time-sec-%d", i, j); - ret = dict_get_int32 (dict, key, (int32_t *)&time_sec); - if (ret) - goto out; - snprintf (key, sizeof (key), "%d-time-usec-%d", i, j); - ret = dict_get_int32 (dict, key, (int32_t *)&time_usec); - if (ret) - goto out; - gf_time_fmt (timestr, sizeof timestr, - time_sec, gf_timefmt_FT); - snprintf (timestr + strlen (timestr), sizeof timestr - strlen (timestr), - ".%ld", time_usec); - if (strlen (filename) < VOL_TOP_PERF_FILENAME_DEF_WIDTH) - cli_out ("%*"PRIu64" %-*s %-*s", - VOL_TOP_PERF_SPEED_WIDTH, - value, - VOL_TOP_PERF_FILENAME_DEF_WIDTH, - filename, - VOL_TOP_PERF_TIME_WIDTH, - timestr); - else - cli_out ("%*"PRIu64" ...%-*s %-*s", - VOL_TOP_PERF_SPEED_WIDTH, - value, - VOL_TOP_PERF_FILENAME_ALT_WIDTH , - filename + strlen (filename) - - VOL_TOP_PERF_FILENAME_ALT_WIDTH, - VOL_TOP_PERF_TIME_WIDTH, - timestr); - } else { - cli_out ("%"PRIu64"\t\t%s", value, filename); - } - } + cli_out("%*" PRIu64 " ...%-*s %-*s", + VOL_TOP_PERF_SPEED_WIDTH, value, + VOL_TOP_PERF_FILENAME_ALT_WIDTH, + filename + strlen(filename) - + VOL_TOP_PERF_FILENAME_ALT_WIDTH, + VOL_TOP_PERF_TIME_WIDTH, timestr); + } else { + cli_out("%" PRIu64 "\t\t%s", value, filename); + } } - ret = rsp.op_ret; + } + ret = rsp.op_ret; out: - cli_cmd_broadcast_response (ret); + cli_cmd_broadcast_response(ret); - if (dict) - dict_unref (dict); + if (dict) + dict_unref(dict); - free (rsp.dict.dict_val); - return ret; + free(rsp.dict.dict_val); + return ret; } int32_t -gf_cli_top_volume (call_frame_t *frame, xlator_t *this, void *data) +gf_cli_top_volume(call_frame_t *frame, xlator_t *this, void *data) { - int ret = -1; - gf_cli_req req = {{0,}}; - dict_t *dict = NULL; + int ret = -1; + gf_cli_req req = {{ + 0, + }}; + dict_t *dict = NULL; - GF_ASSERT (frame); - GF_ASSERT (this); - GF_ASSERT (data); + GF_ASSERT(frame); + GF_ASSERT(this); + GF_ASSERT(data); - if (!frame || !this || !data) - goto out; - dict = data; + if (!frame || !this || !data) + goto out; + dict = data; - ret = cli_to_glusterd (&req, frame, gf_cli_top_volume_cbk, - (xdrproc_t) xdr_gf_cli_req, dict, - GLUSTER_CLI_PROFILE_VOLUME, this, cli_rpc_prog, - NULL); + ret = cli_to_glusterd(&req, frame, gf_cli_top_volume_cbk, + (xdrproc_t)xdr_gf_cli_req, dict, + GLUSTER_CLI_PROFILE_VOLUME, this, cli_rpc_prog, NULL); out: - gf_log ("cli", GF_LOG_DEBUG, "Returning %d", ret); - GF_FREE (req.dict.dict_val); - return ret; + gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret); + GF_FREE(req.dict.dict_val); + return ret; } - int -gf_cli_getwd_cbk (struct rpc_req *req, struct iovec *iov, - int count, void *myframe) +gf_cli_getwd_cbk(struct rpc_req *req, struct iovec *iov, int count, + void *myframe) { - gf1_cli_getwd_rsp rsp = {0,}; - int ret = -1; + gf1_cli_getwd_rsp rsp = { + 0, + }; + int ret = -1; - GF_ASSERT (myframe); + GF_ASSERT(myframe); - if (-1 == req->rpc_status) { - goto out; - } + if (-1 == req->rpc_status) { + goto out; + } - ret = xdr_to_generic (*iov, &rsp, (xdrproc_t)xdr_gf1_cli_getwd_rsp); - if (ret < 0) { - gf_log (((call_frame_t *) myframe)->this->name, GF_LOG_ERROR, - "Failed to decode xdr response"); - goto out; - } + ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf1_cli_getwd_rsp); + if (ret < 0) { + gf_log(((call_frame_t *)myframe)->this->name, GF_LOG_ERROR, + "Failed to decode xdr response"); + goto out; + } - if (rsp.op_ret == -1) { - cli_err ("getwd failed"); - ret = rsp.op_ret; - goto out; - } + if (rsp.op_ret == -1) { + cli_err("getwd failed"); + ret = rsp.op_ret; + goto out; + } - gf_log ("cli", GF_LOG_INFO, "Received resp to getwd"); + gf_log("cli", GF_LOG_INFO, "Received resp to getwd"); - cli_out ("%s", rsp.wd); + cli_out("%s", rsp.wd); - ret = 0; + ret = 0; out: - cli_cmd_broadcast_response (ret); - return ret; + cli_cmd_broadcast_response(ret); + return ret; } int32_t -gf_cli_getwd (call_frame_t *frame, xlator_t *this, void *data) +gf_cli_getwd(call_frame_t *frame, xlator_t *this, void *data) { - int ret = -1; - gf1_cli_getwd_req req = {0,}; + int ret = -1; + gf1_cli_getwd_req req = { + 0, + }; - GF_ASSERT (frame); - GF_ASSERT (this); + GF_ASSERT(frame); + GF_ASSERT(this); - if (!frame || !this) - goto out; + if (!frame || !this) + goto out; - ret = cli_cmd_submit (NULL, &req, frame, cli_rpc_prog, - GLUSTER_CLI_GETWD, NULL, - this, gf_cli_getwd_cbk, - (xdrproc_t) xdr_gf1_cli_getwd_req); + ret = cli_cmd_submit(NULL, &req, frame, cli_rpc_prog, GLUSTER_CLI_GETWD, + NULL, this, gf_cli_getwd_cbk, + (xdrproc_t)xdr_gf1_cli_getwd_req); out: - gf_log ("cli", GF_LOG_DEBUG, "Returning %d", ret); + gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret); - return ret; + return ret; } void -cli_print_volume_status_mempool (dict_t *dict, char *prefix) -{ - int ret = -1; - int32_t mempool_count = 0; - char *name = NULL; - int32_t hotcount = 0; - int32_t coldcount = 0; - uint64_t paddedsizeof = 0; - uint64_t alloccount = 0; - int32_t maxalloc = 0; - uint64_t pool_misses = 0; - int32_t maxstdalloc = 0; - char key[1024] = {0,}; - int i = 0; - - GF_ASSERT (dict); - GF_ASSERT (prefix); - - snprintf (key, sizeof (key), "%s.mempool-count",prefix); - ret = dict_get_int32 (dict, key, &mempool_count); +cli_print_volume_status_mempool(dict_t *dict, char *prefix) +{ + int ret = -1; + int32_t mempool_count = 0; + char *name = NULL; + int32_t hotcount = 0; + int32_t coldcount = 0; + uint64_t paddedsizeof = 0; + uint64_t alloccount = 0; + int32_t maxalloc = 0; + uint64_t pool_misses = 0; + int32_t maxstdalloc = 0; + char key[1024] = { + 0, + }; + int i = 0; + + GF_ASSERT(dict); + GF_ASSERT(prefix); + + snprintf(key, sizeof(key), "%s.mempool-count", prefix); + ret = dict_get_int32(dict, key, &mempool_count); + if (ret) + goto out; + + cli_out("Mempool Stats\n-------------"); + cli_out("%-30s %9s %9s %12s %10s %8s %8s %12s", "Name", "HotCount", + "ColdCount", "PaddedSizeof", "AllocCount", "MaxAlloc", "Misses", + "Max-StdAlloc"); + cli_out("%-30s %9s %9s %12s %10s %8s %8s %12s", "----", "--------", + "---------", "------------", "----------", "--------", "--------", + "------------"); + + for (i = 0; i < mempool_count; i++) { + snprintf(key, sizeof(key), "%s.pool%d.name", prefix, i); + ret = dict_get_str(dict, key, &name); if (ret) - goto out; + goto out; - cli_out ("Mempool Stats\n-------------"); - cli_out ("%-30s %9s %9s %12s %10s %8s %8s %12s", "Name", "HotCount", - "ColdCount", "PaddedSizeof", "AllocCount", "MaxAlloc", - "Misses", "Max-StdAlloc"); - cli_out ("%-30s %9s %9s %12s %10s %8s %8s %12s", "----", "--------", - "---------", "------------", "----------", - "--------", "--------", "------------"); - - for (i = 0; i < mempool_count; i++) { - snprintf (key, sizeof (key), "%s.pool%d.name", prefix, i); - ret = dict_get_str (dict, key, &name); - if (ret) - goto out; - - snprintf (key, sizeof (key), "%s.pool%d.hotcount", prefix, i); - ret = dict_get_int32 (dict, key, &hotcount); - if (ret) - goto out; + snprintf(key, sizeof(key), "%s.pool%d.hotcount", prefix, i); + ret = dict_get_int32(dict, key, &hotcount); + if (ret) + goto out; - snprintf (key, sizeof (key), "%s.pool%d.coldcount", prefix, i); - ret = dict_get_int32 (dict, key, &coldcount); - if (ret) - goto out; + snprintf(key, sizeof(key), "%s.pool%d.coldcount", prefix, i); + ret = dict_get_int32(dict, key, &coldcount); + if (ret) + goto out; - snprintf (key, sizeof (key), "%s.pool%d.paddedsizeof", - prefix, i); - ret = dict_get_uint64 (dict, key, &paddedsizeof); - if (ret) - goto out; + snprintf(key, sizeof(key), "%s.pool%d.paddedsizeof", prefix, i); + ret = dict_get_uint64(dict, key, &paddedsizeof); + if (ret) + goto out; - snprintf (key, sizeof (key), "%s.pool%d.alloccount", prefix, i); - ret = dict_get_uint64 (dict, key, &alloccount); - if (ret) - goto out; + snprintf(key, sizeof(key), "%s.pool%d.alloccount", prefix, i); + ret = dict_get_uint64(dict, key, &alloccount); + if (ret) + goto out; - snprintf (key, sizeof (key), "%s.pool%d.max_alloc", prefix, i); - ret = dict_get_int32 (dict, key, &maxalloc); - if (ret) - goto out; + snprintf(key, sizeof(key), "%s.pool%d.max_alloc", prefix, i); + ret = dict_get_int32(dict, key, &maxalloc); + if (ret) + goto out; - snprintf (key, sizeof (key), "%s.pool%d.max-stdalloc", prefix, i); - ret = dict_get_int32 (dict, key, &maxstdalloc); - if (ret) - goto out; + snprintf(key, sizeof(key), "%s.pool%d.max-stdalloc", prefix, i); + ret = dict_get_int32(dict, key, &maxstdalloc); + if (ret) + goto out; - snprintf (key, sizeof (key), "%s.pool%d.pool-misses", prefix, i); - ret = dict_get_uint64 (dict, key, &pool_misses); - if (ret) - goto out; + snprintf(key, sizeof(key), "%s.pool%d.pool-misses", prefix, i); + ret = dict_get_uint64(dict, key, &pool_misses); + if (ret) + goto out; - cli_out ("%-30s %9d %9d %12"PRIu64" %10"PRIu64" %8d %8"PRIu64 - " %12d", name, hotcount, coldcount, paddedsizeof, - alloccount, maxalloc, pool_misses, maxstdalloc); - } + cli_out("%-30s %9d %9d %12" PRIu64 " %10" PRIu64 " %8d %8" PRIu64 + " %12d", + name, hotcount, coldcount, paddedsizeof, alloccount, maxalloc, + pool_misses, maxstdalloc); + } out: - return; - + return; } void -cli_print_volume_status_mem (dict_t *dict, gf_boolean_t notbrick) -{ - int ret = -1; - char *volname = NULL; - char *hostname = NULL; - char *path = NULL; - int online = -1; - char key[1024] = {0,}; - int brick_index_max = -1; - int other_count = 0; - int index_max = 0; - int val = 0; - int i = 0; - - GF_ASSERT (dict); - - ret = dict_get_str (dict, "volname", &volname); +cli_print_volume_status_mem(dict_t *dict, gf_boolean_t notbrick) +{ + int ret = -1; + char *volname = NULL; + char *hostname = NULL; + char *path = NULL; + int online = -1; + char key[1024] = { + 0, + }; + int brick_index_max = -1; + int other_count = 0; + int index_max = 0; + int val = 0; + int i = 0; + + GF_ASSERT(dict); + + ret = dict_get_str(dict, "volname", &volname); + if (ret) + goto out; + cli_out("Memory status for volume : %s", volname); + + ret = dict_get_int32(dict, "brick-index-max", &brick_index_max); + if (ret) + goto out; + ret = dict_get_int32(dict, "other-count", &other_count); + if (ret) + goto out; + + index_max = brick_index_max + other_count; + + for (i = 0; i <= index_max; i++) { + cli_out("----------------------------------------------"); + + snprintf(key, sizeof(key), "brick%d.hostname", i); + ret = dict_get_str(dict, key, &hostname); if (ret) - goto out; - cli_out ("Memory status for volume : %s", volname); - - ret = dict_get_int32 (dict, "brick-index-max", &brick_index_max); + continue; + snprintf(key, sizeof(key), "brick%d.path", i); + ret = dict_get_str(dict, key, &path); if (ret) - goto out; - ret = dict_get_int32 (dict, "other-count", &other_count); - if (ret) - goto out; - - index_max = brick_index_max + other_count; - - for (i = 0; i <= index_max; i++) { - cli_out ("----------------------------------------------"); - - snprintf (key, sizeof (key), "brick%d.hostname", i); - ret = dict_get_str (dict, key, &hostname); - if (ret) - continue; - snprintf (key, sizeof (key), "brick%d.path", i); - ret = dict_get_str (dict, key, &path); - if (ret) - continue; - if (notbrick) - cli_out ("%s : %s", hostname, path); - else - cli_out ("Brick : %s:%s", hostname, path); - - snprintf (key, sizeof (key), "brick%d.status", i); - ret = dict_get_int32 (dict, key, &online); - if (ret) - goto out; - if (!online) { - if (notbrick) - cli_out ("%s is offline", hostname); - else - cli_out ("Brick is offline"); - continue; - } - - cli_out ("Mallinfo\n--------"); - - snprintf (key, sizeof (key), "brick%d.mallinfo.arena", i); - ret = dict_get_int32 (dict, key, &val); - if (ret) - goto out; - cli_out ("%-8s : %d","Arena", val); - - snprintf (key, sizeof (key), "brick%d.mallinfo.ordblks", i); - ret = dict_get_int32 (dict, key, &val); - if(ret) - goto out; - cli_out ("%-8s : %d","Ordblks", val); - - snprintf (key, sizeof (key), "brick%d.mallinfo.smblks", i); - ret = dict_get_int32 (dict, key, &val); - if(ret) - goto out; - cli_out ("%-8s : %d","Smblks", val); - - snprintf (key, sizeof (key), "brick%d.mallinfo.hblks", i); - ret = dict_get_int32 (dict, key, &val); - if(ret) - goto out; - cli_out ("%-8s : %d", "Hblks", val); - - snprintf (key, sizeof (key), "brick%d.mallinfo.hblkhd", i); - ret = dict_get_int32 (dict, key, &val); - if (ret) - goto out; - cli_out ("%-8s : %d", "Hblkhd", val); - - snprintf (key, sizeof (key), "brick%d.mallinfo.usmblks", i); - ret = dict_get_int32 (dict, key, &val); - if (ret) - goto out; - cli_out ("%-8s : %d", "Usmblks", val); - - snprintf (key, sizeof (key), "brick%d.mallinfo.fsmblks", i); - ret = dict_get_int32 (dict, key, &val); - if (ret) - goto out; - cli_out ("%-8s : %d", "Fsmblks", val); - - snprintf (key, sizeof (key), "brick%d.mallinfo.uordblks", i); - ret = dict_get_int32 (dict, key, &val); - if (ret) - goto out; - cli_out ("%-8s : %d", "Uordblks", val); - - snprintf (key, sizeof (key), "brick%d.mallinfo.fordblks", i); - ret = dict_get_int32 (dict, key, &val); - if (ret) - goto out; - cli_out ("%-8s : %d", "Fordblks", val); - - snprintf (key, sizeof (key), "brick%d.mallinfo.keepcost", i); - ret = dict_get_int32 (dict, key, &val); - if (ret) - goto out; - cli_out ("%-8s : %d", "Keepcost", val); + continue; + if (notbrick) + cli_out("%s : %s", hostname, path); + else + cli_out("Brick : %s:%s", hostname, path); - cli_out (" "); - snprintf (key, sizeof (key), "brick%d", i); - cli_print_volume_status_mempool (dict, key); + snprintf(key, sizeof(key), "brick%d.status", i); + ret = dict_get_int32(dict, key, &online); + if (ret) + goto out; + if (!online) { + if (notbrick) + cli_out("%s is offline", hostname); + else + cli_out("Brick is offline"); + continue; } -out: - cli_out ("----------------------------------------------\n"); - return; -} -void -cli_print_volume_status_client_list (dict_t *dict, gf_boolean_t notbrick) -{ - int ret = -1; - char *volname = NULL; - int client_count = 0; - int current_count = 0; - char key[1024] = {0,}; - int i = 0; - int total = 0; - char *name = NULL; - gf_boolean_t is_fuse_done = _gf_false; - gf_boolean_t is_gfapi_done = _gf_false; - gf_boolean_t is_tierd_done = _gf_false; - gf_boolean_t is_rebalance_done = _gf_false; - gf_boolean_t is_glustershd_done = _gf_false; - gf_boolean_t is_quotad_done = _gf_false; - gf_boolean_t is_snapd_done = _gf_false; - - GF_ASSERT (dict); - - ret = dict_get_str (dict, "volname", &volname); - if (ret) - goto out; - cli_out ("Client connections for volume %s", volname); + cli_out("Mallinfo\n--------"); - ret = dict_get_int32 (dict, "client-count", &client_count); + snprintf(key, sizeof(key), "brick%d.mallinfo.arena", i); + ret = dict_get_int32(dict, key, &val); if (ret) - goto out; + goto out; + cli_out("%-8s : %d", "Arena", val); - cli_out ("%-48s %15s", "Name", "count"); - cli_out ("%-48s %15s", "-----", "------"); - for (i = 0; i < client_count; i++) { - name = NULL; - snprintf (key, sizeof (key), - "client%d.name", i); - ret = dict_get_str (dict, key, &name); - - if (!strncmp (name, "fuse", 4)) { - if (!is_fuse_done) { - is_fuse_done = _gf_true; - ret = dict_get_int32 (dict, - "fuse-count", - ¤t_count); - if (ret) - goto out; - total = total + current_count; - goto print; - } - continue; - } else if (!strncmp (name, "gfapi", 5)) { - if (!is_gfapi_done) { - is_gfapi_done = _gf_true; - ret = dict_get_int32 (dict, - "gfapi-count", - ¤t_count); - if (ret) - goto out; - total = total + current_count; - goto print; - } - continue; - } else if (!strcmp(name, "tierd")) { - if (!is_tierd_done) { - is_tierd_done = _gf_true; - ret = dict_get_int32 (dict, - "tierd-count", - ¤t_count); - if (ret) - goto out; - total = total + current_count; - goto print; - } - continue; - } else if (!strcmp(name, "rebalance")) { - if (!is_rebalance_done) { - is_rebalance_done = _gf_true; - ret = dict_get_int32 (dict, - "rebalance-count", - ¤t_count); - if (ret) - goto out; - total = total + current_count; - goto print; - } - continue; - } else if (!strcmp(name, "glustershd")) { - if (!is_glustershd_done) { - is_glustershd_done = _gf_true; - ret = dict_get_int32 (dict, - "glustershd-count", - ¤t_count); - if (ret) - goto out; - total = total + current_count; - goto print; - } - continue; - } else if (!strcmp(name, "quotad")) { - if (!is_quotad_done) { - is_quotad_done = _gf_true; - ret = dict_get_int32 (dict, - "quotad-count", - ¤t_count); - if (ret) - goto out; - total = total + current_count; - goto print; - } - continue; - } else if (!strcmp(name, "snapd")) { - if (!is_snapd_done) { - is_snapd_done = _gf_true; - ret = dict_get_int32 (dict, - "snapd-count", - ¤t_count); - if (ret) - goto out; - total = total + current_count; - goto print; - } - continue; - } - -print: - cli_out ("%-48s %15d", name, current_count); - - } -out: - cli_out ("\ntotal clients for volume %s : %d ", volname, total); - cli_out ("-----------------------------------------------------------------\n"); - return; -} - -void -cli_print_volume_status_clients (dict_t *dict, gf_boolean_t notbrick) -{ - int ret = -1; - char *volname = NULL; - int brick_index_max = -1; - int other_count = 0; - int index_max = 0; - char *hostname = NULL; - char *path = NULL; - int online = -1; - int client_count = 0; - char *clientname = NULL; - uint64_t bytesread = 0; - uint64_t byteswrite = 0; - uint32_t opversion = 0; - char key[1024] = {0,}; - int i = 0; - int j = 0; - - GF_ASSERT (dict); - - ret = dict_get_str (dict, "volname", &volname); + snprintf(key, sizeof(key), "brick%d.mallinfo.ordblks", i); + ret = dict_get_int32(dict, key, &val); if (ret) - goto out; - cli_out ("Client connections for volume %s", volname); + goto out; + cli_out("%-8s : %d", "Ordblks", val); - ret = dict_get_int32 (dict, "brick-index-max", &brick_index_max); + snprintf(key, sizeof(key), "brick%d.mallinfo.smblks", i); + ret = dict_get_int32(dict, key, &val); if (ret) - goto out; - ret = dict_get_int32 (dict, "other-count", &other_count); + goto out; + cli_out("%-8s : %d", "Smblks", val); + + snprintf(key, sizeof(key), "brick%d.mallinfo.hblks", i); + ret = dict_get_int32(dict, key, &val); if (ret) - goto out; + goto out; + cli_out("%-8s : %d", "Hblks", val); - index_max = brick_index_max + other_count; + snprintf(key, sizeof(key), "brick%d.mallinfo.hblkhd", i); + ret = dict_get_int32(dict, key, &val); + if (ret) + goto out; + cli_out("%-8s : %d", "Hblkhd", val); - for (i = 0; i <= index_max; i++) { - hostname = NULL; - path = NULL; - online = -1; - client_count = 0; - clientname = NULL; - bytesread = 0; - byteswrite = 0; + snprintf(key, sizeof(key), "brick%d.mallinfo.usmblks", i); + ret = dict_get_int32(dict, key, &val); + if (ret) + goto out; + cli_out("%-8s : %d", "Usmblks", val); - cli_out ("----------------------------------------------"); + snprintf(key, sizeof(key), "brick%d.mallinfo.fsmblks", i); + ret = dict_get_int32(dict, key, &val); + if (ret) + goto out; + cli_out("%-8s : %d", "Fsmblks", val); - snprintf (key, sizeof (key), "brick%d.hostname", i); - ret = dict_get_str (dict, key, &hostname); + snprintf(key, sizeof(key), "brick%d.mallinfo.uordblks", i); + ret = dict_get_int32(dict, key, &val); + if (ret) + goto out; + cli_out("%-8s : %d", "Uordblks", val); - snprintf (key, sizeof (key), "brick%d.path", i); - ret = dict_get_str (dict, key, &path); + snprintf(key, sizeof(key), "brick%d.mallinfo.fordblks", i); + ret = dict_get_int32(dict, key, &val); + if (ret) + goto out; + cli_out("%-8s : %d", "Fordblks", val); - if (hostname && path) { - if (notbrick) - cli_out ("%s : %s", hostname, path); - else - cli_out ("Brick : %s:%s", hostname, path); - } - snprintf (key, sizeof (key), "brick%d.status", i); - ret = dict_get_int32 (dict, key, &online); - if (!online) { - if (notbrick) - cli_out ("%s is offline", hostname); - else - cli_out ("Brick is offline"); - continue; - } + snprintf(key, sizeof(key), "brick%d.mallinfo.keepcost", i); + ret = dict_get_int32(dict, key, &val); + if (ret) + goto out; + cli_out("%-8s : %d", "Keepcost", val); - snprintf (key, sizeof (key), "brick%d.clientcount", i); - ret = dict_get_int32 (dict, key, &client_count); - - if (hostname && path) - cli_out ("Clients connected : %d", client_count); - if (client_count == 0) - continue; - - cli_out ("%-48s %15s %15s %15s", "Hostname", "BytesRead", - "BytesWritten", "OpVersion"); - cli_out ("%-48s %15s %15s %15s", "--------", "---------", - "------------", "---------"); - for (j =0; j < client_count; j++) { - snprintf (key, sizeof (key), - "brick%d.client%d.hostname", i, j); - ret = dict_get_str (dict, key, &clientname); - - snprintf (key, sizeof (key), - "brick%d.client%d.bytesread", i, j); - ret = dict_get_uint64 (dict, key, &bytesread); - - snprintf (key, sizeof (key), - "brick%d.client%d.byteswrite", i, j); - ret = dict_get_uint64 (dict, key, &byteswrite); - - snprintf (key, sizeof (key), - "brick%d.client%d.opversion", i, j); - ret = dict_get_uint32 (dict, key, &opversion); - - cli_out ("%-48s %15"PRIu64" %15"PRIu64" %15"PRIu32, - clientname, bytesread, byteswrite, - opversion); - } - } + cli_out(" "); + snprintf(key, sizeof(key), "brick%d", i); + cli_print_volume_status_mempool(dict, key); + } out: - cli_out ("----------------------------------------------\n"); - return; + cli_out("----------------------------------------------\n"); + return; } void -cli_print_volume_status_inode_entry (dict_t *dict, char *prefix) -{ - int ret = -1; - char key[1024] = {0,}; - char *gfid = NULL; - uint64_t nlookup = 0; - uint32_t ref = 0; - int ia_type = 0; - char inode_type; - - GF_ASSERT (dict); - GF_ASSERT (prefix); - - snprintf (key, sizeof (key), "%s.gfid", prefix); - ret = dict_get_str (dict, key, &gfid); - if (ret) - goto out; - - snprintf (key, sizeof (key), "%s.nlookup", prefix); - ret = dict_get_uint64 (dict, key, &nlookup); - if (ret) - goto out; - - snprintf (key, sizeof (key), "%s.ref", prefix); - ret = dict_get_uint32 (dict, key, &ref); - if (ret) - goto out; +cli_print_volume_status_client_list(dict_t *dict, gf_boolean_t notbrick) +{ + int ret = -1; + char *volname = NULL; + int client_count = 0; + int current_count = 0; + char key[1024] = { + 0, + }; + int i = 0; + int total = 0; + char *name = NULL; + gf_boolean_t is_fuse_done = _gf_false; + gf_boolean_t is_gfapi_done = _gf_false; + gf_boolean_t is_tierd_done = _gf_false; + gf_boolean_t is_rebalance_done = _gf_false; + gf_boolean_t is_glustershd_done = _gf_false; + gf_boolean_t is_quotad_done = _gf_false; + gf_boolean_t is_snapd_done = _gf_false; + + GF_ASSERT(dict); + + ret = dict_get_str(dict, "volname", &volname); + if (ret) + goto out; + cli_out("Client connections for volume %s", volname); + + ret = dict_get_int32(dict, "client-count", &client_count); + if (ret) + goto out; + + cli_out("%-48s %15s", "Name", "count"); + cli_out("%-48s %15s", "-----", "------"); + for (i = 0; i < client_count; i++) { + name = NULL; + snprintf(key, sizeof(key), "client%d.name", i); + ret = dict_get_str(dict, key, &name); + + if (!strncmp(name, "fuse", 4)) { + if (!is_fuse_done) { + is_fuse_done = _gf_true; + ret = dict_get_int32(dict, "fuse-count", ¤t_count); + if (ret) + goto out; + total = total + current_count; + goto print; + } + continue; + } else if (!strncmp(name, "gfapi", 5)) { + if (!is_gfapi_done) { + is_gfapi_done = _gf_true; + ret = dict_get_int32(dict, "gfapi-count", ¤t_count); + if (ret) + goto out; + total = total + current_count; + goto print; + } + continue; + } else if (!strcmp(name, "tierd")) { + if (!is_tierd_done) { + is_tierd_done = _gf_true; + ret = dict_get_int32(dict, "tierd-count", ¤t_count); + if (ret) + goto out; + total = total + current_count; + goto print; + } + continue; + } else if (!strcmp(name, "rebalance")) { + if (!is_rebalance_done) { + is_rebalance_done = _gf_true; + ret = dict_get_int32(dict, "rebalance-count", ¤t_count); + if (ret) + goto out; + total = total + current_count; + goto print; + } + continue; + } else if (!strcmp(name, "glustershd")) { + if (!is_glustershd_done) { + is_glustershd_done = _gf_true; + ret = dict_get_int32(dict, "glustershd-count", ¤t_count); + if (ret) + goto out; + total = total + current_count; + goto print; + } + continue; + } else if (!strcmp(name, "quotad")) { + if (!is_quotad_done) { + is_quotad_done = _gf_true; + ret = dict_get_int32(dict, "quotad-count", ¤t_count); + if (ret) + goto out; + total = total + current_count; + goto print; + } + continue; + } else if (!strcmp(name, "snapd")) { + if (!is_snapd_done) { + is_snapd_done = _gf_true; + ret = dict_get_int32(dict, "snapd-count", ¤t_count); + if (ret) + goto out; + total = total + current_count; + goto print; + } + continue; + } + + print: + cli_out("%-48s %15d", name, current_count); + } +out: + cli_out("\ntotal clients for volume %s : %d ", volname, total); + cli_out( + "-----------------------------------------------------------------\n"); + return; +} - snprintf (key, sizeof (key), "%s.ia_type", prefix); - ret = dict_get_int32 (dict, key, &ia_type); - if (ret) - goto out; +void +cli_print_volume_status_clients(dict_t *dict, gf_boolean_t notbrick) +{ + int ret = -1; + char *volname = NULL; + int brick_index_max = -1; + int other_count = 0; + int index_max = 0; + char *hostname = NULL; + char *path = NULL; + int online = -1; + int client_count = 0; + char *clientname = NULL; + uint64_t bytesread = 0; + uint64_t byteswrite = 0; + uint32_t opversion = 0; + char key[1024] = { + 0, + }; + int i = 0; + int j = 0; + + GF_ASSERT(dict); + + ret = dict_get_str(dict, "volname", &volname); + if (ret) + goto out; + cli_out("Client connections for volume %s", volname); + + ret = dict_get_int32(dict, "brick-index-max", &brick_index_max); + if (ret) + goto out; + ret = dict_get_int32(dict, "other-count", &other_count); + if (ret) + goto out; + + index_max = brick_index_max + other_count; + + for (i = 0; i <= index_max; i++) { + hostname = NULL; + path = NULL; + online = -1; + client_count = 0; + clientname = NULL; + bytesread = 0; + byteswrite = 0; + + cli_out("----------------------------------------------"); + + snprintf(key, sizeof(key), "brick%d.hostname", i); + ret = dict_get_str(dict, key, &hostname); + + snprintf(key, sizeof(key), "brick%d.path", i); + ret = dict_get_str(dict, key, &path); + + if (hostname && path) { + if (notbrick) + cli_out("%s : %s", hostname, path); + else + cli_out("Brick : %s:%s", hostname, path); + } + snprintf(key, sizeof(key), "brick%d.status", i); + ret = dict_get_int32(dict, key, &online); + if (!online) { + if (notbrick) + cli_out("%s is offline", hostname); + else + cli_out("Brick is offline"); + continue; + } + + snprintf(key, sizeof(key), "brick%d.clientcount", i); + ret = dict_get_int32(dict, key, &client_count); + + if (hostname && path) + cli_out("Clients connected : %d", client_count); + if (client_count == 0) + continue; + + cli_out("%-48s %15s %15s %15s", "Hostname", "BytesRead", "BytesWritten", + "OpVersion"); + cli_out("%-48s %15s %15s %15s", "--------", "---------", "------------", + "---------"); + for (j = 0; j < client_count; j++) { + snprintf(key, sizeof(key), "brick%d.client%d.hostname", i, j); + ret = dict_get_str(dict, key, &clientname); + + snprintf(key, sizeof(key), "brick%d.client%d.bytesread", i, j); + ret = dict_get_uint64(dict, key, &bytesread); + + snprintf(key, sizeof(key), "brick%d.client%d.byteswrite", i, j); + ret = dict_get_uint64(dict, key, &byteswrite); + + snprintf(key, sizeof(key), "brick%d.client%d.opversion", i, j); + ret = dict_get_uint32(dict, key, &opversion); + + cli_out("%-48s %15" PRIu64 " %15" PRIu64 " %15" PRIu32, clientname, + bytesread, byteswrite, opversion); + } + } +out: + cli_out("----------------------------------------------\n"); + return; +} - switch (ia_type) { +void +cli_print_volume_status_inode_entry(dict_t *dict, char *prefix) +{ + int ret = -1; + char key[1024] = { + 0, + }; + char *gfid = NULL; + uint64_t nlookup = 0; + uint32_t ref = 0; + int ia_type = 0; + char inode_type; + + GF_ASSERT(dict); + GF_ASSERT(prefix); + + snprintf(key, sizeof(key), "%s.gfid", prefix); + ret = dict_get_str(dict, key, &gfid); + if (ret) + goto out; + + snprintf(key, sizeof(key), "%s.nlookup", prefix); + ret = dict_get_uint64(dict, key, &nlookup); + if (ret) + goto out; + + snprintf(key, sizeof(key), "%s.ref", prefix); + ret = dict_get_uint32(dict, key, &ref); + if (ret) + goto out; + + snprintf(key, sizeof(key), "%s.ia_type", prefix); + ret = dict_get_int32(dict, key, &ia_type); + if (ret) + goto out; + + switch (ia_type) { case IA_IFREG: - inode_type = 'R'; - break; + inode_type = 'R'; + break; case IA_IFDIR: - inode_type = 'D'; - break; + inode_type = 'D'; + break; case IA_IFLNK: - inode_type = 'L'; - break; + inode_type = 'L'; + break; case IA_IFBLK: - inode_type = 'B'; - break; + inode_type = 'B'; + break; case IA_IFCHR: - inode_type = 'C'; - break; + inode_type = 'C'; + break; case IA_IFIFO: - inode_type = 'F'; - break; + inode_type = 'F'; + break; case IA_IFSOCK: - inode_type = 'S'; - break; + inode_type = 'S'; + break; default: - inode_type = 'I'; - break; - } + inode_type = 'I'; + break; + } - cli_out ("%-40s %14"PRIu64" %14"PRIu32" %9c", - gfid, nlookup, ref, inode_type); + cli_out("%-40s %14" PRIu64 " %14" PRIu32 " %9c", gfid, nlookup, ref, + inode_type); out: - return; - + return; } void -cli_print_volume_status_itables (dict_t *dict, char *prefix) -{ - int ret = -1; - char key[1024] = {0,}; - uint32_t active_size = 0; - uint32_t lru_size = 0; - uint32_t purge_size = 0; - int i =0; - - GF_ASSERT (dict); - GF_ASSERT (prefix); +cli_print_volume_status_itables(dict_t *dict, char *prefix) +{ + int ret = -1; + char key[1024] = { + 0, + }; + uint32_t active_size = 0; + uint32_t lru_size = 0; + uint32_t purge_size = 0; + int i = 0; + + GF_ASSERT(dict); + GF_ASSERT(prefix); + + snprintf(key, sizeof(key), "%s.active_size", prefix); + ret = dict_get_uint32(dict, key, &active_size); + if (ret) + goto out; + if (active_size != 0) { + cli_out("Active inodes:"); + cli_out("%-40s %14s %14s %9s", "GFID", "Lookups", "Ref", "IA type"); + cli_out("%-40s %14s %14s %9s", "----", "-------", "---", "-------"); + } + for (i = 0; i < active_size; i++) { + snprintf(key, sizeof(key), "%s.active%d", prefix, i); + cli_print_volume_status_inode_entry(dict, key); + } + cli_out(" "); + + snprintf(key, sizeof(key), "%s.lru_size", prefix); + ret = dict_get_uint32(dict, key, &lru_size); + if (ret) + goto out; + if (lru_size != 0) { + cli_out("LRU inodes:"); + cli_out("%-40s %14s %14s %9s", "GFID", "Lookups", "Ref", "IA type"); + cli_out("%-40s %14s %14s %9s", "----", "-------", "---", "-------"); + } + for (i = 0; i < lru_size; i++) { + snprintf(key, sizeof(key), "%s.lru%d", prefix, i); + cli_print_volume_status_inode_entry(dict, key); + } + cli_out(" "); + + snprintf(key, sizeof(key), "%s.purge_size", prefix); + ret = dict_get_uint32(dict, key, &purge_size); + if (ret) + goto out; + if (purge_size != 0) { + cli_out("Purged inodes:"); + cli_out("%-40s %14s %14s %9s", "GFID", "Lookups", "Ref", "IA type"); + cli_out("%-40s %14s %14s %9s", "----", "-------", "---", "-------"); + } + for (i = 0; i < purge_size; i++) { + snprintf(key, sizeof(key), "%s.purge%d", prefix, i); + cli_print_volume_status_inode_entry(dict, key); + } + +out: + return; +} - snprintf (key, sizeof (key), "%s.active_size", prefix); - ret = dict_get_uint32 (dict, key, &active_size); +void +cli_print_volume_status_inode(dict_t *dict, gf_boolean_t notbrick) +{ + int ret = -1; + char *volname = NULL; + int brick_index_max = -1; + int other_count = 0; + int index_max = 0; + char *hostname = NULL; + char *path = NULL; + int online = -1; + int conn_count = 0; + char key[1024] = { + 0, + }; + int i = 0; + int j = 0; + + GF_ASSERT(dict); + + ret = dict_get_str(dict, "volname", &volname); + if (ret) + goto out; + cli_out("Inode tables for volume %s", volname); + + ret = dict_get_int32(dict, "brick-index-max", &brick_index_max); + if (ret) + goto out; + ret = dict_get_int32(dict, "other-count", &other_count); + if (ret) + goto out; + + index_max = brick_index_max + other_count; + + for (i = 0; i <= index_max; i++) { + cli_out("----------------------------------------------"); + + snprintf(key, sizeof(key), "brick%d.hostname", i); + ret = dict_get_str(dict, key, &hostname); if (ret) - goto out; - if (active_size != 0) { - cli_out ("Active inodes:"); - cli_out ("%-40s %14s %14s %9s", "GFID", "Lookups", "Ref", - "IA type"); - cli_out ("%-40s %14s %14s %9s", "----", "-------", "---", - "-------"); - } - for (i = 0; i < active_size; i++) { - snprintf (key, sizeof (key), "%s.active%d", prefix, i); - cli_print_volume_status_inode_entry (dict, key); - } - cli_out (" "); + goto out; + snprintf(key, sizeof(key), "brick%d.path", i); + ret = dict_get_str(dict, key, &path); + if (ret) + goto out; + if (notbrick) + cli_out("%s : %s", hostname, path); + else + cli_out("Brick : %s:%s", hostname, path); - snprintf (key, sizeof (key), "%s.lru_size", prefix); - ret = dict_get_uint32 (dict, key, &lru_size); + snprintf(key, sizeof(key), "brick%d.status", i); + ret = dict_get_int32(dict, key, &online); if (ret) - goto out; - if (lru_size != 0) { - cli_out ("LRU inodes:"); - cli_out ("%-40s %14s %14s %9s", "GFID", "Lookups", "Ref", - "IA type"); - cli_out ("%-40s %14s %14s %9s", "----", "-------", "---", - "-------"); - } - for (i = 0; i < lru_size; i++) { - snprintf (key, sizeof (key), "%s.lru%d", prefix, i); - cli_print_volume_status_inode_entry (dict, key); + goto out; + if (!online) { + if (notbrick) + cli_out("%s is offline", hostname); + else + cli_out("Brick is offline"); + continue; } - cli_out (" "); - snprintf (key, sizeof (key), "%s.purge_size", prefix); - ret = dict_get_uint32 (dict, key, &purge_size); + snprintf(key, sizeof(key), "brick%d.conncount", i); + ret = dict_get_int32(dict, key, &conn_count); if (ret) - goto out; - if (purge_size != 0) { - cli_out ("Purged inodes:"); - cli_out ("%-40s %14s %14s %9s", "GFID", "Lookups", "Ref", - "IA type"); - cli_out ("%-40s %14s %14s %9s", "----", "-------", "---", - "-------"); - } - for (i = 0; i < purge_size; i++) { - snprintf (key, sizeof (key), "%s.purge%d", prefix, i); - cli_print_volume_status_inode_entry (dict, key); - } + goto out; + + for (j = 0; j < conn_count; j++) { + if (conn_count > 1) + cli_out("Connection %d:", j + 1); + snprintf(key, sizeof(key), "brick%d.conn%d.itable", i, j); + cli_print_volume_status_itables(dict, key); + cli_out(" "); + } + } out: - return; + cli_out("----------------------------------------------"); + return; } void -cli_print_volume_status_inode (dict_t *dict, gf_boolean_t notbrick) -{ - int ret = -1; - char *volname = NULL; - int brick_index_max = -1; - int other_count = 0; - int index_max = 0; - char *hostname = NULL; - char *path = NULL; - int online = -1; - int conn_count = 0; - char key[1024] = {0,}; - int i = 0; - int j = 0; - - GF_ASSERT (dict); - - ret = dict_get_str (dict, "volname", &volname); +cli_print_volume_status_fdtable(dict_t *dict, char *prefix) +{ + int ret = -1; + char key[1024] = { + 0, + }; + int refcount = 0; + uint32_t maxfds = 0; + int firstfree = 0; + int openfds = 0; + int fd_pid = 0; + int fd_refcount = 0; + int fd_flags = 0; + int i = 0; + + GF_ASSERT(dict); + GF_ASSERT(prefix); + + snprintf(key, sizeof(key), "%s.refcount", prefix); + ret = dict_get_int32(dict, key, &refcount); + if (ret) + goto out; + + snprintf(key, sizeof(key), "%s.maxfds", prefix); + ret = dict_get_uint32(dict, key, &maxfds); + if (ret) + goto out; + + snprintf(key, sizeof(key), "%s.firstfree", prefix); + ret = dict_get_int32(dict, key, &firstfree); + if (ret) + goto out; + + cli_out("RefCount = %d MaxFDs = %d FirstFree = %d", refcount, maxfds, + firstfree); + + snprintf(key, sizeof(key), "%s.openfds", prefix); + ret = dict_get_int32(dict, key, &openfds); + if (ret) + goto out; + if (0 == openfds) { + cli_err("No open fds"); + goto out; + } + + cli_out("%-19s %-19s %-19s %-19s", "FD Entry", "PID", "RefCount", "Flags"); + cli_out("%-19s %-19s %-19s %-19s", "--------", "---", "--------", "-----"); + + for (i = 0; i < maxfds; i++) { + snprintf(key, sizeof(key), "%s.fdentry%d.pid", prefix, i); + ret = dict_get_int32(dict, key, &fd_pid); if (ret) - goto out; - cli_out ("Inode tables for volume %s", volname); + continue; - ret = dict_get_int32 (dict, "brick-index-max", &brick_index_max); + snprintf(key, sizeof(key), "%s.fdentry%d.refcount", prefix, i); + ret = dict_get_int32(dict, key, &fd_refcount); if (ret) - goto out; - ret = dict_get_int32 (dict, "other-count", &other_count); - if (ret) - goto out; - - index_max = brick_index_max + other_count; + continue; - for ( i = 0; i <= index_max; i++) { - cli_out ("----------------------------------------------"); - - snprintf (key, sizeof (key), "brick%d.hostname", i); - ret = dict_get_str (dict, key, &hostname); - if (ret) - goto out; - snprintf (key, sizeof (key), "brick%d.path", i); - ret = dict_get_str (dict, key, &path); - if (ret) - goto out; - if (notbrick) - cli_out ("%s : %s", hostname, path); - else - cli_out ("Brick : %s:%s", hostname, path); - - snprintf (key, sizeof (key), "brick%d.status", i); - ret = dict_get_int32 (dict, key, &online); - if (ret) - goto out; - if (!online) { - if (notbrick) - cli_out ("%s is offline", hostname); - else - cli_out ("Brick is offline"); - continue; - } - - snprintf (key, sizeof (key), "brick%d.conncount", i); - ret = dict_get_int32 (dict, key, &conn_count); - if (ret) - goto out; + snprintf(key, sizeof(key), "%s.fdentry%d.flags", prefix, i); + ret = dict_get_int32(dict, key, &fd_flags); + if (ret) + continue; - for (j = 0; j < conn_count; j++) { - if (conn_count > 1) - cli_out ("Connection %d:", j+1); + cli_out("%-19d %-19d %-19d %-19d", i, fd_pid, fd_refcount, fd_flags); + } - snprintf (key, sizeof (key), "brick%d.conn%d.itable", - i, j); - cli_print_volume_status_itables (dict, key); - cli_out (" "); - } - } out: - cli_out ("----------------------------------------------"); - return; + return; } void -cli_print_volume_status_fdtable (dict_t *dict, char *prefix) -{ - int ret = -1; - char key[1024] = {0,}; - int refcount = 0; - uint32_t maxfds = 0; - int firstfree = 0; - int openfds = 0; - int fd_pid = 0; - int fd_refcount = 0; - int fd_flags = 0; - int i = 0; - - GF_ASSERT (dict); - GF_ASSERT (prefix); - - snprintf (key, sizeof (key), "%s.refcount", prefix); - ret = dict_get_int32 (dict, key, &refcount); +cli_print_volume_status_fd(dict_t *dict, gf_boolean_t notbrick) +{ + int ret = -1; + char *volname = NULL; + int brick_index_max = -1; + int other_count = 0; + int index_max = 0; + char *hostname = NULL; + char *path = NULL; + int online = -1; + int conn_count = 0; + char key[1024] = { + 0, + }; + int i = 0; + int j = 0; + + GF_ASSERT(dict); + + ret = dict_get_str(dict, "volname", &volname); + if (ret) + goto out; + cli_out("FD tables for volume %s", volname); + + ret = dict_get_int32(dict, "brick-index-max", &brick_index_max); + if (ret) + goto out; + ret = dict_get_int32(dict, "other-count", &other_count); + if (ret) + goto out; + + index_max = brick_index_max + other_count; + + for (i = 0; i <= index_max; i++) { + cli_out("----------------------------------------------"); + + snprintf(key, sizeof(key), "brick%d.hostname", i); + ret = dict_get_str(dict, key, &hostname); if (ret) - goto out; - - snprintf (key, sizeof (key), "%s.maxfds", prefix); - ret = dict_get_uint32 (dict, key, &maxfds); + goto out; + snprintf(key, sizeof(key), "brick%d.path", i); + ret = dict_get_str(dict, key, &path); if (ret) - goto out; - - snprintf (key, sizeof (key), "%s.firstfree", prefix); - ret = dict_get_int32 (dict, key, &firstfree); - if (ret) - goto out; + goto out; - cli_out ("RefCount = %d MaxFDs = %d FirstFree = %d", - refcount, maxfds, firstfree); + if (notbrick) + cli_out("%s : %s", hostname, path); + else + cli_out("Brick : %s:%s", hostname, path); - snprintf (key, sizeof (key), "%s.openfds", prefix); - ret = dict_get_int32 (dict, key, &openfds); + snprintf(key, sizeof(key), "brick%d.status", i); + ret = dict_get_int32(dict, key, &online); if (ret) - goto out; - if (0 == openfds) { - cli_err ("No open fds"); - goto out; + goto out; + if (!online) { + if (notbrick) + cli_out("%s is offline", hostname); + else + cli_out("Brick is offline"); + continue; } - cli_out ("%-19s %-19s %-19s %-19s", "FD Entry", "PID", - "RefCount", "Flags"); - cli_out ("%-19s %-19s %-19s %-19s", "--------", "---", - "--------", "-----"); - - for (i = 0; i < maxfds ; i++) { - snprintf (key, sizeof (key), "%s.fdentry%d.pid", prefix, i); - ret = dict_get_int32 (dict, key, &fd_pid); - if (ret) - continue; - - snprintf (key, sizeof (key), "%s.fdentry%d.refcount", - prefix, i); - ret = dict_get_int32 (dict, key, &fd_refcount); - if (ret) - continue; + snprintf(key, sizeof(key), "brick%d.conncount", i); + ret = dict_get_int32(dict, key, &conn_count); + if (ret) + goto out; - snprintf (key, sizeof (key), "%s.fdentry%d.flags", prefix, i); - ret = dict_get_int32 (dict, key, &fd_flags); - if (ret) - continue; + for (j = 0; j < conn_count; j++) { + cli_out("Connection %d:", j + 1); - cli_out ("%-19d %-19d %-19d %-19d", i, fd_pid, fd_refcount, - fd_flags); + snprintf(key, sizeof(key), "brick%d.conn%d.fdtable", i, j); + cli_print_volume_status_fdtable(dict, key); + cli_out(" "); } - + } out: - return; + cli_out("----------------------------------------------"); + return; } void -cli_print_volume_status_fd (dict_t *dict, gf_boolean_t notbrick) -{ - int ret = -1; - char *volname = NULL; - int brick_index_max = -1; - int other_count = 0; - int index_max = 0; - char *hostname = NULL; - char *path = NULL; - int online = -1; - int conn_count = 0; - char key[1024] = {0,}; - int i = 0; - int j = 0; - - GF_ASSERT (dict); - - ret = dict_get_str (dict, "volname", &volname); - if (ret) - goto out; - cli_out ("FD tables for volume %s", volname); +cli_print_volume_status_call_frame(dict_t *dict, char *prefix) +{ + int ret = -1; + char key[1024] = { + 0, + }; + int ref_count = 0; + char *translator = 0; + int complete = 0; + char *parent = NULL; + char *wind_from = NULL; + char *wind_to = NULL; + char *unwind_from = NULL; + char *unwind_to = NULL; + + if (!dict || !prefix) + return; - ret = dict_get_int32 (dict, "brick-index-max", &brick_index_max); - if (ret) - goto out; - ret = dict_get_int32 (dict, "other-count", &other_count); - if (ret) - goto out; + snprintf(key, sizeof(key), "%s.refcount", prefix); + ret = dict_get_int32(dict, key, &ref_count); + if (ret) + return; - index_max = brick_index_max + other_count; + snprintf(key, sizeof(key), "%s.translator", prefix); + ret = dict_get_str(dict, key, &translator); + if (ret) + return; - for (i = 0; i <= index_max; i++) { - cli_out ("----------------------------------------------"); + snprintf(key, sizeof(key), "%s.complete", prefix); + ret = dict_get_int32(dict, key, &complete); + if (ret) + return; - snprintf (key, sizeof (key), "brick%d.hostname", i); - ret = dict_get_str (dict, key, &hostname); - if (ret) - goto out; - snprintf (key, sizeof (key), "brick%d.path", i); - ret = dict_get_str (dict, key, &path); - if (ret) - goto out; + cli_out(" Ref Count = %d", ref_count); + cli_out(" Translator = %s", translator); + cli_out(" Completed = %s", (complete ? "Yes" : "No")); - if (notbrick) - cli_out ("%s : %s", hostname, path); - else - cli_out ("Brick : %s:%s", hostname, path); + snprintf(key, sizeof(key), "%s.parent", prefix); + ret = dict_get_str(dict, key, &parent); + if (!ret) + cli_out(" Parent = %s", parent); - snprintf (key, sizeof (key), "brick%d.status", i); - ret = dict_get_int32 (dict, key, &online); - if (ret) - goto out; - if (!online) { - if (notbrick) - cli_out ("%s is offline", hostname); - else - cli_out ("Brick is offline"); - continue; - } + snprintf(key, sizeof(key), "%s.windfrom", prefix); + ret = dict_get_str(dict, key, &wind_from); + if (!ret) + cli_out(" Wind From = %s", wind_from); - snprintf (key, sizeof (key), "brick%d.conncount", i); - ret = dict_get_int32 (dict, key, &conn_count); - if (ret) - goto out; + snprintf(key, sizeof(key), "%s.windto", prefix); + ret = dict_get_str(dict, key, &wind_to); + if (!ret) + cli_out(" Wind To = %s", wind_to); - for (j = 0; j < conn_count; j++) { - cli_out ("Connection %d:", j+1); + snprintf(key, sizeof(key), "%s.unwindfrom", prefix); + ret = dict_get_str(dict, key, &unwind_from); + if (!ret) + cli_out(" Unwind From = %s", unwind_from); - snprintf (key, sizeof (key), "brick%d.conn%d.fdtable", - i, j); - cli_print_volume_status_fdtable (dict, key); - cli_out (" "); - } - } -out: - cli_out ("----------------------------------------------"); - return; + snprintf(key, sizeof(key), "%s.unwindto", prefix); + ret = dict_get_str(dict, key, &unwind_to); + if (!ret) + cli_out(" Unwind To = %s", unwind_to); } void -cli_print_volume_status_call_frame (dict_t *dict, char *prefix) -{ - int ret = -1; - char key[1024] = {0,}; - int ref_count = 0; - char *translator = 0; - int complete = 0; - char *parent = NULL; - char *wind_from = NULL; - char *wind_to = NULL; - char *unwind_from = NULL; - char *unwind_to = NULL; +cli_print_volume_status_call_stack(dict_t *dict, char *prefix) +{ + int ret = -1; + char key[1024] = { + 0, + }; + int uid = 0; + int gid = 0; + int pid = 0; + uint64_t unique = 0; + // char *op = NULL; + int count = 0; + int i = 0; + + if (!dict || !prefix) + return; - if (!dict || !prefix) - return; + snprintf(key, sizeof(key), "%s.uid", prefix); + ret = dict_get_int32(dict, key, &uid); + if (ret) + return; - snprintf (key, sizeof (key), "%s.refcount", prefix); - ret = dict_get_int32 (dict, key, &ref_count); - if (ret) - return; + snprintf(key, sizeof(key), "%s.gid", prefix); + ret = dict_get_int32(dict, key, &gid); + if (ret) + return; - snprintf (key, sizeof (key), "%s.translator", prefix); - ret = dict_get_str (dict, key, &translator); - if (ret) - return; + snprintf(key, sizeof(key), "%s.pid", prefix); + ret = dict_get_int32(dict, key, &pid); + if (ret) + return; - snprintf (key, sizeof (key), "%s.complete", prefix); - ret = dict_get_int32 (dict, key, &complete); - if (ret) - return; + snprintf(key, sizeof(key), "%s.unique", prefix); + ret = dict_get_uint64(dict, key, &unique); + if (ret) + return; - cli_out (" Ref Count = %d", ref_count); - cli_out (" Translator = %s", translator); - cli_out (" Completed = %s", (complete ? "Yes" : "No")); + /* + snprintf (key, sizeof (key), "%s.op", prefix); + ret = dict_get_str (dict, key, &op); + if (ret) + return; + */ - snprintf (key, sizeof (key), "%s.parent", prefix); - ret = dict_get_str (dict, key, &parent); - if (!ret) - cli_out (" Parent = %s", parent); + snprintf(key, sizeof(key), "%s.count", prefix); + ret = dict_get_int32(dict, key, &count); + if (ret) + return; - snprintf (key, sizeof (key), "%s.windfrom", prefix); - ret = dict_get_str (dict, key, &wind_from); - if (!ret) - cli_out (" Wind From = %s", wind_from); + cli_out(" UID : %d", uid); + cli_out(" GID : %d", gid); + cli_out(" PID : %d", pid); + cli_out(" Unique : %" PRIu64, unique); + // cli_out ("\tOp : %s", op); + cli_out(" Frames : %d", count); - snprintf (key, sizeof (key), "%s.windto", prefix); - ret = dict_get_str (dict, key, &wind_to); - if (!ret) - cli_out (" Wind To = %s", wind_to); + for (i = 0; i < count; i++) { + cli_out(" Frame %d", i + 1); - snprintf (key, sizeof (key), "%s.unwindfrom", prefix); - ret = dict_get_str (dict, key, &unwind_from); - if (!ret) - cli_out (" Unwind From = %s", unwind_from); + snprintf(key, sizeof(key), "%s.frame%d", prefix, i); + cli_print_volume_status_call_frame(dict, key); + } - snprintf (key, sizeof (key), "%s.unwindto", prefix); - ret = dict_get_str (dict, key, &unwind_to); - if (!ret) - cli_out (" Unwind To = %s", unwind_to); + cli_out(" "); } void -cli_print_volume_status_call_stack (dict_t *dict, char *prefix) -{ - int ret = -1; - char key[1024] = {0,}; - int uid = 0; - int gid = 0; - int pid = 0; - uint64_t unique = 0; - //char *op = NULL; - int count = 0; - int i = 0; - - if (!dict || !prefix) - return; - - snprintf (key, sizeof (key), "%s.uid", prefix); - ret = dict_get_int32 (dict, key, &uid); - if (ret) - return; - - snprintf (key, sizeof (key), "%s.gid", prefix); - ret = dict_get_int32 (dict, key, &gid); +cli_print_volume_status_callpool(dict_t *dict, gf_boolean_t notbrick) +{ + int ret = -1; + char *volname = NULL; + int brick_index_max = -1; + int other_count = 0; + int index_max = 0; + char *hostname = NULL; + char *path = NULL; + int online = -1; + int call_count = 0; + char key[1024] = { + 0, + }; + int i = 0; + int j = 0; + + GF_ASSERT(dict); + + ret = dict_get_str(dict, "volname", &volname); + if (ret) + goto out; + cli_out("Pending calls for volume %s", volname); + + ret = dict_get_int32(dict, "brick-index-max", &brick_index_max); + if (ret) + goto out; + ret = dict_get_int32(dict, "other-count", &other_count); + if (ret) + goto out; + + index_max = brick_index_max + other_count; + + for (i = 0; i <= index_max; i++) { + cli_out("----------------------------------------------"); + + snprintf(key, sizeof(key), "brick%d.hostname", i); + ret = dict_get_str(dict, key, &hostname); if (ret) - return; - - snprintf (key, sizeof (key), "%s.pid", prefix); - ret = dict_get_int32 (dict, key, &pid); - if (ret) - return; - - snprintf (key, sizeof (key), "%s.unique", prefix); - ret = dict_get_uint64 (dict, key, &unique); + goto out; + snprintf(key, sizeof(key), "brick%d.path", i); + ret = dict_get_str(dict, key, &path); if (ret) - return; - - /* - snprintf (key, sizeof (key), "%s.op", prefix); - ret = dict_get_str (dict, key, &op); - if (ret) - return; - */ + goto out; + if (notbrick) + cli_out("%s : %s", hostname, path); + else + cli_out("Brick : %s:%s", hostname, path); - snprintf (key, sizeof (key), "%s.count", prefix); - ret = dict_get_int32 (dict, key, &count); + snprintf(key, sizeof(key), "brick%d.status", i); + ret = dict_get_int32(dict, key, &online); if (ret) - return; - - cli_out (" UID : %d", uid); - cli_out (" GID : %d", gid); - cli_out (" PID : %d", pid); - cli_out (" Unique : %"PRIu64, unique); - //cli_out ("\tOp : %s", op); - cli_out (" Frames : %d", count); - - for (i = 0; i < count; i++) { - cli_out (" Frame %d", i+1); - - snprintf (key, sizeof (key), "%s.frame%d", prefix, i); - cli_print_volume_status_call_frame (dict, key); + goto out; + if (!online) { + if (notbrick) + cli_out("%s is offline", hostname); + else + cli_out("Brick is offline"); + continue; } - cli_out (" "); -} - -void -cli_print_volume_status_callpool (dict_t *dict, gf_boolean_t notbrick) -{ - int ret = -1; - char *volname = NULL; - int brick_index_max = -1; - int other_count = 0; - int index_max = 0; - char *hostname = NULL; - char *path = NULL; - int online = -1; - int call_count = 0; - char key[1024] = {0,}; - int i = 0; - int j = 0; - - GF_ASSERT (dict); - - ret = dict_get_str (dict, "volname", &volname); - if (ret) - goto out; - cli_out ("Pending calls for volume %s", volname); - - ret = dict_get_int32 (dict, "brick-index-max", &brick_index_max); + snprintf(key, sizeof(key), "brick%d.callpool.count", i); + ret = dict_get_int32(dict, key, &call_count); if (ret) - goto out; - ret = dict_get_int32 (dict, "other-count", &other_count); - if (ret) - goto out; + goto out; + cli_out("Pending calls: %d", call_count); - index_max = brick_index_max + other_count; + if (0 == call_count) + continue; - for (i = 0; i <= index_max; i++) { - cli_out ("----------------------------------------------"); + for (j = 0; j < call_count; j++) { + cli_out("Call Stack%d", j + 1); - snprintf (key, sizeof (key), "brick%d.hostname", i); - ret = dict_get_str (dict, key, &hostname); - if (ret) - goto out; - snprintf (key, sizeof (key), "brick%d.path", i); - ret = dict_get_str (dict, key, &path); - if (ret) - goto out; - - if (notbrick) - cli_out ("%s : %s", hostname, path); - else - cli_out ("Brick : %s:%s", hostname, path); - - snprintf (key, sizeof (key), "brick%d.status", i); - ret = dict_get_int32 (dict, key, &online); - if (ret) - goto out; - if (!online) { - if (notbrick) - cli_out ("%s is offline", hostname); - else - cli_out ("Brick is offline"); - continue; - } - - snprintf (key, sizeof (key), "brick%d.callpool.count", i); - ret = dict_get_int32 (dict, key, &call_count); - if (ret) - goto out; - cli_out ("Pending calls: %d", call_count); - - if (0 == call_count) - continue; - - for (j = 0; j < call_count; j++) { - cli_out ("Call Stack%d", j+1); - - snprintf (key, sizeof (key), - "brick%d.callpool.stack%d", i, j); - cli_print_volume_status_call_stack (dict, key); - } + snprintf(key, sizeof(key), "brick%d.callpool.stack%d", i, j); + cli_print_volume_status_call_stack(dict, key); } + } out: - cli_out ("----------------------------------------------"); - return; + cli_out("----------------------------------------------"); + return; } static void -cli_print_volume_status_tasks (dict_t *dict) -{ - int ret = -1; - int i = 0; - int j = 0; - int count = 0; - int task_count = 0; - int status = 0; - char *op = NULL; - char *task_id_str = NULL; - char *volname = NULL; - char key[64] = {0,}; - char task[32] = {0,}; - char *brick = NULL; - - ret = dict_get_str (dict, "volname", &volname); - if (ret) - goto out; - - ret = dict_get_int32 (dict, "tasks", &task_count); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Failed to get tasks count"); - return; - } +cli_print_volume_status_tasks(dict_t *dict) +{ + int ret = -1; + int i = 0; + int j = 0; + int count = 0; + int task_count = 0; + int status = 0; + char *op = NULL; + char *task_id_str = NULL; + char *volname = NULL; + char key[64] = { + 0, + }; + char task[32] = { + 0, + }; + char *brick = NULL; + + ret = dict_get_str(dict, "volname", &volname); + if (ret) + goto out; + + ret = dict_get_int32(dict, "tasks", &task_count); + if (ret) { + gf_log("cli", GF_LOG_ERROR, "Failed to get tasks count"); + return; + } - cli_out ("Task Status of Volume %s", volname); - cli_print_line (CLI_BRICK_STATUS_LINE_LEN); + cli_out("Task Status of Volume %s", volname); + cli_print_line(CLI_BRICK_STATUS_LINE_LEN); - if (task_count == 0) { - cli_out ("There are no active volume tasks"); - cli_out (" "); - return; - } + if (task_count == 0) { + cli_out("There are no active volume tasks"); + cli_out(" "); + return; + } - for (i = 0; i < task_count; i++) { - snprintf (key, sizeof (key), "task%d.type", i); - ret = dict_get_str(dict, key, &op); - if (ret) - return; - cli_out ("%-20s : %-20s", "Task", op); + for (i = 0; i < task_count; i++) { + snprintf(key, sizeof(key), "task%d.type", i); + ret = dict_get_str(dict, key, &op); + if (ret) + return; + cli_out("%-20s : %-20s", "Task", op); - snprintf (key, sizeof (key), "task%d.id", i); - ret = dict_get_str (dict, key, &task_id_str); - if (ret) - return; - cli_out ("%-20s : %-20s", "ID", task_id_str); + snprintf(key, sizeof(key), "task%d.id", i); + ret = dict_get_str(dict, key, &task_id_str); + if (ret) + return; + cli_out("%-20s : %-20s", "ID", task_id_str); - snprintf (key, sizeof (key), "task%d.status", i); - ret = dict_get_int32 (dict, key, &status); - if (ret) - return; + snprintf(key, sizeof(key), "task%d.status", i); + ret = dict_get_int32(dict, key, &status); + if (ret) + return; - snprintf (task, sizeof (task), "task%d", i); + snprintf(task, sizeof(task), "task%d", i); - if (!strcmp (op, "Remove brick")) { - snprintf (key, sizeof (key), "%s.count", task); - ret = dict_get_int32 (dict, key, &count); - if (ret) - goto out; + if (!strcmp(op, "Remove brick")) { + snprintf(key, sizeof(key), "%s.count", task); + ret = dict_get_int32(dict, key, &count); + if (ret) + goto out; - cli_out ("%-20s", "Removed bricks:"); + cli_out("%-20s", "Removed bricks:"); - for (j = 1; j <= count; j++) { - snprintf (key, sizeof (key),"%s.brick%d", - task, j); - ret = dict_get_str (dict, key, &brick); - if (ret) - goto out; + for (j = 1; j <= count; j++) { + snprintf(key, sizeof(key), "%s.brick%d", task, j); + ret = dict_get_str(dict, key, &brick); + if (ret) + goto out; - cli_out ("%-20s", brick); - } - } - cli_out ("%-20s : %-20s", "Status", - cli_vol_task_status_str[status]); - cli_out (" "); + cli_out("%-20s", brick); + } } + cli_out("%-20s : %-20s", "Status", cli_vol_task_status_str[status]); + cli_out(" "); + } out: - return; + return; } static int -gf_cli_status_cbk (struct rpc_req *req, struct iovec *iov, - int count, void *myframe) -{ - int ret = -1; - int brick_index_max = -1; - int other_count = 0; - int index_max = 0; - int i = 0; - int type = -1; - int hot_brick_count = -1; - int pid = -1; - uint32_t cmd = 0; - gf_boolean_t notbrick = _gf_false; - char key[1024] = {0,}; - char *hostname = NULL; - char *path = NULL; - char *volname = NULL; - dict_t *dict = NULL; - gf_cli_rsp rsp = {0,}; - cli_volume_status_t status = {0}; - cli_local_t *local = NULL; - gf_boolean_t wipe_local = _gf_false; - char msg[1024] = {0,}; - - GF_ASSERT (myframe); - - if (req->rpc_status == -1) - goto out; - - ret = xdr_to_generic (*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); - if (ret < 0) { - gf_log (((call_frame_t *) myframe)->this->name, GF_LOG_ERROR, - "Failed to decode xdr response"); - goto out; - } - - gf_log ("cli", GF_LOG_DEBUG, "Received response to status cmd"); - - local = ((call_frame_t *) myframe)->local; +gf_cli_status_cbk(struct rpc_req *req, struct iovec *iov, int count, + void *myframe) +{ + int ret = -1; + int brick_index_max = -1; + int other_count = 0; + int index_max = 0; + int i = 0; + int type = -1; + int hot_brick_count = -1; + int pid = -1; + uint32_t cmd = 0; + gf_boolean_t notbrick = _gf_false; + char key[1024] = { + 0, + }; + char *hostname = NULL; + char *path = NULL; + char *volname = NULL; + dict_t *dict = NULL; + gf_cli_rsp rsp = { + 0, + }; + cli_volume_status_t status = {0}; + cli_local_t *local = NULL; + gf_boolean_t wipe_local = _gf_false; + char msg[1024] = { + 0, + }; + + GF_ASSERT(myframe); + + if (req->rpc_status == -1) + goto out; + + ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); + if (ret < 0) { + gf_log(((call_frame_t *)myframe)->this->name, GF_LOG_ERROR, + "Failed to decode xdr response"); + goto out; + } + + gf_log("cli", GF_LOG_DEBUG, "Received response to status cmd"); + + local = ((call_frame_t *)myframe)->local; + if (!local) { + local = cli_local_get(); if (!local) { - local = cli_local_get (); - if (!local) { - ret = -1; - gf_log ("cli", GF_LOG_ERROR, "Failed to get local"); - goto out; - } - wipe_local = _gf_true; + ret = -1; + gf_log("cli", GF_LOG_ERROR, "Failed to get local"); + goto out; } + wipe_local = _gf_true; + } - if (rsp.op_ret) { - if (strcmp (rsp.op_errstr, "")) - snprintf (msg, sizeof (msg), "%s", rsp.op_errstr); - else - snprintf (msg, sizeof (msg), "Unable to obtain volume " - "status information."); - - if (global_state->mode & GLUSTER_MODE_XML) { - if (!local->all) - cli_xml_output_str ("volStatus", msg, - rsp.op_ret, rsp.op_errno, - rsp.op_errstr); - ret = 0; - goto out; - } + if (rsp.op_ret) { + if (strcmp(rsp.op_errstr, "")) + snprintf(msg, sizeof(msg), "%s", rsp.op_errstr); + else + snprintf(msg, sizeof(msg), + "Unable to obtain volume " + "status information."); - cli_err ("%s", msg); - if (local && local->all) { - ret = 0; - cli_out (" "); - } else - ret = -1; + if (global_state->mode & GLUSTER_MODE_XML) { + if (!local->all) + cli_xml_output_str("volStatus", msg, rsp.op_ret, rsp.op_errno, + rsp.op_errstr); + ret = 0; + goto out; + } + + cli_err("%s", msg); + if (local && local->all) { + ret = 0; + cli_out(" "); + } else + ret = -1; + + goto out; + } + + dict = dict_new(); + if (!dict) + goto out; + + ret = dict_unserialize(rsp.dict.dict_val, rsp.dict.dict_len, &dict); + if (ret) + goto out; + + ret = dict_get_uint32(dict, "cmd", &cmd); + if (ret) + goto out; + + if ((cmd & GF_CLI_STATUS_ALL)) { + if (local && local->dict) { + dict_ref(dict); + ret = dict_set_static_ptr(local->dict, "rsp-dict", dict); + ret = 0; + } else { + gf_log("cli", GF_LOG_ERROR, "local not found"); + ret = -1; + } + goto out; + } + + if ((cmd & GF_CLI_STATUS_NFS) || (cmd & GF_CLI_STATUS_SHD) || + (cmd & GF_CLI_STATUS_QUOTAD) || (cmd & GF_CLI_STATUS_SNAPD) || + (cmd & GF_CLI_STATUS_BITD) || (cmd & GF_CLI_STATUS_SCRUB) || + (cmd & GF_CLI_STATUS_TIERD)) + notbrick = _gf_true; + + if (global_state->mode & GLUSTER_MODE_XML) { + if (!local->all) { + ret = cli_xml_output_vol_status_begin(local, rsp.op_ret, + rsp.op_errno, rsp.op_errstr); + if (ret) { + gf_log("cli", GF_LOG_ERROR, "Error outputting to xml"); + goto xml_end; + } + } + if (cmd & GF_CLI_STATUS_TASKS) { + ret = cli_xml_output_vol_status_tasks_detail(local, dict); + if (ret) { + gf_log("cli", GF_LOG_ERROR, + "Error outputting " + "to xml"); + goto xml_end; + } + } else { + ret = cli_xml_output_vol_status(local, dict); + if (ret) { + gf_log("cli", GF_LOG_ERROR, "Error outputting to xml"); + goto xml_end; + } + } + + xml_end: + if (!local->all) { + ret = cli_xml_output_vol_status_end(local); + if (ret) { + gf_log("cli", GF_LOG_ERROR, "Error outputting to xml"); + } + } + goto out; + } + + status.brick = GF_MALLOC(PATH_MAX + 256, gf_common_mt_strdup); + if (!status.brick) { + errno = ENOMEM; + ret = -1; + goto out; + } + switch (cmd & GF_CLI_STATUS_MASK) { + case GF_CLI_STATUS_MEM: + cli_print_volume_status_mem(dict, notbrick); + goto cont; + break; + case GF_CLI_STATUS_CLIENTS: + cli_print_volume_status_clients(dict, notbrick); + goto cont; + break; + case GF_CLI_STATUS_CLIENT_LIST: + cli_print_volume_status_client_list(dict, notbrick); + goto cont; + break; + case GF_CLI_STATUS_INODE: + cli_print_volume_status_inode(dict, notbrick); + goto cont; + break; + case GF_CLI_STATUS_FD: + cli_print_volume_status_fd(dict, notbrick); + goto cont; + break; + case GF_CLI_STATUS_CALLPOOL: + cli_print_volume_status_callpool(dict, notbrick); + goto cont; + break; + case GF_CLI_STATUS_TASKS: + cli_print_volume_status_tasks(dict); + goto cont; + break; + default: + break; + } - goto out; - } + ret = dict_get_str(dict, "volname", &volname); + if (ret) + goto out; - dict = dict_new (); - if (!dict) - goto out; + ret = dict_get_int32(dict, "brick-index-max", &brick_index_max); + if (ret) + goto out; - ret = dict_unserialize (rsp.dict.dict_val, - rsp.dict.dict_len, - &dict); - if (ret) - goto out; + ret = dict_get_int32(dict, "other-count", &other_count); + if (ret) + goto out; - ret = dict_get_uint32 (dict, "cmd", &cmd); - if (ret) - goto out; + index_max = brick_index_max + other_count; - if ((cmd & GF_CLI_STATUS_ALL)) { - if (local && local->dict) { - dict_ref (dict); - ret = dict_set_static_ptr (local->dict, "rsp-dict", dict); - ret = 0; - } else { - gf_log ("cli", GF_LOG_ERROR, "local not found"); - ret = -1; - } - goto out; - } - - if ((cmd & GF_CLI_STATUS_NFS) || (cmd & GF_CLI_STATUS_SHD) || - (cmd & GF_CLI_STATUS_QUOTAD) || (cmd & GF_CLI_STATUS_SNAPD) || - (cmd & GF_CLI_STATUS_BITD) || (cmd & GF_CLI_STATUS_SCRUB) || - (cmd & GF_CLI_STATUS_TIERD)) - notbrick = _gf_true; + ret = dict_get_int32(dict, "type", &type); + if (ret) + goto out; - if (global_state->mode & GLUSTER_MODE_XML) { - if (!local->all) { - ret = cli_xml_output_vol_status_begin (local, - rsp.op_ret, - rsp.op_errno, - rsp.op_errstr); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, - "Error outputting to xml"); - goto xml_end; - } - } - if (cmd & GF_CLI_STATUS_TASKS) { - ret = cli_xml_output_vol_status_tasks_detail (local, - dict); - if (ret) { - gf_log ("cli", GF_LOG_ERROR,"Error outputting " - "to xml"); - goto xml_end; - } - } else { - ret = cli_xml_output_vol_status (local, dict); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, - "Error outputting to xml"); - goto xml_end; - } - } + ret = dict_get_int32(dict, "hot_brick_count", &hot_brick_count); + if (ret) + goto out; -xml_end: - if (!local->all) { - ret = cli_xml_output_vol_status_end (local); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, - "Error outputting to xml"); - } - } - goto out; - } + cli_out("Status of volume: %s", volname); - status.brick = GF_MALLOC (PATH_MAX + 256, gf_common_mt_strdup); - if (!status.brick) { - errno = ENOMEM; - ret = -1; - goto out; + if ((cmd & GF_CLI_STATUS_DETAIL) == 0) { + cli_out("%-*s %s %s %s %s", CLI_VOL_STATUS_BRICK_LEN, + "Gluster process", "TCP Port", "RDMA Port", "Online", "Pid"); + cli_print_line(CLI_BRICK_STATUS_LINE_LEN); + } + if (type == GF_CLUSTER_TYPE_TIER) { + cli_out("Hot Bricks:"); + } + for (i = 0; i <= index_max; i++) { + if (type == GF_CLUSTER_TYPE_TIER && i == hot_brick_count) { + cli_out("Cold Bricks:"); } - switch (cmd & GF_CLI_STATUS_MASK) { - case GF_CLI_STATUS_MEM: - cli_print_volume_status_mem (dict, notbrick); - goto cont; - break; - case GF_CLI_STATUS_CLIENTS: - cli_print_volume_status_clients (dict, notbrick); - goto cont; - break; - case GF_CLI_STATUS_CLIENT_LIST: - cli_print_volume_status_client_list (dict, notbrick); - goto cont; - break; - case GF_CLI_STATUS_INODE: - cli_print_volume_status_inode (dict, notbrick); - goto cont; - break; - case GF_CLI_STATUS_FD: - cli_print_volume_status_fd (dict, notbrick); - goto cont; - break; - case GF_CLI_STATUS_CALLPOOL: - cli_print_volume_status_callpool (dict, notbrick); - goto cont; - break; - case GF_CLI_STATUS_TASKS: - cli_print_volume_status_tasks (dict); - goto cont; - break; - default: - break; - } - - ret = dict_get_str (dict, "volname", &volname); - if (ret) - goto out; + status.rdma_port = 0; - ret = dict_get_int32 (dict, "brick-index-max", &brick_index_max); + snprintf(key, sizeof(key), "brick%d.hostname", i); + ret = dict_get_str(dict, key, &hostname); if (ret) - goto out; + continue; - ret = dict_get_int32 (dict, "other-count", &other_count); + snprintf(key, sizeof(key), "brick%d.path", i); + ret = dict_get_str(dict, key, &path); if (ret) - goto out; + continue; - index_max = brick_index_max + other_count; + /* Brick/not-brick is handled separately here as all + * types of nodes are contained in the default output + */ + memset(status.brick, 0, PATH_MAX + 255); + if (!strcmp(hostname, "NFS Server") || + !strcmp(hostname, "Self-heal Daemon") || + !strcmp(hostname, "Quota Daemon") || + !strcmp(hostname, "Snapshot Daemon") || + !strcmp(hostname, "Scrubber Daemon") || + !strcmp(hostname, "Bitrot Daemon") || + !strcmp(hostname, "Tier Daemon")) + snprintf(status.brick, PATH_MAX + 255, "%s on %s", hostname, path); + else { + snprintf(key, sizeof(key), "brick%d.rdma_port", i); + ret = dict_get_int32(dict, key, &(status.rdma_port)); + if (ret) + continue; + snprintf(status.brick, PATH_MAX + 255, "Brick %s:%s", hostname, + path); + } - ret = dict_get_int32 (dict, "type", &type); + snprintf(key, sizeof(key), "brick%d.port", i); + ret = dict_get_int32(dict, key, &(status.port)); if (ret) - goto out; + continue; - ret = dict_get_int32 (dict, "hot_brick_count", &hot_brick_count); + snprintf(key, sizeof(key), "brick%d.status", i); + ret = dict_get_int32(dict, key, &(status.online)); if (ret) - goto out; - - cli_out ("Status of volume: %s", volname); - - if ((cmd & GF_CLI_STATUS_DETAIL) == 0) { - cli_out ("%-*s %s %s %s %s", CLI_VOL_STATUS_BRICK_LEN, - "Gluster process", "TCP Port", "RDMA Port", - "Online", "Pid"); - cli_print_line (CLI_BRICK_STATUS_LINE_LEN); - } - if (type == GF_CLUSTER_TYPE_TIER) { - cli_out ("Hot Bricks:"); - } - for (i = 0; i <= index_max; i++) { - - if (type == GF_CLUSTER_TYPE_TIER && i == hot_brick_count) { - cli_out ("Cold Bricks:"); - } - status.rdma_port = 0; - - snprintf (key, sizeof (key), "brick%d.hostname", i); - ret = dict_get_str (dict, key, &hostname); - if (ret) - continue; - - snprintf (key, sizeof (key), "brick%d.path", i); - ret = dict_get_str (dict, key, &path); - if (ret) - continue; - - /* Brick/not-brick is handled separately here as all - * types of nodes are contained in the default output - */ - memset (status.brick, 0, PATH_MAX + 255); - if (!strcmp (hostname, "NFS Server") || - !strcmp (hostname, "Self-heal Daemon") || - !strcmp (hostname, "Quota Daemon") || - !strcmp (hostname, "Snapshot Daemon") || - !strcmp (hostname, "Scrubber Daemon") || - !strcmp (hostname, "Bitrot Daemon") || - !strcmp (hostname, "Tier Daemon")) - snprintf (status.brick, PATH_MAX + 255, "%s on %s", - hostname, path); - else { - snprintf (key, sizeof (key), "brick%d.rdma_port", i); - ret = dict_get_int32 (dict, key, &(status.rdma_port)); - if (ret) - continue; - snprintf (status.brick, PATH_MAX + 255, "Brick %s:%s", - hostname, path); - } + continue; - snprintf (key, sizeof (key), "brick%d.port", i); - ret = dict_get_int32 (dict, key, &(status.port)); - if (ret) - continue; - - snprintf (key, sizeof (key), "brick%d.status", i); - ret = dict_get_int32 (dict, key, &(status.online)); - if (ret) - continue; - - snprintf (key, sizeof (key), "brick%d.pid", i); - ret = dict_get_int32 (dict, key, &pid); - if (ret) - continue; - if (pid == -1) - ret = gf_asprintf (&(status.pid_str), "%s", "N/A"); - else - ret = gf_asprintf (&(status.pid_str), "%d", pid); - - if (ret == -1) - goto out; + snprintf(key, sizeof(key), "brick%d.pid", i); + ret = dict_get_int32(dict, key, &pid); + if (ret) + continue; + if (pid == -1) + ret = gf_asprintf(&(status.pid_str), "%s", "N/A"); + else + ret = gf_asprintf(&(status.pid_str), "%d", pid); - if ((cmd & GF_CLI_STATUS_DETAIL)) { - ret = cli_get_detail_status (dict, i, &status); - if (ret) - goto out; - cli_print_line (CLI_BRICK_STATUS_LINE_LEN); - cli_print_detailed_status (&status); - } else { - cli_print_brick_status (&status); - } + if (ret == -1) + goto out; + if ((cmd & GF_CLI_STATUS_DETAIL)) { + ret = cli_get_detail_status(dict, i, &status); + if (ret) + goto out; + cli_print_line(CLI_BRICK_STATUS_LINE_LEN); + cli_print_detailed_status(&status); + } else { + cli_print_brick_status(&status); } - cli_out (" "); + } + cli_out(" "); - if ((cmd & GF_CLI_STATUS_MASK) == GF_CLI_STATUS_NONE) - cli_print_volume_status_tasks (dict); + if ((cmd & GF_CLI_STATUS_MASK) == GF_CLI_STATUS_NONE) + cli_print_volume_status_tasks(dict); cont: - ret = rsp.op_ret; + ret = rsp.op_ret; out: - if (dict) - dict_unref (dict); - GF_FREE (status.brick); - if (local && wipe_local) { - cli_local_wipe (local); - } + if (dict) + dict_unref(dict); + GF_FREE(status.brick); + if (local && wipe_local) { + cli_local_wipe(local); + } - cli_cmd_broadcast_response (ret); - return ret; + cli_cmd_broadcast_response(ret); + return ret; } int32_t -gf_cli_status_volume (call_frame_t *frame, xlator_t *this, - void *data) +gf_cli_status_volume(call_frame_t *frame, xlator_t *this, void *data) { - gf_cli_req req = {{0,}}; - int ret = -1; - dict_t *dict = NULL; + gf_cli_req req = {{ + 0, + }}; + int ret = -1; + dict_t *dict = NULL; - if (!frame || !this || !data) - goto out; + if (!frame || !this || !data) + goto out; - dict = data; + dict = data; - ret = cli_to_glusterd (&req, frame, gf_cli_status_cbk, - (xdrproc_t) xdr_gf_cli_req, dict, - GLUSTER_CLI_STATUS_VOLUME, this, cli_rpc_prog, - NULL); - out: - gf_log ("cli", GF_LOG_DEBUG, "Returning: %d", ret); - return ret; + ret = cli_to_glusterd(&req, frame, gf_cli_status_cbk, + (xdrproc_t)xdr_gf_cli_req, dict, + GLUSTER_CLI_STATUS_VOLUME, this, cli_rpc_prog, NULL); +out: + gf_log("cli", GF_LOG_DEBUG, "Returning: %d", ret); + return ret; } int -gf_cli_status_volume_all (call_frame_t *frame, xlator_t *this, void *data) -{ - int i = 0; - int ret = -1; - int vol_count = -1; - uint32_t cmd = 0; - char key[1024] = {0}; - char *volname = NULL; - void *vol_dict = NULL; - dict_t *dict = NULL; - cli_local_t *local = NULL; - - if (!frame) - goto out; +gf_cli_status_volume_all(call_frame_t *frame, xlator_t *this, void *data) +{ + int i = 0; + int ret = -1; + int vol_count = -1; + uint32_t cmd = 0; + char key[1024] = {0}; + char *volname = NULL; + void *vol_dict = NULL; + dict_t *dict = NULL; + cli_local_t *local = NULL; - if (!frame->local) - goto out; + if (!frame) + goto out; - local = frame->local; + if (!frame->local) + goto out; - ret = dict_get_uint32 (local->dict, "cmd", &cmd); - if (ret) - goto out; + local = frame->local; - local->all = _gf_true; + ret = dict_get_uint32(local->dict, "cmd", &cmd); + if (ret) + goto out; - ret = gf_cli_status_volume (frame, this, data); + local->all = _gf_true; - if (ret) - goto out; + ret = gf_cli_status_volume(frame, this, data); - ret = dict_get_ptr (local->dict, "rsp-dict", &vol_dict); - if (ret) - goto out; + if (ret) + goto out; - ret = dict_get_int32 ((dict_t *)vol_dict, "vol_count", &vol_count); - if (ret) { - cli_err ("Failed to get names of volumes"); - goto out; - } + ret = dict_get_ptr(local->dict, "rsp-dict", &vol_dict); + if (ret) + goto out; - /* remove the "all" flag in cmd */ - cmd &= ~GF_CLI_STATUS_ALL; - cmd |= GF_CLI_STATUS_VOL; + ret = dict_get_int32((dict_t *)vol_dict, "vol_count", &vol_count); + if (ret) { + cli_err("Failed to get names of volumes"); + goto out; + } - if (global_state->mode & GLUSTER_MODE_XML) { - //TODO: Pass proper op_* values - ret = cli_xml_output_vol_status_begin (local, 0,0, NULL); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, - "Error outputting to xml"); - goto xml_end; - } - } + /* remove the "all" flag in cmd */ + cmd &= ~GF_CLI_STATUS_ALL; + cmd |= GF_CLI_STATUS_VOL; - if (vol_count == 0 && !(global_state->mode & GLUSTER_MODE_XML)) { - cli_err ("No volumes present"); - ret = 0; - goto out; + if (global_state->mode & GLUSTER_MODE_XML) { + // TODO: Pass proper op_* values + ret = cli_xml_output_vol_status_begin(local, 0, 0, NULL); + if (ret) { + gf_log("cli", GF_LOG_ERROR, "Error outputting to xml"); + goto xml_end; } + } - for (i = 0; i < vol_count; i++) { + if (vol_count == 0 && !(global_state->mode & GLUSTER_MODE_XML)) { + cli_err("No volumes present"); + ret = 0; + goto out; + } - dict = dict_new (); - if (!dict) - goto out; + for (i = 0; i < vol_count; i++) { + dict = dict_new(); + if (!dict) + goto out; - snprintf (key, sizeof (key), "vol%d", i); - ret = dict_get_str (vol_dict, key, &volname); - if (ret) - goto out; + snprintf(key, sizeof(key), "vol%d", i); + ret = dict_get_str(vol_dict, key, &volname); + if (ret) + goto out; - ret = dict_set_str (dict, "volname", volname); - if (ret) - goto out; + ret = dict_set_str(dict, "volname", volname); + if (ret) + goto out; - ret = dict_set_uint32 (dict, "cmd", cmd); - if (ret) - goto out; + ret = dict_set_uint32(dict, "cmd", cmd); + if (ret) + goto out; - ret = gf_cli_status_volume (frame, this, dict); - if (ret) - goto out; + ret = gf_cli_status_volume(frame, this, dict); + if (ret) + goto out; - dict_unref (dict); - } + dict_unref(dict); + } xml_end: - if (global_state->mode & GLUSTER_MODE_XML) { - ret = cli_xml_output_vol_status_end (local); - } + if (global_state->mode & GLUSTER_MODE_XML) { + ret = cli_xml_output_vol_status_end(local); + } - out: - if (ret) - gf_log ("cli", GF_LOG_ERROR, "status all failed"); +out: + if (ret) + gf_log("cli", GF_LOG_ERROR, "status all failed"); - if (vol_dict) - dict_unref (vol_dict); + if (vol_dict) + dict_unref(vol_dict); - if (ret && dict) - dict_unref (dict); + if (ret && dict) + dict_unref(dict); - if (local) - cli_local_wipe (local); + if (local) + cli_local_wipe(local); - if (frame) - frame->local = NULL; + if (frame) + frame->local = NULL; - return ret; + return ret; } static int -gf_cli_mount_cbk (struct rpc_req *req, struct iovec *iov, - int count, void *myframe) +gf_cli_mount_cbk(struct rpc_req *req, struct iovec *iov, int count, + void *myframe) { - gf1_cli_mount_rsp rsp = {0,}; - int ret = -1; + gf1_cli_mount_rsp rsp = { + 0, + }; + int ret = -1; - GF_ASSERT (myframe); + GF_ASSERT(myframe); - if (-1 == req->rpc_status) { - goto out; - } + if (-1 == req->rpc_status) { + goto out; + } - ret = xdr_to_generic (*iov, &rsp, (xdrproc_t)xdr_gf1_cli_mount_rsp); - if (ret < 0) { - gf_log (((call_frame_t *) myframe)->this->name, GF_LOG_ERROR, - "Failed to decode xdr response"); - goto out; - } + ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf1_cli_mount_rsp); + if (ret < 0) { + gf_log(((call_frame_t *)myframe)->this->name, GF_LOG_ERROR, + "Failed to decode xdr response"); + goto out; + } - gf_log ("cli", GF_LOG_INFO, "Received resp to mount"); + gf_log("cli", GF_LOG_INFO, "Received resp to mount"); - if (rsp.op_ret == 0) { - ret = 0; - cli_out ("%s", rsp.path); - } else { - /* weird sounding but easy to parse... */ - cli_err ("%d : failed with this errno (%s)", - rsp.op_errno, strerror (rsp.op_errno)); - ret = -1; - } + if (rsp.op_ret == 0) { + ret = 0; + cli_out("%s", rsp.path); + } else { + /* weird sounding but easy to parse... */ + cli_err("%d : failed with this errno (%s)", rsp.op_errno, + strerror(rsp.op_errno)); + ret = -1; + } out: - cli_cmd_broadcast_response (ret); - return ret; + cli_cmd_broadcast_response(ret); + return ret; } int32_t -gf_cli_mount (call_frame_t *frame, xlator_t *this, void *data) +gf_cli_mount(call_frame_t *frame, xlator_t *this, void *data) { - gf1_cli_mount_req req = {0,}; - int ret = -1; - void **dataa = data; - char *label = NULL; - dict_t *dict = NULL; + gf1_cli_mount_req req = { + 0, + }; + int ret = -1; + void **dataa = data; + char *label = NULL; + dict_t *dict = NULL; - if (!frame || !this || !data) - goto out; + if (!frame || !this || !data) + goto out; - label = dataa[0]; - dict = dataa[1]; + label = dataa[0]; + dict = dataa[1]; - req.label = label; - ret = dict_allocate_and_serialize (dict, &req.dict.dict_val, - &req.dict.dict_len); - if (ret) { - ret = -1; - goto out; - } + req.label = label; + ret = dict_allocate_and_serialize(dict, &req.dict.dict_val, + &req.dict.dict_len); + if (ret) { + ret = -1; + goto out; + } - ret = cli_cmd_submit (NULL, &req, frame, cli_rpc_prog, - GLUSTER_CLI_MOUNT, NULL, - this, gf_cli_mount_cbk, - (xdrproc_t)xdr_gf1_cli_mount_req); + ret = cli_cmd_submit(NULL, &req, frame, cli_rpc_prog, GLUSTER_CLI_MOUNT, + NULL, this, gf_cli_mount_cbk, + (xdrproc_t)xdr_gf1_cli_mount_req); out: - gf_log ("cli", GF_LOG_DEBUG, "Returning %d", ret); - return ret; + gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret); + return ret; } static int -gf_cli_umount_cbk (struct rpc_req *req, struct iovec *iov, - int count, void *myframe) +gf_cli_umount_cbk(struct rpc_req *req, struct iovec *iov, int count, + void *myframe) { - gf1_cli_umount_rsp rsp = {0,}; - int ret = -1; + gf1_cli_umount_rsp rsp = { + 0, + }; + int ret = -1; - GF_ASSERT (myframe); + GF_ASSERT(myframe); - if (-1 == req->rpc_status) { - goto out; - } + if (-1 == req->rpc_status) { + goto out; + } - ret = xdr_to_generic (*iov, &rsp, (xdrproc_t)xdr_gf1_cli_umount_rsp); - if (ret < 0) { - gf_log (((call_frame_t *) myframe)->this->name, GF_LOG_ERROR, - "Failed to decode xdr response"); - goto out; - } + ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf1_cli_umount_rsp); + if (ret < 0) { + gf_log(((call_frame_t *)myframe)->this->name, GF_LOG_ERROR, + "Failed to decode xdr response"); + goto out; + } - gf_log ("cli", GF_LOG_INFO, "Received resp to mount"); + gf_log("cli", GF_LOG_INFO, "Received resp to mount"); - if (rsp.op_ret == 0) - ret = 0; - else { - cli_err ("umount failed"); - ret = -1; - } + if (rsp.op_ret == 0) + ret = 0; + else { + cli_err("umount failed"); + ret = -1; + } out: - cli_cmd_broadcast_response (ret); - return ret; + cli_cmd_broadcast_response(ret); + return ret; } int32_t -gf_cli_umount (call_frame_t *frame, xlator_t *this, void *data) +gf_cli_umount(call_frame_t *frame, xlator_t *this, void *data) { - gf1_cli_umount_req req = {0,}; - int ret = -1; - dict_t *dict = NULL; + gf1_cli_umount_req req = { + 0, + }; + int ret = -1; + dict_t *dict = NULL; - if (!frame || !this || !data) - goto out; + if (!frame || !this || !data) + goto out; - dict = data; + dict = data; - ret = dict_get_str (dict, "path", &req.path); - if (ret == 0) - ret = dict_get_int32 (dict, "lazy", &req.lazy); + ret = dict_get_str(dict, "path", &req.path); + if (ret == 0) + ret = dict_get_int32(dict, "lazy", &req.lazy); - if (ret) { - ret = -1; - goto out; - } + if (ret) { + ret = -1; + goto out; + } - ret = cli_cmd_submit (NULL, &req, frame, cli_rpc_prog, - GLUSTER_CLI_UMOUNT, NULL, - this, gf_cli_umount_cbk, - (xdrproc_t)xdr_gf1_cli_umount_req); + ret = cli_cmd_submit(NULL, &req, frame, cli_rpc_prog, GLUSTER_CLI_UMOUNT, + NULL, this, gf_cli_umount_cbk, + (xdrproc_t)xdr_gf1_cli_umount_req); - out: - gf_log ("cli", GF_LOG_DEBUG, "Returning %d", ret); - return ret; +out: + gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret); + return ret; } void -cmd_heal_volume_statistics_out (dict_t *dict, int brick) -{ - - uint64_t num_entries = 0; - int ret = 0; - char key[256] = {0}; - char *hostname = NULL; - uint64_t i = 0; - uint64_t healed_count = 0; - uint64_t split_brain_count = 0; - uint64_t heal_failed_count = 0; - char *start_time_str = NULL; - char *end_time_str = NULL; - char *crawl_type = NULL; - int progress = -1; - - snprintf (key, sizeof key, "%d-hostname", brick); - ret = dict_get_str (dict, key, &hostname); +cmd_heal_volume_statistics_out(dict_t *dict, int brick) +{ + uint64_t num_entries = 0; + int ret = 0; + char key[256] = {0}; + char *hostname = NULL; + uint64_t i = 0; + uint64_t healed_count = 0; + uint64_t split_brain_count = 0; + uint64_t heal_failed_count = 0; + char *start_time_str = NULL; + char *end_time_str = NULL; + char *crawl_type = NULL; + int progress = -1; + + snprintf(key, sizeof key, "%d-hostname", brick); + ret = dict_get_str(dict, key, &hostname); + if (ret) + goto out; + cli_out("------------------------------------------------"); + cli_out("\nCrawl statistics for brick no %d", brick); + cli_out("Hostname of brick %s", hostname); + + snprintf(key, sizeof key, "statistics-%d-count", brick); + ret = dict_get_uint64(dict, key, &num_entries); + if (ret) + goto out; + + for (i = 0; i < num_entries; i++) { + snprintf(key, sizeof key, "statistics_crawl_type-%d-%" PRIu64, brick, + i); + ret = dict_get_str(dict, key, &crawl_type); if (ret) - goto out; - cli_out ("------------------------------------------------"); - cli_out ("\nCrawl statistics for brick no %d", brick); - cli_out ("Hostname of brick %s", hostname); + goto out; - snprintf (key, sizeof key, "statistics-%d-count", brick); - ret = dict_get_uint64 (dict, key, &num_entries); + snprintf(key, sizeof key, "statistics_healed_cnt-%d-%" PRIu64, brick, + i); + ret = dict_get_uint64(dict, key, &healed_count); if (ret) - goto out; - - for (i = 0; i < num_entries; i++) - { - snprintf (key, sizeof key, "statistics_crawl_type-%d-%"PRIu64, - brick, i); - ret = dict_get_str (dict, key, &crawl_type); - if (ret) - goto out; - - snprintf (key, sizeof key, "statistics_healed_cnt-%d-%"PRIu64, - brick,i); - ret = dict_get_uint64 (dict, key, &healed_count); - if (ret) - goto out; - - snprintf (key, sizeof key, "statistics_sb_cnt-%d-%"PRIu64, - brick, i); - ret = dict_get_uint64 (dict, key, &split_brain_count); - if (ret) - goto out; - snprintf (key, sizeof key, "statistics_heal_failed_cnt-%d-%"PRIu64, - brick, i); - ret = dict_get_uint64 (dict, key, &heal_failed_count); - if (ret) - goto out; - snprintf (key, sizeof key, "statistics_strt_time-%d-%"PRIu64, - brick, i); - ret = dict_get_str (dict, key, &start_time_str); - if (ret) - goto out; - snprintf (key, sizeof key, "statistics_end_time-%d-%"PRIu64, - brick, i); - ret = dict_get_str (dict, key, &end_time_str); - if (ret) - goto out; - snprintf (key, sizeof key, "statistics_inprogress-%d-%"PRIu64, - brick, i); - ret = dict_get_int32 (dict, key, &progress); - if (ret) - goto out; + goto out; - cli_out ("\nStarting time of crawl: %s", start_time_str); - if (progress == 1) - cli_out ("Crawl is in progress"); - else - cli_out ("Ending time of crawl: %s", end_time_str); - - cli_out ("Type of crawl: %s", crawl_type); - cli_out ("No. of entries healed: %"PRIu64, - healed_count); - cli_out ("No. of entries in split-brain: %"PRIu64, - split_brain_count); - cli_out ("No. of heal failed entries: %"PRIu64, - heal_failed_count); + snprintf(key, sizeof key, "statistics_sb_cnt-%d-%" PRIu64, brick, i); + ret = dict_get_uint64(dict, key, &split_brain_count); + if (ret) + goto out; + snprintf(key, sizeof key, "statistics_heal_failed_cnt-%d-%" PRIu64, + brick, i); + ret = dict_get_uint64(dict, key, &heal_failed_count); + if (ret) + goto out; + snprintf(key, sizeof key, "statistics_strt_time-%d-%" PRIu64, brick, i); + ret = dict_get_str(dict, key, &start_time_str); + if (ret) + goto out; + snprintf(key, sizeof key, "statistics_end_time-%d-%" PRIu64, brick, i); + ret = dict_get_str(dict, key, &end_time_str); + if (ret) + goto out; + snprintf(key, sizeof key, "statistics_inprogress-%d-%" PRIu64, brick, + i); + ret = dict_get_int32(dict, key, &progress); + if (ret) + goto out; - } + cli_out("\nStarting time of crawl: %s", start_time_str); + if (progress == 1) + cli_out("Crawl is in progress"); + else + cli_out("Ending time of crawl: %s", end_time_str); + cli_out("Type of crawl: %s", crawl_type); + cli_out("No. of entries healed: %" PRIu64, healed_count); + cli_out("No. of entries in split-brain: %" PRIu64, split_brain_count); + cli_out("No. of heal failed entries: %" PRIu64, heal_failed_count); + } out: - return; + return; } void -cmd_heal_volume_brick_out (dict_t *dict, int brick) -{ - uint64_t num_entries = 0; - int ret = 0; - char key[256] = {0}; - char *hostname = NULL; - char *path = NULL; - char *status = NULL; - uint64_t i = 0; - uint32_t time = 0; - char timestr[32] = {0}; - char *shd_status = NULL; - - snprintf (key, sizeof key, "%d-hostname", brick); - ret = dict_get_str (dict, key, &hostname); - if (ret) - goto out; - snprintf (key, sizeof key, "%d-path", brick); - ret = dict_get_str (dict, key, &path); - if (ret) - goto out; - cli_out ("\nBrick %s:%s", hostname, path); - - snprintf (key, sizeof key, "%d-status", brick); - ret = dict_get_str (dict, key, &status); - if (status && strlen (status)) - cli_out ("Status: %s", status); - - snprintf (key, sizeof key, "%d-shd-status",brick); - ret = dict_get_str (dict, key, &shd_status); - - if (!shd_status) { - snprintf (key, sizeof key, "%d-count", brick); - ret = dict_get_uint64 (dict, key, &num_entries); - cli_out ("Number of entries: %"PRIu64, num_entries); - - - for (i = 0; i < num_entries; i++) { - snprintf (key, sizeof key, "%d-%"PRIu64, brick, i); - ret = dict_get_str (dict, key, &path); - if (ret) - continue; - time = 0; - snprintf (key, sizeof key, "%d-%"PRIu64"-time", - brick, i); - ret = dict_get_uint32 (dict, key, &time); - if (ret || !time) { - cli_out ("%s", path); - } else { - gf_time_fmt (timestr, sizeof timestr, - time, gf_timefmt_FT); - if (i == 0) { - cli_out ("at path on brick"); - cli_out ("-----------------------------------"); - } - cli_out ("%s %s", timestr, path); - } - } - } - -out: - return; +cmd_heal_volume_brick_out(dict_t *dict, int brick) +{ + uint64_t num_entries = 0; + int ret = 0; + char key[256] = {0}; + char *hostname = NULL; + char *path = NULL; + char *status = NULL; + uint64_t i = 0; + uint32_t time = 0; + char timestr[32] = {0}; + char *shd_status = NULL; + + snprintf(key, sizeof key, "%d-hostname", brick); + ret = dict_get_str(dict, key, &hostname); + if (ret) + goto out; + snprintf(key, sizeof key, "%d-path", brick); + ret = dict_get_str(dict, key, &path); + if (ret) + goto out; + cli_out("\nBrick %s:%s", hostname, path); + + snprintf(key, sizeof key, "%d-status", brick); + ret = dict_get_str(dict, key, &status); + if (status && strlen(status)) + cli_out("Status: %s", status); + + snprintf(key, sizeof key, "%d-shd-status", brick); + ret = dict_get_str(dict, key, &shd_status); + + if (!shd_status) { + snprintf(key, sizeof key, "%d-count", brick); + ret = dict_get_uint64(dict, key, &num_entries); + cli_out("Number of entries: %" PRIu64, num_entries); + + for (i = 0; i < num_entries; i++) { + snprintf(key, sizeof key, "%d-%" PRIu64, brick, i); + ret = dict_get_str(dict, key, &path); + if (ret) + continue; + time = 0; + snprintf(key, sizeof key, "%d-%" PRIu64 "-time", brick, i); + ret = dict_get_uint32(dict, key, &time); + if (ret || !time) { + cli_out("%s", path); + } else { + gf_time_fmt(timestr, sizeof timestr, time, gf_timefmt_FT); + if (i == 0) { + cli_out("at path on brick"); + cli_out("-----------------------------------"); + } + cli_out("%s %s", timestr, path); + } + } + } + +out: + return; } - void -cmd_heal_volume_statistics_heal_count_out (dict_t *dict, int brick) -{ - uint64_t num_entries = 0; - int ret = 0; - char key[256] = {0}; - char *hostname = NULL; - char *path = NULL; - char *status = NULL; - char *shd_status = NULL; - - snprintf (key, sizeof key, "%d-hostname", brick); - ret = dict_get_str (dict, key, &hostname); +cmd_heal_volume_statistics_heal_count_out(dict_t *dict, int brick) +{ + uint64_t num_entries = 0; + int ret = 0; + char key[256] = {0}; + char *hostname = NULL; + char *path = NULL; + char *status = NULL; + char *shd_status = NULL; + + snprintf(key, sizeof key, "%d-hostname", brick); + ret = dict_get_str(dict, key, &hostname); + if (ret) + goto out; + snprintf(key, sizeof key, "%d-path", brick); + ret = dict_get_str(dict, key, &path); + if (ret) + goto out; + cli_out("\nBrick %s:%s", hostname, path); + + snprintf(key, sizeof key, "%d-status", brick); + ret = dict_get_str(dict, key, &status); + if (status && strlen(status)) + cli_out("Status: %s", status); + + snprintf(key, sizeof key, "%d-shd-status", brick); + ret = dict_get_str(dict, key, &shd_status); + + if (!shd_status) { + snprintf(key, sizeof key, "%d-hardlinks", brick); + ret = dict_get_uint64(dict, key, &num_entries); if (ret) - goto out; - snprintf (key, sizeof key, "%d-path", brick); - ret = dict_get_str (dict, key, &path); - if (ret) - goto out; - cli_out ("\nBrick %s:%s", hostname, path); - - snprintf (key, sizeof key, "%d-status", brick); - ret = dict_get_str (dict, key, &status); - if (status && strlen (status)) - cli_out ("Status: %s", status); - - snprintf (key, sizeof key, "%d-shd-status",brick); - ret = dict_get_str (dict, key, &shd_status); - - if(!shd_status) - { - snprintf (key, sizeof key, "%d-hardlinks", brick); - ret = dict_get_uint64 (dict, key, &num_entries); - if (ret) - cli_out ("No gathered input for this brick"); - else - cli_out ("Number of entries: %"PRIu64, num_entries); - - - } + cli_out("No gathered input for this brick"); + else + cli_out("Number of entries: %" PRIu64, num_entries); + } out: - return; + return; } int -gf_is_cli_heal_get_command (gf_xl_afr_op_t heal_op) -{ - /* If the command is get command value is 1 otherwise 0, for - invalid commands -1 */ - int get_cmds[GF_SHD_OP_HEAL_DISABLE + 1] = { - [GF_SHD_OP_INVALID] = -1, - [GF_SHD_OP_HEAL_INDEX] = 0, - [GF_SHD_OP_HEAL_FULL] = 0, - [GF_SHD_OP_INDEX_SUMMARY] = 1, - [GF_SHD_OP_HEALED_FILES] = 1, - [GF_SHD_OP_HEAL_FAILED_FILES] = 1, - [GF_SHD_OP_SPLIT_BRAIN_FILES] = 1, - [GF_SHD_OP_STATISTICS] = 1, - [GF_SHD_OP_STATISTICS_HEAL_COUNT] = 1, - [GF_SHD_OP_STATISTICS_HEAL_COUNT_PER_REPLICA] = 1, - [GF_SHD_OP_HEAL_ENABLE] = 0, - [GF_SHD_OP_HEAL_DISABLE] = 0, - }; - - if (heal_op > GF_SHD_OP_INVALID && heal_op <= GF_SHD_OP_HEAL_DISABLE) - return get_cmds[heal_op] == 1; - return _gf_false; +gf_is_cli_heal_get_command(gf_xl_afr_op_t heal_op) +{ + /* If the command is get command value is 1 otherwise 0, for + invalid commands -1 */ + int get_cmds[GF_SHD_OP_HEAL_DISABLE + 1] = { + [GF_SHD_OP_INVALID] = -1, + [GF_SHD_OP_HEAL_INDEX] = 0, + [GF_SHD_OP_HEAL_FULL] = 0, + [GF_SHD_OP_INDEX_SUMMARY] = 1, + [GF_SHD_OP_HEALED_FILES] = 1, + [GF_SHD_OP_HEAL_FAILED_FILES] = 1, + [GF_SHD_OP_SPLIT_BRAIN_FILES] = 1, + [GF_SHD_OP_STATISTICS] = 1, + [GF_SHD_OP_STATISTICS_HEAL_COUNT] = 1, + [GF_SHD_OP_STATISTICS_HEAL_COUNT_PER_REPLICA] = 1, + [GF_SHD_OP_HEAL_ENABLE] = 0, + [GF_SHD_OP_HEAL_DISABLE] = 0, + }; + + if (heal_op > GF_SHD_OP_INVALID && heal_op <= GF_SHD_OP_HEAL_DISABLE) + return get_cmds[heal_op] == 1; + return _gf_false; } int -gf_cli_heal_volume_cbk (struct rpc_req *req, struct iovec *iov, - int count, void *myframe) -{ - gf_cli_rsp rsp = {0,}; - int ret = -1; - cli_local_t *local = NULL; - char *volname = NULL; - call_frame_t *frame = NULL; - dict_t *dict = NULL; - int brick_count = 0; - int i = 0; - gf_xl_afr_op_t heal_op = GF_SHD_OP_INVALID; - char *operation = NULL; - char *substr = NULL; - char *heal_op_str = NULL; - - GF_ASSERT (myframe); - - if (-1 == req->rpc_status) { - goto out; - } - - frame = myframe; - - GF_ASSERT (frame->local); - - local = frame->local; - - ret = xdr_to_generic (*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); - if (ret < 0) { - gf_log (frame->this->name, GF_LOG_ERROR, - "Failed to decode xdr response"); - goto out; - } - - ret = dict_get_int32 (local->dict, "heal-op", (int32_t *)&heal_op); -//TODO: Proper XML output -//#if (HAVE_LIB_XML) -// if (global_state->mode & GLUSTER_MODE_XML) { -// ret = cli_xml_output_dict ("volHeal", dict, rsp.op_ret, -// rsp.op_errno, rsp.op_errstr); -// if (ret) -// gf_log ("cli", GF_LOG_ERROR, -// "Error outputting to xml"); -// goto out; -// } -//#endif - - ret = dict_get_str (local->dict, "volname", &volname); - if (ret) { - gf_log (frame->this->name, GF_LOG_ERROR, "failed to get volname"); - goto out; - } - - gf_log ("cli", GF_LOG_INFO, "Received resp to heal volume"); - - operation = "Gathering "; - substr = ""; - switch (heal_op) { - case GF_SHD_OP_HEAL_INDEX: - operation = "Launching heal operation "; - heal_op_str = "to perform index self heal"; - substr = "\nUse heal info commands to check" - " status."; - break; - case GF_SHD_OP_HEAL_FULL: - operation = "Launching heal operation "; - heal_op_str = "to perform full self heal"; - substr = "\nUse heal info commands to check" - " status."; - break; - case GF_SHD_OP_INDEX_SUMMARY: - heal_op_str = "list of entries to be healed"; - break; - case GF_SHD_OP_HEALED_FILES: - heal_op_str = "list of healed entries"; - break; - case GF_SHD_OP_HEAL_FAILED_FILES: - heal_op_str = "list of heal failed entries"; - break; - case GF_SHD_OP_SPLIT_BRAIN_FILES: - heal_op_str = "list of split brain entries"; - break; - case GF_SHD_OP_STATISTICS: - heal_op_str = "crawl statistics"; - break; - case GF_SHD_OP_STATISTICS_HEAL_COUNT: - heal_op_str = "count of entries to be healed"; - break; - case GF_SHD_OP_STATISTICS_HEAL_COUNT_PER_REPLICA: - heal_op_str = "count of entries to be healed per replica"; - break; +gf_cli_heal_volume_cbk(struct rpc_req *req, struct iovec *iov, int count, + void *myframe) +{ + gf_cli_rsp rsp = { + 0, + }; + int ret = -1; + cli_local_t *local = NULL; + char *volname = NULL; + call_frame_t *frame = NULL; + dict_t *dict = NULL; + int brick_count = 0; + int i = 0; + gf_xl_afr_op_t heal_op = GF_SHD_OP_INVALID; + char *operation = NULL; + char *substr = NULL; + char *heal_op_str = NULL; + + GF_ASSERT(myframe); + + if (-1 == req->rpc_status) { + goto out; + } + + frame = myframe; + + GF_ASSERT(frame->local); + + local = frame->local; + + ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); + if (ret < 0) { + gf_log(frame->this->name, GF_LOG_ERROR, + "Failed to decode xdr response"); + goto out; + } + + ret = dict_get_int32(local->dict, "heal-op", (int32_t *)&heal_op); + // TODO: Proper XML output + //#if (HAVE_LIB_XML) + // if (global_state->mode & GLUSTER_MODE_XML) { + // ret = cli_xml_output_dict ("volHeal", dict, rsp.op_ret, + // rsp.op_errno, rsp.op_errstr); + // if (ret) + // gf_log ("cli", GF_LOG_ERROR, + // "Error outputting to xml"); + // goto out; + // } + //#endif + + ret = dict_get_str(local->dict, "volname", &volname); + if (ret) { + gf_log(frame->this->name, GF_LOG_ERROR, "failed to get volname"); + goto out; + } + + gf_log("cli", GF_LOG_INFO, "Received resp to heal volume"); + + operation = "Gathering "; + substr = ""; + switch (heal_op) { + case GF_SHD_OP_HEAL_INDEX: + operation = "Launching heal operation "; + heal_op_str = "to perform index self heal"; + substr = + "\nUse heal info commands to check" + " status."; + break; + case GF_SHD_OP_HEAL_FULL: + operation = "Launching heal operation "; + heal_op_str = "to perform full self heal"; + substr = + "\nUse heal info commands to check" + " status."; + break; + case GF_SHD_OP_INDEX_SUMMARY: + heal_op_str = "list of entries to be healed"; + break; + case GF_SHD_OP_HEALED_FILES: + heal_op_str = "list of healed entries"; + break; + case GF_SHD_OP_HEAL_FAILED_FILES: + heal_op_str = "list of heal failed entries"; + break; + case GF_SHD_OP_SPLIT_BRAIN_FILES: + heal_op_str = "list of split brain entries"; + break; + case GF_SHD_OP_STATISTICS: + heal_op_str = "crawl statistics"; + break; + case GF_SHD_OP_STATISTICS_HEAL_COUNT: + heal_op_str = "count of entries to be healed"; + break; + case GF_SHD_OP_STATISTICS_HEAL_COUNT_PER_REPLICA: + heal_op_str = "count of entries to be healed per replica"; + break; /* The below 4 cases are never hit; they're coded only to make * compiler warnings go away.*/ - case GF_SHD_OP_SBRAIN_HEAL_FROM_BIGGER_FILE: - case GF_SHD_OP_SBRAIN_HEAL_FROM_LATEST_MTIME: - case GF_SHD_OP_SBRAIN_HEAL_FROM_BRICK: - case GF_SHD_OP_HEAL_SUMMARY: - break; - - case GF_SHD_OP_INVALID: - heal_op_str = "invalid heal op"; - break; - case GF_SHD_OP_HEAL_ENABLE: - operation = ""; - heal_op_str = "Enable heal"; - break; - case GF_SHD_OP_HEAL_DISABLE: - operation = ""; - heal_op_str = "Disable heal"; - break; - case GF_SHD_OP_GRANULAR_ENTRY_HEAL_ENABLE: - operation = ""; - heal_op_str = "Enable granular entry heal"; - break; - case GF_SHD_OP_GRANULAR_ENTRY_HEAL_DISABLE: - operation = ""; - heal_op_str = "Disable granular entry heal"; - break; - } - - if (rsp.op_ret) { - if (strcmp (rsp.op_errstr, "")) { - cli_err ("%s%s on volume %s has been unsuccessful:", - operation, heal_op_str, volname); - cli_err ("%s", rsp.op_errstr); - } - ret = rsp.op_ret; - goto out; - } else { - cli_out ("%s%s on volume %s has been successful %s", operation, - heal_op_str, volname, substr); + case GF_SHD_OP_SBRAIN_HEAL_FROM_BIGGER_FILE: + case GF_SHD_OP_SBRAIN_HEAL_FROM_LATEST_MTIME: + case GF_SHD_OP_SBRAIN_HEAL_FROM_BRICK: + case GF_SHD_OP_HEAL_SUMMARY: + break; + + case GF_SHD_OP_INVALID: + heal_op_str = "invalid heal op"; + break; + case GF_SHD_OP_HEAL_ENABLE: + operation = ""; + heal_op_str = "Enable heal"; + break; + case GF_SHD_OP_HEAL_DISABLE: + operation = ""; + heal_op_str = "Disable heal"; + break; + case GF_SHD_OP_GRANULAR_ENTRY_HEAL_ENABLE: + operation = ""; + heal_op_str = "Enable granular entry heal"; + break; + case GF_SHD_OP_GRANULAR_ENTRY_HEAL_DISABLE: + operation = ""; + heal_op_str = "Disable granular entry heal"; + break; + } + + if (rsp.op_ret) { + if (strcmp(rsp.op_errstr, "")) { + cli_err("%s%s on volume %s has been unsuccessful:", operation, + heal_op_str, volname); + cli_err("%s", rsp.op_errstr); } - ret = rsp.op_ret; - if (!gf_is_cli_heal_get_command (heal_op)) - goto out; - - dict = dict_new (); - if (!dict) { - ret = -1; - goto out; - } - - ret = dict_unserialize (rsp.dict.dict_val, - rsp.dict.dict_len, - &dict); - - if (ret) { - gf_log ("", GF_LOG_ERROR, - "Unable to allocate memory"); - goto out; - } else { - dict->extra_stdfree = rsp.dict.dict_val; - } - ret = dict_get_int32 (dict, "count", &brick_count); - if (ret) - goto out; - - if (!brick_count) { - cli_err ("All bricks of volume %s are down.", volname); - goto out; - } - - switch (heal_op) { + goto out; + } else { + cli_out("%s%s on volume %s has been successful %s", operation, + heal_op_str, volname, substr); + } + + ret = rsp.op_ret; + if (!gf_is_cli_heal_get_command(heal_op)) + goto out; + + dict = dict_new(); + if (!dict) { + ret = -1; + goto out; + } + + ret = dict_unserialize(rsp.dict.dict_val, rsp.dict.dict_len, &dict); + + if (ret) { + gf_log("", GF_LOG_ERROR, "Unable to allocate memory"); + goto out; + } else { + dict->extra_stdfree = rsp.dict.dict_val; + } + ret = dict_get_int32(dict, "count", &brick_count); + if (ret) + goto out; + + if (!brick_count) { + cli_err("All bricks of volume %s are down.", volname); + goto out; + } + + switch (heal_op) { case GF_SHD_OP_STATISTICS: - for (i = 0; i < brick_count; i++) - cmd_heal_volume_statistics_out (dict, i); - break; + for (i = 0; i < brick_count; i++) + cmd_heal_volume_statistics_out(dict, i); + break; case GF_SHD_OP_STATISTICS_HEAL_COUNT: case GF_SHD_OP_STATISTICS_HEAL_COUNT_PER_REPLICA: - for (i = 0; i < brick_count; i++) - cmd_heal_volume_statistics_heal_count_out (dict, - i); - break; + for (i = 0; i < brick_count; i++) + cmd_heal_volume_statistics_heal_count_out(dict, i); + break; case GF_SHD_OP_INDEX_SUMMARY: case GF_SHD_OP_HEALED_FILES: case GF_SHD_OP_HEAL_FAILED_FILES: case GF_SHD_OP_SPLIT_BRAIN_FILES: - for (i = 0; i < brick_count; i++) - cmd_heal_volume_brick_out (dict, i); - break; + for (i = 0; i < brick_count; i++) + cmd_heal_volume_brick_out(dict, i); + break; default: - break; - } + break; + } - ret = rsp.op_ret; + ret = rsp.op_ret; out: - cli_cmd_broadcast_response (ret); - free (rsp.op_errstr); - if (dict) - dict_unref (dict); - return ret; + cli_cmd_broadcast_response(ret); + free(rsp.op_errstr); + if (dict) + dict_unref(dict); + return ret; } int32_t -gf_cli_heal_volume (call_frame_t *frame, xlator_t *this, - void *data) +gf_cli_heal_volume(call_frame_t *frame, xlator_t *this, void *data) { - gf_cli_req req = {{0,}}; - int ret = 0; - dict_t *dict = NULL; + gf_cli_req req = {{ + 0, + }}; + int ret = 0; + dict_t *dict = NULL; - if (!frame || !this || !data) { - ret = -1; - goto out; - } + if (!frame || !this || !data) { + ret = -1; + goto out; + } - dict = data; + dict = data; - ret = cli_to_glusterd (&req, frame, gf_cli_heal_volume_cbk, - (xdrproc_t) xdr_gf_cli_req, dict, - GLUSTER_CLI_HEAL_VOLUME, this, cli_rpc_prog, - NULL); + ret = cli_to_glusterd(&req, frame, gf_cli_heal_volume_cbk, + (xdrproc_t)xdr_gf_cli_req, dict, + GLUSTER_CLI_HEAL_VOLUME, this, cli_rpc_prog, NULL); out: - gf_log ("cli", GF_LOG_DEBUG, "Returning %d", ret); + gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret); - GF_FREE (req.dict.dict_val); + GF_FREE(req.dict.dict_val); - return ret; + return ret; } int32_t -gf_cli_statedump_volume_cbk (struct rpc_req *req, struct iovec *iov, - int count, void *myframe) -{ - gf_cli_rsp rsp = {0,}; - int ret = -1; - char msg[1024] = {0,}; - - GF_ASSERT (myframe); - - if (-1 == req->rpc_status) - goto out; - - ret = xdr_to_generic (*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); - if (ret < 0) { - gf_log (((call_frame_t *) myframe)->this->name, GF_LOG_ERROR, - "Failed to decode xdr response"); - goto out; - } - gf_log ("cli", GF_LOG_DEBUG, "Received response to statedump"); - if (rsp.op_ret) - snprintf (msg, sizeof(msg), "%s", rsp.op_errstr); - else - snprintf (msg, sizeof (msg), "Volume statedump successful"); - - if (global_state->mode & GLUSTER_MODE_XML) { - ret = cli_xml_output_str ("volStatedump", msg, rsp.op_ret, - rsp.op_errno, rsp.op_errstr); - if (ret) - gf_log ("cli", GF_LOG_ERROR, - "Error outputting to xml"); - goto out; - } +gf_cli_statedump_volume_cbk(struct rpc_req *req, struct iovec *iov, int count, + void *myframe) +{ + gf_cli_rsp rsp = { + 0, + }; + int ret = -1; + char msg[1024] = { + 0, + }; + + GF_ASSERT(myframe); + + if (-1 == req->rpc_status) + goto out; + + ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); + if (ret < 0) { + gf_log(((call_frame_t *)myframe)->this->name, GF_LOG_ERROR, + "Failed to decode xdr response"); + goto out; + } + gf_log("cli", GF_LOG_DEBUG, "Received response to statedump"); + if (rsp.op_ret) + snprintf(msg, sizeof(msg), "%s", rsp.op_errstr); + else + snprintf(msg, sizeof(msg), "Volume statedump successful"); + + if (global_state->mode & GLUSTER_MODE_XML) { + ret = cli_xml_output_str("volStatedump", msg, rsp.op_ret, rsp.op_errno, + rsp.op_errstr); + if (ret) + gf_log("cli", GF_LOG_ERROR, "Error outputting to xml"); + goto out; + } - if (rsp.op_ret) - cli_err ("volume statedump: failed: %s", msg); - else - cli_out ("volume statedump: success"); - ret = rsp.op_ret; + if (rsp.op_ret) + cli_err("volume statedump: failed: %s", msg); + else + cli_out("volume statedump: success"); + ret = rsp.op_ret; out: - cli_cmd_broadcast_response (ret); - return ret; + cli_cmd_broadcast_response(ret); + return ret; } int32_t -gf_cli_statedump_volume (call_frame_t *frame, xlator_t *this, - void *data) +gf_cli_statedump_volume(call_frame_t *frame, xlator_t *this, void *data) { - gf_cli_req req = {{0,}}; - dict_t *options = NULL; - int ret = -1; + gf_cli_req req = {{ + 0, + }}; + dict_t *options = NULL; + int ret = -1; - if (!frame || !this || !data) - goto out; + if (!frame || !this || !data) + goto out; - options = data; + options = data; - ret = cli_to_glusterd (&req, frame, gf_cli_statedump_volume_cbk, - (xdrproc_t) xdr_gf_cli_req, options, - GLUSTER_CLI_STATEDUMP_VOLUME, this, cli_rpc_prog, - NULL); + ret = cli_to_glusterd( + &req, frame, gf_cli_statedump_volume_cbk, (xdrproc_t)xdr_gf_cli_req, + options, GLUSTER_CLI_STATEDUMP_VOLUME, this, cli_rpc_prog, NULL); out: - gf_log ("cli", GF_LOG_DEBUG, "Returning %d", ret); + gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret); - GF_FREE (req.dict.dict_val); - return ret; + GF_FREE(req.dict.dict_val); + return ret; } int32_t -gf_cli_list_volume_cbk (struct rpc_req *req, struct iovec *iov, - int count, void *myframe) -{ - int ret = -1; - gf_cli_rsp rsp = {0,}; - dict_t *dict = NULL; - int vol_count = 0;; - char *volname = NULL; - char key[1024] = {0,}; - int i = 0; - - GF_ASSERT (myframe); - - if (-1 == req->rpc_status) - goto out; - - ret = xdr_to_generic (*iov, &rsp, - (xdrproc_t)xdr_gf_cli_rsp); - if (ret < 0) { - gf_log (((call_frame_t *) myframe)->this->name, GF_LOG_ERROR, - "Failed to decode xdr response"); - goto out; - } - - dict = dict_new (); - if (!dict) { - ret = -1; - goto out; - } +gf_cli_list_volume_cbk(struct rpc_req *req, struct iovec *iov, int count, + void *myframe) +{ + int ret = -1; + gf_cli_rsp rsp = { + 0, + }; + dict_t *dict = NULL; + int vol_count = 0; + ; + char *volname = NULL; + char key[1024] = { + 0, + }; + int i = 0; + + GF_ASSERT(myframe); + + if (-1 == req->rpc_status) + goto out; + + ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); + if (ret < 0) { + gf_log(((call_frame_t *)myframe)->this->name, GF_LOG_ERROR, + "Failed to decode xdr response"); + goto out; + } + + dict = dict_new(); + if (!dict) { + ret = -1; + goto out; + } + + ret = dict_unserialize(rsp.dict.dict_val, rsp.dict.dict_len, &dict); + if (ret) { + gf_log("cli", GF_LOG_ERROR, "Unable to allocate memory"); + goto out; + } + + if (global_state->mode & GLUSTER_MODE_XML) { + ret = cli_xml_output_vol_list(dict, rsp.op_ret, rsp.op_errno, + rsp.op_errstr); + if (ret) + gf_log("cli", GF_LOG_ERROR, "Error outputting to xml"); + goto out; + } + + if (rsp.op_ret) + cli_err("%s", rsp.op_errstr); + else { + ret = dict_get_int32(dict, "count", &vol_count); + if (ret) + goto out; - ret = dict_unserialize (rsp.dict.dict_val, rsp.dict.dict_len, &dict); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Unable to allocate memory"); - goto out; + if (vol_count == 0) { + cli_err("No volumes present in cluster"); + goto out; } - - if (global_state->mode & GLUSTER_MODE_XML) { - ret = cli_xml_output_vol_list (dict, rsp.op_ret, rsp.op_errno, - rsp.op_errstr); - if (ret) - gf_log ("cli", GF_LOG_ERROR, - "Error outputting to xml"); + for (i = 0; i < vol_count; i++) { + snprintf(key, sizeof(key), "volume%d", i); + ret = dict_get_str(dict, key, &volname); + if (ret) goto out; + cli_out("%s", volname); } + } - if (rsp.op_ret) - cli_err ("%s", rsp.op_errstr); - else { - ret = dict_get_int32 (dict, "count", &vol_count); - if (ret) - goto out; - - if (vol_count == 0) { - cli_err ("No volumes present in cluster"); - goto out; - } - for (i = 0; i < vol_count; i++) { - snprintf (key, sizeof (key), "volume%d", i); - ret = dict_get_str (dict, key, &volname); - if (ret) - goto out; - cli_out ("%s", volname); - } - } - - ret = rsp.op_ret; + ret = rsp.op_ret; out: - cli_cmd_broadcast_response (ret); - return ret; + cli_cmd_broadcast_response(ret); + return ret; } int32_t -gf_cli_list_volume (call_frame_t *frame, xlator_t *this, void *data) +gf_cli_list_volume(call_frame_t *frame, xlator_t *this, void *data) { - int ret = -1; - gf_cli_req req = {{0,}}; + int ret = -1; + gf_cli_req req = {{ + 0, + }}; - if (!frame || !this) - goto out; + if (!frame || !this) + goto out; - ret = cli_cmd_submit (NULL, &req, frame, cli_rpc_prog, - GLUSTER_CLI_LIST_VOLUME, NULL, - this, gf_cli_list_volume_cbk, - (xdrproc_t)xdr_gf_cli_req); + ret = cli_cmd_submit(NULL, &req, frame, cli_rpc_prog, + GLUSTER_CLI_LIST_VOLUME, NULL, this, + gf_cli_list_volume_cbk, (xdrproc_t)xdr_gf_cli_req); out: - gf_log ("cli", GF_LOG_DEBUG, "Returning %d", ret); - return ret; + gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret); + return ret; } int32_t -gf_cli_clearlocks_volume_cbk (struct rpc_req *req, struct iovec *iov, - int count, void *myframe) -{ - gf_cli_rsp rsp = {0,}; - int ret = -1; - char *lk_summary = NULL; - char *volname = NULL; - dict_t *dict = NULL; - - GF_ASSERT (myframe); - - if (-1 == req->rpc_status) - goto out; - - ret = xdr_to_generic (*iov, &rsp, - (xdrproc_t)xdr_gf_cli_rsp); - if (ret < 0) { - gf_log (((call_frame_t *) myframe)->this->name, GF_LOG_ERROR, - "Failed to decode xdr response"); - goto out; +gf_cli_clearlocks_volume_cbk(struct rpc_req *req, struct iovec *iov, int count, + void *myframe) +{ + gf_cli_rsp rsp = { + 0, + }; + int ret = -1; + char *lk_summary = NULL; + char *volname = NULL; + dict_t *dict = NULL; + + GF_ASSERT(myframe); + + if (-1 == req->rpc_status) + goto out; + + ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); + if (ret < 0) { + gf_log(((call_frame_t *)myframe)->this->name, GF_LOG_ERROR, + "Failed to decode xdr response"); + goto out; + } + gf_log("cli", GF_LOG_DEBUG, "Received response to clear-locks"); + + if (rsp.op_ret) { + cli_err("Volume clear-locks unsuccessful"); + cli_err("%s", rsp.op_errstr); + + } else { + if (!rsp.dict.dict_len) { + cli_err("Possibly no locks cleared"); + ret = 0; + goto out; } - gf_log ("cli", GF_LOG_DEBUG, "Received response to clear-locks"); - - if (rsp.op_ret) { - cli_err ("Volume clear-locks unsuccessful"); - cli_err ("%s", rsp.op_errstr); - - } else { - if (!rsp.dict.dict_len) { - cli_err ("Possibly no locks cleared"); - ret = 0; - goto out; - } - dict = dict_new (); + dict = dict_new(); - if (!dict) { - ret = -1; - goto out; - } + if (!dict) { + ret = -1; + goto out; + } - ret = dict_unserialize (rsp.dict.dict_val, - rsp.dict.dict_len, - &dict); + ret = dict_unserialize(rsp.dict.dict_val, rsp.dict.dict_len, &dict); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, - "Unable to serialize response dictionary"); - goto out; - } - - ret = dict_get_str (dict, "volname", &volname); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Unable to get volname " - "from dictionary"); - goto out; - } + if (ret) { + gf_log("cli", GF_LOG_ERROR, + "Unable to serialize response dictionary"); + goto out; + } - ret = dict_get_str (dict, "lk-summary", &lk_summary); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Unable to get lock " - "summary from dictionary"); - goto out; - } - cli_out ("Volume clear-locks successful"); - cli_out ("%s", lk_summary); + ret = dict_get_str(dict, "volname", &volname); + if (ret) { + gf_log("cli", GF_LOG_ERROR, + "Unable to get volname " + "from dictionary"); + goto out; + } + ret = dict_get_str(dict, "lk-summary", &lk_summary); + if (ret) { + gf_log("cli", GF_LOG_ERROR, + "Unable to get lock " + "summary from dictionary"); + goto out; } + cli_out("Volume clear-locks successful"); + cli_out("%s", lk_summary); + } - ret = rsp.op_ret; + ret = rsp.op_ret; out: - if (dict) - dict_unref (dict); - cli_cmd_broadcast_response (ret); - return ret; + if (dict) + dict_unref(dict); + cli_cmd_broadcast_response(ret); + return ret; } int32_t -gf_cli_clearlocks_volume (call_frame_t *frame, xlator_t *this, - void *data) +gf_cli_clearlocks_volume(call_frame_t *frame, xlator_t *this, void *data) { - gf_cli_req req = {{0,}}; - dict_t *options = NULL; - int ret = -1; + gf_cli_req req = {{ + 0, + }}; + dict_t *options = NULL; + int ret = -1; - if (!frame || !this || !data) - goto out; + if (!frame || !this || !data) + goto out; - options = data; + options = data; - ret = cli_to_glusterd (&req, frame, gf_cli_clearlocks_volume_cbk, - (xdrproc_t) xdr_gf_cli_req, options, - GLUSTER_CLI_CLRLOCKS_VOLUME, this, cli_rpc_prog, - NULL); + ret = cli_to_glusterd( + &req, frame, gf_cli_clearlocks_volume_cbk, (xdrproc_t)xdr_gf_cli_req, + options, GLUSTER_CLI_CLRLOCKS_VOLUME, this, cli_rpc_prog, NULL); out: - gf_log ("cli", GF_LOG_DEBUG, "Returning %d", ret); + gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret); - GF_FREE (req.dict.dict_val); - return ret; + GF_FREE(req.dict.dict_val); + return ret; } int32_t -cli_snapshot_remove_reply (gf_cli_rsp *rsp, dict_t *dict, call_frame_t *frame) +cli_snapshot_remove_reply(gf_cli_rsp *rsp, dict_t *dict, call_frame_t *frame) { - int32_t ret = -1; - char *snap_name = NULL; - int32_t delete_cmd = -1; - cli_local_t *local = NULL; + int32_t ret = -1; + char *snap_name = NULL; + int32_t delete_cmd = -1; + cli_local_t *local = NULL; - GF_ASSERT (frame); - GF_ASSERT (rsp); - GF_ASSERT (dict); + GF_ASSERT(frame); + GF_ASSERT(rsp); + GF_ASSERT(dict); - local = frame->local; + local = frame->local; - ret = dict_get_int32 (dict, "sub-cmd", &delete_cmd); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Could not get sub-cmd"); - goto end; - } + ret = dict_get_int32(dict, "sub-cmd", &delete_cmd); + if (ret) { + gf_log("cli", GF_LOG_ERROR, "Could not get sub-cmd"); + goto end; + } - if ((global_state->mode & GLUSTER_MODE_XML) && - (delete_cmd == GF_SNAP_DELETE_TYPE_SNAP)) { - ret = cli_xml_output_snap_delete_begin (local, rsp->op_ret, - rsp->op_errno, - rsp->op_errstr); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Failed to create " - "xml output for delete"); - goto end; - } + if ((global_state->mode & GLUSTER_MODE_XML) && + (delete_cmd == GF_SNAP_DELETE_TYPE_SNAP)) { + ret = cli_xml_output_snap_delete_begin(local, rsp->op_ret, + rsp->op_errno, rsp->op_errstr); + if (ret) { + gf_log("cli", GF_LOG_ERROR, + "Failed to create " + "xml output for delete"); + goto end; + } + } + + if (rsp->op_ret && !(global_state->mode & GLUSTER_MODE_XML)) { + cli_err("snapshot delete: failed: %s", + rsp->op_errstr ? rsp->op_errstr + : "Please check log file for details"); + ret = rsp->op_ret; + goto out; + } + + if (delete_cmd == GF_SNAP_DELETE_TYPE_ALL || + delete_cmd == GF_SNAP_DELETE_TYPE_VOL) { + local = ((call_frame_t *)frame)->local; + if (!local) { + ret = -1; + gf_log("cli", GF_LOG_ERROR, "frame->local is NULL"); + goto out; } + /* During first call back of snapshot delete of type + * ALL and VOL, We will get the snapcount and snapnames. + * Hence to make the subsequent rpc calls for individual + * snapshot delete, We need to save it in local dictionary. + */ + dict_copy(dict, local->dict); + ret = 0; + goto out; + } - if (rsp->op_ret && !(global_state->mode & GLUSTER_MODE_XML)) { - cli_err ("snapshot delete: failed: %s", - rsp->op_errstr ? rsp->op_errstr : - "Please check log file for details"); - ret = rsp->op_ret; - goto out; + if (global_state->mode & GLUSTER_MODE_XML) { + ret = cli_xml_snapshot_delete(local, dict, rsp); + if (ret) { + gf_log("cli", GF_LOG_ERROR, + "Failed to create " + "xml output for snapshot delete command"); + goto out; } - - if (delete_cmd == GF_SNAP_DELETE_TYPE_ALL || - delete_cmd == GF_SNAP_DELETE_TYPE_VOL) { - local = ((call_frame_t *) frame) -> local; - if (!local) { - ret = -1; - gf_log ("cli", GF_LOG_ERROR, "frame->local is NULL"); - goto out; - } - - /* During first call back of snapshot delete of type - * ALL and VOL, We will get the snapcount and snapnames. - * Hence to make the subsequent rpc calls for individual - * snapshot delete, We need to save it in local dictionary. - */ - dict_copy (dict, local->dict); - ret = 0; - goto out; + /* Error out in case of the op already failed */ + if (rsp->op_ret) { + ret = rsp->op_ret; + goto out; } - - if (global_state->mode & GLUSTER_MODE_XML) { - ret = cli_xml_snapshot_delete (local, dict, rsp); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Failed to create " - "xml output for snapshot delete command"); - goto out; - } - /* Error out in case of the op already failed */ - if (rsp->op_ret) { - ret = rsp->op_ret; - goto out; - } - } else { - ret = dict_get_str (dict, "snapname", &snap_name); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Failed to get snapname"); - goto out; - } - - cli_out ("snapshot delete: %s: snap removed successfully", - snap_name); + } else { + ret = dict_get_str(dict, "snapname", &snap_name); + if (ret) { + gf_log("cli", GF_LOG_ERROR, "Failed to get snapname"); + goto out; } - ret = 0; + + cli_out("snapshot delete: %s: snap removed successfully", snap_name); + } + ret = 0; out: - if ((global_state->mode & GLUSTER_MODE_XML) && - (delete_cmd == GF_SNAP_DELETE_TYPE_SNAP)) { - ret = cli_xml_output_snap_delete_end (local); - } + if ((global_state->mode & GLUSTER_MODE_XML) && + (delete_cmd == GF_SNAP_DELETE_TYPE_SNAP)) { + ret = cli_xml_output_snap_delete_end(local); + } end: - return ret; + return ret; } int -cli_snapshot_config_display (dict_t *dict, gf_cli_rsp *rsp) -{ - char buf[PATH_MAX] = ""; - char *volname = NULL; - int ret = -1; - int config_command = 0; - uint64_t value = 0; - uint64_t hard_limit = 0; - uint64_t soft_limit = 0; - uint64_t i = 0; - uint64_t voldisplaycount = 0; - char *auto_delete = NULL; - char *snap_activate = NULL; - - GF_ASSERT (dict); - GF_ASSERT (rsp); +cli_snapshot_config_display(dict_t *dict, gf_cli_rsp *rsp) +{ + char buf[PATH_MAX] = ""; + char *volname = NULL; + int ret = -1; + int config_command = 0; + uint64_t value = 0; + uint64_t hard_limit = 0; + uint64_t soft_limit = 0; + uint64_t i = 0; + uint64_t voldisplaycount = 0; + char *auto_delete = NULL; + char *snap_activate = NULL; + + GF_ASSERT(dict); + GF_ASSERT(rsp); + + if (rsp->op_ret) { + cli_err("Snapshot Config : failed: %s", + rsp->op_errstr ? rsp->op_errstr + : "Please check log file for details"); + ret = rsp->op_ret; + goto out; + } + + ret = dict_get_int32(dict, "config-command", &config_command); + if (ret) { + gf_log("cli", GF_LOG_ERROR, "Could not fetch config type"); + goto out; + } + + ret = dict_get_str(dict, "volname", &volname); + /* Ignore the error, as volname is optional */ + + if (!volname) { + volname = "System"; + } + + ret = dict_get_uint64(dict, "snap-max-hard-limit", &hard_limit); + /* Ignore the error, as the key specified is optional */ + ret = dict_get_uint64(dict, "snap-max-soft-limit", &soft_limit); + + ret = dict_get_str(dict, "auto-delete", &auto_delete); + + ret = dict_get_str(dict, "snap-activate-on-create", &snap_activate); + + if (!hard_limit && !soft_limit && + config_command != GF_SNAP_CONFIG_DISPLAY && !auto_delete && + !snap_activate) { + ret = -1; + gf_log(THIS->name, GF_LOG_ERROR, "Could not fetch config-key"); + goto out; + } + + switch (config_command) { + case GF_SNAP_CONFIG_TYPE_SET: + if (hard_limit && soft_limit) { + cli_out( + "snapshot config: snap-max-hard-limit " + "& snap-max-soft-limit for system set " + "successfully"); + } else if (hard_limit) { + cli_out( + "snapshot config: snap-max-hard-limit " + "for %s set successfully", + volname); + } else if (soft_limit) { + cli_out( + "snapshot config: snap-max-soft-limit " + "for %s set successfully", + volname); + } else if (auto_delete) { + cli_out( + "snapshot config: auto-delete " + "successfully set"); + } else if (snap_activate) { + cli_out( + "snapshot config: activate-on-create " + "successfully set"); + } + break; - if (rsp->op_ret) { - cli_err ("Snapshot Config : failed: %s", - rsp->op_errstr ? rsp->op_errstr : - "Please check log file for details"); - ret = rsp->op_ret; + case GF_SNAP_CONFIG_DISPLAY: + cli_out("\nSnapshot System Configuration:"); + ret = dict_get_uint64(dict, "snap-max-hard-limit", &value); + if (ret) { + gf_log("cli", GF_LOG_ERROR, + "Could not fetch " + "snap_max_hard_limit for %s", + volname); + ret = -1; goto out; - } + } + cli_out("snap-max-hard-limit : %" PRIu64, value); - ret = dict_get_int32 (dict, "config-command", &config_command); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Could not fetch config type"); + ret = dict_get_uint64(dict, "snap-max-soft-limit", &soft_limit); + if (ret) { + gf_log("cli", GF_LOG_ERROR, + "Could not fetch " + "snap-max-soft-limit for %s", + volname); + ret = -1; goto out; - } - - ret = dict_get_str (dict, "volname", &volname); - /* Ignore the error, as volname is optional */ + } + cli_out("snap-max-soft-limit : %" PRIu64 "%%", soft_limit); - if (!volname) { - volname = "System"; - } + cli_out("auto-delete : %s", auto_delete); - ret = dict_get_uint64 (dict, "snap-max-hard-limit", &hard_limit); - /* Ignore the error, as the key specified is optional */ - ret = dict_get_uint64 (dict, "snap-max-soft-limit", &soft_limit); + cli_out("activate-on-create : %s\n", snap_activate); - ret = dict_get_str (dict, "auto-delete", &auto_delete); + cli_out("Snapshot Volume Configuration:"); - ret = dict_get_str (dict, "snap-activate-on-create", &snap_activate); - - if (!hard_limit && !soft_limit - && config_command != GF_SNAP_CONFIG_DISPLAY - && !auto_delete && !snap_activate) { + ret = dict_get_uint64(dict, "voldisplaycount", &voldisplaycount); + if (ret) { + gf_log("cli", GF_LOG_ERROR, "Could not fetch voldisplaycount"); ret = -1; - gf_log(THIS->name, GF_LOG_ERROR, - "Could not fetch config-key"); goto out; - } - - switch (config_command) { - case GF_SNAP_CONFIG_TYPE_SET: - if (hard_limit && soft_limit) { - cli_out ("snapshot config: snap-max-hard-limit " - "& snap-max-soft-limit for system set " - "successfully"); - } else if (hard_limit) { - cli_out ("snapshot config: snap-max-hard-limit " - "for %s set successfully", - volname); - } else if (soft_limit) { - cli_out ("snapshot config: snap-max-soft-limit " - "for %s set successfully", - volname); - } else if (auto_delete) { - cli_out ("snapshot config: auto-delete " - "successfully set"); - } else if (snap_activate) { - cli_out ("snapshot config: activate-on-create " - "successfully set"); - } - break; + } - case GF_SNAP_CONFIG_DISPLAY: - cli_out ("\nSnapshot System Configuration:"); - ret = dict_get_uint64 (dict, "snap-max-hard-limit", - &value); + for (i = 0; i < voldisplaycount; i++) { + snprintf(buf, sizeof(buf), "volume%" PRIu64 "-volname", i); + ret = dict_get_str(dict, buf, &volname); if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Could not fetch " - "snap_max_hard_limit for %s", volname); - ret = -1; - goto out; - } - cli_out ("snap-max-hard-limit : %"PRIu64, value); - - ret = dict_get_uint64 (dict, "snap-max-soft-limit", - &soft_limit); + gf_log("cli", GF_LOG_ERROR, + "Could not fetch " + " %s", + buf); + ret = -1; + goto out; + } + cli_out("\nVolume : %s", volname); + + snprintf(buf, sizeof(buf), + "volume%" PRIu64 "-snap-max-hard-limit", i); + ret = dict_get_uint64(dict, buf, &value); if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Could not fetch " - "snap-max-soft-limit for %s", volname); - ret = -1; - goto out; - } - cli_out ("snap-max-soft-limit : %"PRIu64"%%", - soft_limit); - - cli_out ("auto-delete : %s", auto_delete); - - cli_out ("activate-on-create : %s\n", snap_activate); - - cli_out ("Snapshot Volume Configuration:"); - - ret = dict_get_uint64 (dict, "voldisplaycount", - &voldisplaycount); + gf_log("cli", GF_LOG_ERROR, + "Could not fetch " + " %s", + buf); + ret = -1; + goto out; + } + cli_out("snap-max-hard-limit : %" PRIu64, value); + + snprintf(buf, sizeof(buf), + "volume%" PRIu64 "-active-hard-limit", i); + ret = dict_get_uint64(dict, buf, &value); if (ret) { - gf_log ("cli", GF_LOG_ERROR, - "Could not fetch voldisplaycount"); - ret = -1; - goto out; - } - - for (i = 0; i < voldisplaycount; i++) { - snprintf (buf, sizeof(buf), "volume%"PRIu64"-volname", i); - ret = dict_get_str (dict, buf, &volname); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Could not fetch " - " %s", buf); - ret = -1; - goto out; - } - cli_out ("\nVolume : %s", volname); - - snprintf (buf, sizeof(buf), - "volume%"PRIu64"-snap-max-hard-limit", i); - ret = dict_get_uint64 (dict, buf, &value); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Could not fetch " - " %s", buf); - ret = -1; - goto out; - } - cli_out ("snap-max-hard-limit : %"PRIu64, value); - - snprintf (buf, sizeof(buf), - "volume%"PRIu64"-active-hard-limit", i); - ret = dict_get_uint64 (dict, buf, &value); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Could not fetch" - " effective snap_max_hard_limit for " - "%s", volname); - ret = -1; - goto out; - } - cli_out ("Effective snap-max-hard-limit : %"PRIu64, - value); - - snprintf (buf, sizeof(buf), - "volume%"PRIu64"-snap-max-soft-limit", i); - ret = dict_get_uint64 (dict, buf, &value); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Could not fetch " - " %s", buf); - ret = -1; - goto out; - } - cli_out ("Effective snap-max-soft-limit : %"PRIu64" " - "(%"PRIu64"%%)", value, soft_limit); - } - break; + gf_log("cli", GF_LOG_ERROR, + "Could not fetch" + " effective snap_max_hard_limit for " + "%s", + volname); + ret = -1; + goto out; + } + cli_out("Effective snap-max-hard-limit : %" PRIu64, value); + + snprintf(buf, sizeof(buf), + "volume%" PRIu64 "-snap-max-soft-limit", i); + ret = dict_get_uint64(dict, buf, &value); + if (ret) { + gf_log("cli", GF_LOG_ERROR, + "Could not fetch " + " %s", + buf); + ret = -1; + goto out; + } + cli_out("Effective snap-max-soft-limit : %" PRIu64 + " " + "(%" PRIu64 "%%)", + value, soft_limit); + } + break; default: - break; - } + break; + } - ret = 0; + ret = 0; out: - return ret; + return ret; } /* This function is used to print the volume related information @@ -9749,91 +9706,86 @@ out: * arg - 1, prefix str : snaplist.snap{0..}.vol{0..}.* */ int -cli_get_each_volinfo_in_snap (dict_t *dict, char *keyprefix, - gf_boolean_t snap_driven) { - char key[PATH_MAX] = ""; - char *get_buffer = NULL; - int value = 0; - int ret = -1; - char indent[5] = "\t"; - char *volname = NULL; - - GF_ASSERT (dict); - GF_ASSERT (keyprefix); - - if (snap_driven) { - ret = snprintf (key, sizeof (key), "%s.volname", keyprefix); - if (ret < 0) { - goto out; - } - - ret = dict_get_str (dict, key, &get_buffer); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Failed to get %s", key); - goto out; - } - cli_out ("%s" INDENT_MAIN_HEAD "%s", indent, - "Snap Volume Name", ":", get_buffer); - - ret = snprintf (key, sizeof (key), - "%s.origin-volname", keyprefix); - if (ret < 0) { - goto out; - } +cli_get_each_volinfo_in_snap(dict_t *dict, char *keyprefix, + gf_boolean_t snap_driven) +{ + char key[PATH_MAX] = ""; + char *get_buffer = NULL; + int value = 0; + int ret = -1; + char indent[5] = "\t"; + char *volname = NULL; - ret = dict_get_str (dict, key, &volname); - if (ret) { - gf_log ("cli", GF_LOG_WARNING, "Failed to get %s", key); - cli_out ("%-12s", "Origin:"); - } - cli_out ("%s" INDENT_MAIN_HEAD "%s", indent, - "Origin Volume name", ":", volname); + GF_ASSERT(dict); + GF_ASSERT(keyprefix); + if (snap_driven) { + ret = snprintf(key, sizeof(key), "%s.volname", keyprefix); + if (ret < 0) { + goto out; + } - ret = snprintf (key, sizeof (key), "%s.snapcount", - keyprefix); - if (ret < 0) { - goto out; - } + ret = dict_get_str(dict, key, &get_buffer); + if (ret) { + gf_log("cli", GF_LOG_ERROR, "Failed to get %s", key); + goto out; + } + cli_out("%s" INDENT_MAIN_HEAD "%s", indent, "Snap Volume Name", ":", + get_buffer); - ret = dict_get_int32 (dict, key, &value); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Failed to get %s", key); - goto out; - } - cli_out ("%s%s %s %s %d", indent, "Snaps taken for", - volname, ":", value); + ret = snprintf(key, sizeof(key), "%s.origin-volname", keyprefix); + if (ret < 0) { + goto out; + } - ret = snprintf (key, sizeof (key), "%s.snaps-available", - keyprefix); - if (ret < 0) { - goto out; - } + ret = dict_get_str(dict, key, &volname); + if (ret) { + gf_log("cli", GF_LOG_WARNING, "Failed to get %s", key); + cli_out("%-12s", "Origin:"); + } + cli_out("%s" INDENT_MAIN_HEAD "%s", indent, "Origin Volume name", ":", + volname); - ret = dict_get_int32 (dict, key, &value); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Failed to get %s", key); - goto out; - } - cli_out ("%s%s %s %s %d", indent, "Snaps available for", - volname, ":", value); + ret = snprintf(key, sizeof(key), "%s.snapcount", keyprefix); + if (ret < 0) { + goto out; } + ret = dict_get_int32(dict, key, &value); + if (ret) { + gf_log("cli", GF_LOG_ERROR, "Failed to get %s", key); + goto out; + } + cli_out("%s%s %s %s %d", indent, "Snaps taken for", volname, ":", + value); - ret = snprintf (key, sizeof (key), "%s.vol-status", keyprefix); + ret = snprintf(key, sizeof(key), "%s.snaps-available", keyprefix); if (ret < 0) { - goto out; + goto out; } - ret = dict_get_str (dict, key, &get_buffer); + ret = dict_get_int32(dict, key, &value); if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Failed to get %s", key); - goto out; + gf_log("cli", GF_LOG_ERROR, "Failed to get %s", key); + goto out; } - cli_out ("%s" INDENT_MAIN_HEAD "%s", indent, "Status", - ":", get_buffer); + cli_out("%s%s %s %s %d", indent, "Snaps available for", volname, ":", + value); + } + + ret = snprintf(key, sizeof(key), "%s.vol-status", keyprefix); + if (ret < 0) { + goto out; + } + + ret = dict_get_str(dict, key, &get_buffer); + if (ret) { + gf_log("cli", GF_LOG_ERROR, "Failed to get %s", key); + goto out; + } + cli_out("%s" INDENT_MAIN_HEAD "%s", indent, "Status", ":", get_buffer); out: - return ret; + return ret; } /* This function is used to print snap related information @@ -9841,924 +9793,890 @@ out: * arg - 1, prefix_str : snaplist.snap{0..}.* */ int -cli_get_volinfo_in_snap (dict_t *dict, char *keyprefix) { +cli_get_volinfo_in_snap(dict_t *dict, char *keyprefix) +{ + char key[PATH_MAX] = ""; + int i = 0; + int volcount = 0; + int ret = -1; - char key[PATH_MAX] = ""; - int i = 0; - int volcount = 0; - int ret = -1; + GF_ASSERT(dict); + GF_ASSERT(keyprefix); - GF_ASSERT (dict); - GF_ASSERT (keyprefix); + ret = snprintf(key, sizeof(key), "%s.vol-count", keyprefix); + if (ret < 0) { + goto out; + } - ret = snprintf (key, sizeof (key), "%s.vol-count", keyprefix); + ret = dict_get_int32(dict, key, &volcount); + for (i = 1; i <= volcount; i++) { + ret = snprintf(key, sizeof(key), "%s.vol%d", keyprefix, i); if (ret < 0) { - goto out; + goto out; } - - ret = dict_get_int32 (dict, key, &volcount); - for (i = 1 ; i <= volcount ; i++) { - ret = snprintf (key, sizeof (key), - "%s.vol%d", keyprefix, i); - if (ret < 0) { - goto out; - } - ret = cli_get_each_volinfo_in_snap (dict, key, _gf_true); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Could not list " - "details of volume in a snap"); - goto out; - } - cli_out (" "); + ret = cli_get_each_volinfo_in_snap(dict, key, _gf_true); + if (ret) { + gf_log("cli", GF_LOG_ERROR, + "Could not list " + "details of volume in a snap"); + goto out; } + cli_out(" "); + } out: - return ret; + return ret; } int -cli_get_each_snap_info (dict_t *dict, char *prefix_str, - gf_boolean_t snap_driven) { - char key_buffer[PATH_MAX] = ""; - char *get_buffer = NULL; - int ret = -1; - char indent[5] = ""; - - GF_ASSERT (dict); - GF_ASSERT (prefix_str); - - if (!snap_driven) - strcat (indent, "\t"); - - ret = snprintf (key_buffer, sizeof (key_buffer), "%s.snapname", - prefix_str); - if (ret < 0) { - goto out; - } - - ret = dict_get_str (dict, key_buffer, &get_buffer); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Unable to fetch snapname %s ", - key_buffer); - goto out; - } - cli_out ("%s" INDENT_MAIN_HEAD "%s", indent, "Snapshot", - ":", get_buffer); - - ret = snprintf (key_buffer, sizeof (key_buffer), "%s.snap-id", - prefix_str); - if (ret < 0) { - goto out; - } - - ret = dict_get_str (dict, key_buffer, &get_buffer); +cli_get_each_snap_info(dict_t *dict, char *prefix_str, gf_boolean_t snap_driven) +{ + char key_buffer[PATH_MAX] = ""; + char *get_buffer = NULL; + int ret = -1; + char indent[5] = ""; + + GF_ASSERT(dict); + GF_ASSERT(prefix_str); + + if (!snap_driven) + strcat(indent, "\t"); + + ret = snprintf(key_buffer, sizeof(key_buffer), "%s.snapname", prefix_str); + if (ret < 0) { + goto out; + } + + ret = dict_get_str(dict, key_buffer, &get_buffer); + if (ret) { + gf_log("cli", GF_LOG_ERROR, "Unable to fetch snapname %s ", key_buffer); + goto out; + } + cli_out("%s" INDENT_MAIN_HEAD "%s", indent, "Snapshot", ":", get_buffer); + + ret = snprintf(key_buffer, sizeof(key_buffer), "%s.snap-id", prefix_str); + if (ret < 0) { + goto out; + } + + ret = dict_get_str(dict, key_buffer, &get_buffer); + if (ret) { + gf_log("cli", GF_LOG_ERROR, "Unable to fetch snap-id %s ", key_buffer); + goto out; + } + cli_out("%s" INDENT_MAIN_HEAD "%s", indent, "Snap UUID", ":", get_buffer); + + ret = snprintf(key_buffer, sizeof(key_buffer), "%s.snap-desc", prefix_str); + if (ret < 0) { + goto out; + } + + ret = dict_get_str(dict, key_buffer, &get_buffer); + if (!ret) { + /* Ignore error for description */ + cli_out("%s" INDENT_MAIN_HEAD "%s", indent, "Description", ":", + get_buffer); + } + + ret = snprintf(key_buffer, sizeof(key_buffer), "%s.snap-time", prefix_str); + if (ret < 0) { + goto out; + } + + ret = dict_get_str(dict, key_buffer, &get_buffer); + if (ret) { + gf_log("cli", GF_LOG_ERROR, "Unable to fetch snap-time %s ", + prefix_str); + goto out; + } + cli_out("%s" INDENT_MAIN_HEAD "%s", indent, "Created", ":", get_buffer); + + if (snap_driven) { + cli_out("%-12s", "Snap Volumes:\n"); + ret = cli_get_volinfo_in_snap(dict, prefix_str); if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Unable to fetch snap-id %s ", - key_buffer); - goto out; - } - cli_out ("%s" INDENT_MAIN_HEAD "%s", indent, "Snap UUID", - ":", get_buffer); - - ret = snprintf (key_buffer, sizeof (key_buffer), "%s.snap-desc", - prefix_str); - if (ret < 0) { - goto out; - } - - ret = dict_get_str (dict, key_buffer, &get_buffer); - if (!ret) { - /* Ignore error for description */ - cli_out ("%s" INDENT_MAIN_HEAD "%s", indent, - "Description", ":", get_buffer); - } - - ret = snprintf (key_buffer, sizeof (key_buffer), "%s.snap-time", - prefix_str); - if (ret < 0) { - goto out; - } - - ret = dict_get_str (dict, key_buffer, &get_buffer); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Unable to fetch snap-time %s ", - prefix_str); - goto out; - } - cli_out ("%s" INDENT_MAIN_HEAD "%s", indent, "Created", - ":", get_buffer); - - if (snap_driven) { - cli_out ("%-12s", "Snap Volumes:\n"); - ret = cli_get_volinfo_in_snap (dict, prefix_str); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Unable to list details " - "of the snaps"); - goto out; - } + gf_log("cli", GF_LOG_ERROR, + "Unable to list details " + "of the snaps"); + goto out; } + } out: - return ret; + return ret; } /* This is a generic function to print snap related information. * arg - 0, dict : Response Dictionary */ int -cli_call_snapshot_info (dict_t *dict, gf_boolean_t bool_snap_driven) { - int snap_count = 0; - char key[PATH_MAX] = ""; - int ret = -1; - int i = 0; +cli_call_snapshot_info(dict_t *dict, gf_boolean_t bool_snap_driven) +{ + int snap_count = 0; + char key[PATH_MAX] = ""; + int ret = -1; + int i = 0; - GF_ASSERT (dict); + GF_ASSERT(dict); - ret = dict_get_int32 (dict, "snapcount", &snap_count); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Unable to get snapcount"); - goto out; - } + ret = dict_get_int32(dict, "snapcount", &snap_count); + if (ret) { + gf_log("cli", GF_LOG_ERROR, "Unable to get snapcount"); + goto out; + } - if (snap_count == 0) { - cli_out ("No snapshots present"); - } + if (snap_count == 0) { + cli_out("No snapshots present"); + } - for (i = 1 ; i <= snap_count ; i++) { - ret = snprintf (key, sizeof (key), "snap%d", i); - if (ret < 0) { - goto out; - } - ret = cli_get_each_snap_info (dict, key, bool_snap_driven); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, - "Unable to print snap details"); - goto out; - } + for (i = 1; i <= snap_count; i++) { + ret = snprintf(key, sizeof(key), "snap%d", i); + if (ret < 0) { + goto out; + } + ret = cli_get_each_snap_info(dict, key, bool_snap_driven); + if (ret) { + gf_log("cli", GF_LOG_ERROR, "Unable to print snap details"); + goto out; } + } out: - return ret; + return ret; } int -cli_get_snaps_in_volume (dict_t *dict) { - int ret = -1; - int i = 0; - int count = 0; - int avail = 0; - char key[PATH_MAX] = ""; - char *get_buffer = NULL; - - GF_ASSERT (dict); - - ret = dict_get_str (dict, "origin-volname", &get_buffer); +cli_get_snaps_in_volume(dict_t *dict) +{ + int ret = -1; + int i = 0; + int count = 0; + int avail = 0; + char key[PATH_MAX] = ""; + char *get_buffer = NULL; + + GF_ASSERT(dict); + + ret = dict_get_str(dict, "origin-volname", &get_buffer); + if (ret) { + gf_log("cli", GF_LOG_ERROR, "Could not fetch origin-volname"); + goto out; + } + cli_out(INDENT_MAIN_HEAD "%s", "Volume Name", ":", get_buffer); + + ret = dict_get_int32(dict, "snapcount", &avail); + if (ret) { + gf_log("cli", GF_LOG_ERROR, "Could not fetch snapcount"); + goto out; + } + cli_out(INDENT_MAIN_HEAD "%d", "Snaps Taken", ":", avail); + + ret = dict_get_int32(dict, "snaps-available", &count); + if (ret) { + gf_log("cli", GF_LOG_ERROR, "Could not fetch snaps-available"); + goto out; + } + cli_out(INDENT_MAIN_HEAD "%d", "Snaps Available", ":", count); + + for (i = 1; i <= avail; i++) { + snprintf(key, sizeof(key), "snap%d", i); + ret = cli_get_each_snap_info(dict, key, _gf_false); if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Could not fetch origin-volname"); - goto out; + gf_log("cli", GF_LOG_ERROR, "Unable to print snap details"); + goto out; } - cli_out (INDENT_MAIN_HEAD "%s", "Volume Name", ":", get_buffer); - ret = dict_get_int32 (dict, "snapcount", &avail); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Could not fetch snapcount"); - goto out; + ret = snprintf(key, sizeof(key), "snap%d.vol1", i); + if (ret < 0) { + goto out; } - cli_out (INDENT_MAIN_HEAD "%d", "Snaps Taken", ":", avail); - - ret = dict_get_int32 (dict, "snaps-available", &count); + ret = cli_get_each_volinfo_in_snap(dict, key, _gf_false); if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Could not fetch snaps-available"); - goto out; + gf_log("cli", GF_LOG_ERROR, + "Could not get volume " + "related information"); + goto out; } - cli_out (INDENT_MAIN_HEAD "%d", "Snaps Available", ":", count); - - for (i = 1 ; i <= avail ; i++) { - snprintf (key, sizeof (key), "snap%d", i); - ret = cli_get_each_snap_info (dict, key, _gf_false); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, - "Unable to print snap details"); - goto out; - } - - ret = snprintf (key, sizeof (key), "snap%d.vol1", i); - if (ret < 0) { - goto out; - } - ret = cli_get_each_volinfo_in_snap (dict, key, _gf_false); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Could not get volume " - "related information"); - goto out; - } - cli_out (" "); - } + cli_out(" "); + } out: - return ret; + return ret; } int -cli_snapshot_list (dict_t *dict) { - int snapcount = 0; - char key[PATH_MAX] = ""; - int ret = -1; - int i = 0; - char *get_buffer = NULL; - - GF_ASSERT (dict); - - ret = dict_get_int32 (dict, "snapcount", &snapcount); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Could not fetch snap count"); - goto out; - } - - if (snapcount == 0) { - cli_out ("No snapshots present"); - } - - for (i = 1 ; i <= snapcount ; i++) { - ret = snprintf (key, sizeof (key), "snapname%d", i); - if (ret < 0) { - goto out; - } +cli_snapshot_list(dict_t *dict) +{ + int snapcount = 0; + char key[PATH_MAX] = ""; + int ret = -1; + int i = 0; + char *get_buffer = NULL; - ret = dict_get_str (dict, key, &get_buffer); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Could not get %s ", key); - goto out; - } else { - cli_out ("%s", get_buffer); - } - } -out: - return ret; -} + GF_ASSERT(dict); -int -cli_get_snap_volume_status (dict_t *dict, char *key_prefix) -{ - int ret = -1; - char key[PATH_MAX] = ""; - char *buffer = NULL; - int brickcount = 0; - int i = 0; - int pid = 0; + ret = dict_get_int32(dict, "snapcount", &snapcount); + if (ret) { + gf_log("cli", GF_LOG_ERROR, "Could not fetch snap count"); + goto out; + } - GF_ASSERT (dict); - GF_ASSERT (key_prefix); + if (snapcount == 0) { + cli_out("No snapshots present"); + } - ret = snprintf (key, sizeof (key), "%s.brickcount", key_prefix); + for (i = 1; i <= snapcount; i++) { + ret = snprintf(key, sizeof(key), "snapname%d", i); if (ret < 0) { - goto out; - } - ret = dict_get_int32 (dict, key, &brickcount); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Failed to fetch brickcount"); - goto out; + goto out; } - for (i = 0 ; i < brickcount ; i++) { - ret = snprintf (key, sizeof (key), "%s.brick%d.path", - key_prefix, i); - if (ret < 0) { - goto out; - } - - ret = dict_get_str (dict, key, &buffer); - if (ret) { - gf_log ("cli", GF_LOG_INFO, - "Unable to get Brick Path"); - continue; - } - cli_out ("\n\t%-17s %s %s", "Brick Path", ":", buffer); - - ret = snprintf (key, sizeof (key), "%s.brick%d.vgname", - key_prefix, i); - if (ret < 0) { - goto out; - } - - ret = dict_get_str (dict, key, &buffer); - if (ret) { - gf_log ("cli", GF_LOG_INFO, - "Unable to get Volume Group"); - cli_out ("\t%-17s %s %s", "Volume Group", ":", - "N/A"); - } else - cli_out ("\t%-17s %s %s", "Volume Group", ":", - buffer); - - ret = snprintf (key, sizeof (key), "%s.brick%d.status", - key_prefix, i); - if (ret < 0) { - goto out; - } - - ret = dict_get_str (dict, key, &buffer); - if (ret) { - gf_log ("cli", GF_LOG_INFO, - "Unable to get Brick Running"); - cli_out ("\t%-17s %s %s", "Brick Running", ":", - "N/A"); - } else - cli_out ("\t%-17s %s %s", "Brick Running", ":", - buffer); - - ret = snprintf (key, sizeof (key), "%s.brick%d.pid", - key_prefix, i); - if (ret < 0) { - goto out; - } - - ret = dict_get_int32 (dict, key, &pid); - if (ret) { - gf_log ("cli", GF_LOG_INFO, - "Unable to get pid"); - cli_out ("\t%-17s %s %s", "Brick PID", ":", "N/A"); - } else - cli_out ("\t%-17s %s %d", "Brick PID", ":", pid); - - ret = snprintf (key, sizeof (key), "%s.brick%d.data", - key_prefix, i); - if (ret < 0) { - goto out; - } - - ret = dict_get_str (dict, key, &buffer); - if (ret) { - gf_log ("cli", GF_LOG_INFO, - "Unable to get Data Percent"); - cli_out ("\t%-17s %s %s", "Data Percentage", ":", - "N/A"); - } else - cli_out ("\t%-17s %s %s", "Data Percentage", ":", - buffer); - - ret = snprintf (key, sizeof (key), "%s.brick%d.lvsize", - key_prefix, i); - if (ret < 0) { - goto out; - } - ret = dict_get_str (dict, key, &buffer); - if (ret) { - gf_log ("cli", GF_LOG_INFO, "Unable to get LV Size"); - cli_out ("\t%-17s %s %s", "LV Size", ":", "N/A"); - } else - cli_out ("\t%-17s %s %s", "LV Size", ":", buffer); - + ret = dict_get_str(dict, key, &get_buffer); + if (ret) { + gf_log("cli", GF_LOG_ERROR, "Could not get %s ", key); + goto out; + } else { + cli_out("%s", get_buffer); } - - ret = 0; + } out: - return ret; + return ret; } - - int -cli_get_single_snap_status (dict_t *dict, char *keyprefix) -{ - int ret = -1; - char key[PATH_MAX] = ""; - int i = 0; - int volcount = 0; - char *get_buffer = NULL; - - GF_ASSERT (dict); - GF_ASSERT (keyprefix); - - ret = snprintf (key, sizeof (key), "%s.snapname", keyprefix); +cli_get_snap_volume_status(dict_t *dict, char *key_prefix) +{ + int ret = -1; + char key[PATH_MAX] = ""; + char *buffer = NULL; + int brickcount = 0; + int i = 0; + int pid = 0; + + GF_ASSERT(dict); + GF_ASSERT(key_prefix); + + ret = snprintf(key, sizeof(key), "%s.brickcount", key_prefix); + if (ret < 0) { + goto out; + } + ret = dict_get_int32(dict, key, &brickcount); + if (ret) { + gf_log("cli", GF_LOG_ERROR, "Failed to fetch brickcount"); + goto out; + } + + for (i = 0; i < brickcount; i++) { + ret = snprintf(key, sizeof(key), "%s.brick%d.path", key_prefix, i); if (ret < 0) { - goto out; + goto out; } - ret = dict_get_str (dict, key, &get_buffer); + ret = dict_get_str(dict, key, &buffer); if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Unable to get snapname"); - goto out; + gf_log("cli", GF_LOG_INFO, "Unable to get Brick Path"); + continue; } - cli_out ("\nSnap Name : %s", get_buffer); + cli_out("\n\t%-17s %s %s", "Brick Path", ":", buffer); - ret = snprintf (key, sizeof (key), "%s.uuid", keyprefix); + ret = snprintf(key, sizeof(key), "%s.brick%d.vgname", key_prefix, i); if (ret < 0) { - goto out; + goto out; } - ret = dict_get_str (dict, key, &get_buffer); + ret = dict_get_str(dict, key, &buffer); if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Unable to get snap UUID"); - goto out; - } - cli_out ("Snap UUID : %s", get_buffer); + gf_log("cli", GF_LOG_INFO, "Unable to get Volume Group"); + cli_out("\t%-17s %s %s", "Volume Group", ":", "N/A"); + } else + cli_out("\t%-17s %s %s", "Volume Group", ":", buffer); - ret = snprintf (key, sizeof (key), "%s.volcount", keyprefix); + ret = snprintf(key, sizeof(key), "%s.brick%d.status", key_prefix, i); if (ret < 0) { - goto out; + goto out; } - ret = dict_get_int32 (dict, key, &volcount); + ret = dict_get_str(dict, key, &buffer); if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Unable to get volume count"); - goto out; - } + gf_log("cli", GF_LOG_INFO, "Unable to get Brick Running"); + cli_out("\t%-17s %s %s", "Brick Running", ":", "N/A"); + } else + cli_out("\t%-17s %s %s", "Brick Running", ":", buffer); - for (i = 0 ; i < volcount ; i++) { - ret = snprintf (key, sizeof (key), "%s.vol%d", keyprefix, i); - if (ret < 0) { - goto out; - } - - ret = cli_get_snap_volume_status (dict, key); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, - "Could not get snap volume status"); - goto out; - } + ret = snprintf(key, sizeof(key), "%s.brick%d.pid", key_prefix, i); + if (ret < 0) { + goto out; } -out: - return ret; -} - -int32_t -cli_populate_req_dict_for_delete (dict_t *snap_dict, dict_t *dict, size_t index) -{ - int32_t ret = -1; - char key[PATH_MAX] = ""; - char *buffer = NULL; - - GF_ASSERT (snap_dict); - GF_ASSERT (dict); - ret = dict_set_int32 (snap_dict, "sub-cmd", - GF_SNAP_DELETE_TYPE_ITER); + ret = dict_get_int32(dict, key, &pid); if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Could not save command " - "type in snap dictionary"); - goto out; - } + gf_log("cli", GF_LOG_INFO, "Unable to get pid"); + cli_out("\t%-17s %s %s", "Brick PID", ":", "N/A"); + } else + cli_out("\t%-17s %s %d", "Brick PID", ":", pid); - ret = snprintf (key, sizeof (key), "snapname%zu", index); + ret = snprintf(key, sizeof(key), "%s.brick%d.data", key_prefix, i); if (ret < 0) { - goto out; + goto out; } - ret = dict_get_str (dict, key, &buffer); + ret = dict_get_str(dict, key, &buffer); if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Failed to get snapname"); - goto out; - } + gf_log("cli", GF_LOG_INFO, "Unable to get Data Percent"); + cli_out("\t%-17s %s %s", "Data Percentage", ":", "N/A"); + } else + cli_out("\t%-17s %s %s", "Data Percentage", ":", buffer); - ret = dict_set_dynstr_with_alloc (snap_dict, "snapname", buffer); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Failed to save snapname"); - goto out; + ret = snprintf(key, sizeof(key), "%s.brick%d.lvsize", key_prefix, i); + if (ret < 0) { + goto out; } - - ret = dict_set_int32 (snap_dict, "type", GF_SNAP_OPTION_TYPE_DELETE); + ret = dict_get_str(dict, key, &buffer); if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Failed to save command type"); - goto out; - } + gf_log("cli", GF_LOG_INFO, "Unable to get LV Size"); + cli_out("\t%-17s %s %s", "LV Size", ":", "N/A"); + } else + cli_out("\t%-17s %s %s", "LV Size", ":", buffer); + } - ret = dict_set_dynstr_with_alloc (snap_dict, "cmd-str", - "snapshot delete"); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, - "Could not save command string as delete"); - goto out; - } + ret = 0; out: - return ret; + return ret; } int -cli_populate_req_dict_for_status (dict_t *snap_dict, dict_t *dict, int index) -{ - int ret = -1; - char key[PATH_MAX] = ""; - char *buffer = NULL; - - GF_ASSERT (snap_dict); - GF_ASSERT (dict); - - ret = dict_set_uint32 (snap_dict, "sub-cmd", - GF_SNAP_STATUS_TYPE_ITER); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Could not save command " - "type in snap dict"); - goto out; +cli_get_single_snap_status(dict_t *dict, char *keyprefix) +{ + int ret = -1; + char key[PATH_MAX] = ""; + int i = 0; + int volcount = 0; + char *get_buffer = NULL; + + GF_ASSERT(dict); + GF_ASSERT(keyprefix); + + ret = snprintf(key, sizeof(key), "%s.snapname", keyprefix); + if (ret < 0) { + goto out; + } + + ret = dict_get_str(dict, key, &get_buffer); + if (ret) { + gf_log("cli", GF_LOG_ERROR, "Unable to get snapname"); + goto out; + } + cli_out("\nSnap Name : %s", get_buffer); + + ret = snprintf(key, sizeof(key), "%s.uuid", keyprefix); + if (ret < 0) { + goto out; + } + + ret = dict_get_str(dict, key, &get_buffer); + if (ret) { + gf_log("cli", GF_LOG_ERROR, "Unable to get snap UUID"); + goto out; + } + cli_out("Snap UUID : %s", get_buffer); + + ret = snprintf(key, sizeof(key), "%s.volcount", keyprefix); + if (ret < 0) { + goto out; + } + + ret = dict_get_int32(dict, key, &volcount); + if (ret) { + gf_log("cli", GF_LOG_ERROR, "Unable to get volume count"); + goto out; + } + + for (i = 0; i < volcount; i++) { + ret = snprintf(key, sizeof(key), "%s.vol%d", keyprefix, i); + if (ret < 0) { + goto out; } - ret = snprintf (key, sizeof (key), "status.snap%d.snapname", index); - if (ret < 0) { - goto out; - } - - ret = dict_get_str (dict, key, &buffer); + ret = cli_get_snap_volume_status(dict, key); if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Could not get snapname"); - goto out; + gf_log("cli", GF_LOG_ERROR, "Could not get snap volume status"); + goto out; } + } +out: + return ret; +} - ret = dict_set_str (snap_dict, "snapname", buffer); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Could not save snapname " - "in snap dict"); - goto out; - - } +int32_t +cli_populate_req_dict_for_delete(dict_t *snap_dict, dict_t *dict, size_t index) +{ + int32_t ret = -1; + char key[PATH_MAX] = ""; + char *buffer = NULL; + + GF_ASSERT(snap_dict); + GF_ASSERT(dict); + + ret = dict_set_int32(snap_dict, "sub-cmd", GF_SNAP_DELETE_TYPE_ITER); + if (ret) { + gf_log("cli", GF_LOG_ERROR, + "Could not save command " + "type in snap dictionary"); + goto out; + } + + ret = snprintf(key, sizeof(key), "snapname%zu", index); + if (ret < 0) { + goto out; + } + + ret = dict_get_str(dict, key, &buffer); + if (ret) { + gf_log("cli", GF_LOG_ERROR, "Failed to get snapname"); + goto out; + } + + ret = dict_set_dynstr_with_alloc(snap_dict, "snapname", buffer); + if (ret) { + gf_log("cli", GF_LOG_ERROR, "Failed to save snapname"); + goto out; + } + + ret = dict_set_int32(snap_dict, "type", GF_SNAP_OPTION_TYPE_DELETE); + if (ret) { + gf_log("cli", GF_LOG_ERROR, "Failed to save command type"); + goto out; + } + + ret = dict_set_dynstr_with_alloc(snap_dict, "cmd-str", "snapshot delete"); + if (ret) { + gf_log("cli", GF_LOG_ERROR, "Could not save command string as delete"); + goto out; + } +out: + return ret; +} - ret = dict_set_int32 (snap_dict, "type", GF_SNAP_OPTION_TYPE_STATUS); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, - "Could not save command type"); - goto out; - } +int +cli_populate_req_dict_for_status(dict_t *snap_dict, dict_t *dict, int index) +{ + int ret = -1; + char key[PATH_MAX] = ""; + char *buffer = NULL; + + GF_ASSERT(snap_dict); + GF_ASSERT(dict); + + ret = dict_set_uint32(snap_dict, "sub-cmd", GF_SNAP_STATUS_TYPE_ITER); + if (ret) { + gf_log("cli", GF_LOG_ERROR, + "Could not save command " + "type in snap dict"); + goto out; + } + + ret = snprintf(key, sizeof(key), "status.snap%d.snapname", index); + if (ret < 0) { + goto out; + } + + ret = dict_get_str(dict, key, &buffer); + if (ret) { + gf_log("cli", GF_LOG_ERROR, "Could not get snapname"); + goto out; + } + + ret = dict_set_str(snap_dict, "snapname", buffer); + if (ret) { + gf_log("cli", GF_LOG_ERROR, + "Could not save snapname " + "in snap dict"); + goto out; + } + + ret = dict_set_int32(snap_dict, "type", GF_SNAP_OPTION_TYPE_STATUS); + if (ret) { + gf_log("cli", GF_LOG_ERROR, "Could not save command type"); + goto out; + } + + ret = dict_set_dynstr_with_alloc(snap_dict, "cmd-str", "snapshot status"); + if (ret) { + gf_log("cli", GF_LOG_ERROR, "Could not save command string as status"); + goto out; + } + + ret = dict_set_int32(snap_dict, "hold_vol_locks", _gf_false); + if (ret) { + gf_log("cli", GF_LOG_ERROR, "Setting volume lock flag failed"); + goto out; + } + +out: + return ret; +} - ret = dict_set_dynstr_with_alloc (snap_dict, "cmd-str", - "snapshot status"); +int +cli_snapshot_status(dict_t *dict, gf_cli_rsp *rsp, call_frame_t *frame) +{ + char key[PATH_MAX] = ""; + int ret = -1; + int status_cmd = -1; + cli_local_t *local = NULL; + + GF_ASSERT(dict); + GF_ASSERT(rsp); + GF_ASSERT(frame); + + local = ((call_frame_t *)frame)->local; + if (!local) { + gf_log("cli", GF_LOG_ERROR, "frame->local is NULL"); + goto out; + } + + if (rsp->op_ret) { + if (rsp->op_errstr) { + ret = dict_set_dynstr_with_alloc(local->dict, "op_err_str", + rsp->op_errstr); + if (ret) { + gf_log("cli", GF_LOG_ERROR, + "Failed to set " + "op_errstr in local dictionary"); + goto out; + } + } + ret = rsp->op_ret; + goto out; + } + + ret = dict_get_int32(dict, "sub-cmd", &status_cmd); + if (ret) { + gf_log("cli", GF_LOG_ERROR, "Could not fetch status type"); + goto out; + } + + if ((status_cmd != GF_SNAP_STATUS_TYPE_SNAP) && + (status_cmd != GF_SNAP_STATUS_TYPE_ITER)) { + dict_copy(dict, local->dict); + goto out; + } + + ret = snprintf(key, sizeof(key), "status.snap0"); + if (ret < 0) { + goto out; + } + + if (global_state->mode & GLUSTER_MODE_XML) { + ret = cli_xml_snapshot_status_single_snap(local, dict, key); if (ret) { - gf_log ("cli", GF_LOG_ERROR, - "Could not save command string as status"); - goto out; + gf_log("cli", GF_LOG_ERROR, + "Failed to create " + "xml output for snapshot status"); + goto out; } - - ret = dict_set_int32 (snap_dict, "hold_vol_locks", _gf_false); + } else { + ret = cli_get_single_snap_status(dict, key); if (ret) { - gf_log ("cli", GF_LOG_ERROR, - "Setting volume lock flag failed"); - goto out; + gf_log("cli", GF_LOG_ERROR, + "Could not fetch " + "status of snap"); + goto out; } + } + ret = 0; out: - return ret; + return ret; } int -cli_snapshot_status (dict_t *dict, gf_cli_rsp *rsp, - call_frame_t *frame) -{ - char key[PATH_MAX] = ""; - int ret = -1; - int status_cmd = -1; - cli_local_t *local = NULL; - - GF_ASSERT (dict); - GF_ASSERT (rsp); - GF_ASSERT (frame); - - local = ((call_frame_t *) frame) -> local; - if (!local) { - gf_log ("cli", GF_LOG_ERROR, "frame->local is NULL"); - goto out; - } - - if (rsp->op_ret) { - if (rsp->op_errstr) { - ret = dict_set_dynstr_with_alloc (local->dict, - "op_err_str", - rsp->op_errstr); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Failed to set " - "op_errstr in local dictionary"); - goto out; - } - } - ret = rsp->op_ret; - goto out; - } - - ret = dict_get_int32 (dict, "sub-cmd", &status_cmd); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Could not fetch status type"); +gf_cli_generate_snapshot_event(gf_cli_rsp *rsp, dict_t *dict, int32_t type, + char *snap_name, char *volname, char *snap_uuid, + char *clone_name) +{ + int ret = -1; + int config_command = 0; + int32_t delete_cmd = -1; + uint64_t hard_limit = 0; + uint64_t soft_limit = 0; + char *auto_delete = NULL; + char *snap_activate = NULL; + char msg[PATH_MAX] = { + 0, + }; + char option[512] = { + 0, + }; + + GF_VALIDATE_OR_GOTO("cli", dict, out); + GF_VALIDATE_OR_GOTO("cli", rsp, out); + + switch (type) { + case GF_SNAP_OPTION_TYPE_CREATE: + if (!snap_name) { + gf_log("cli", GF_LOG_ERROR, "Failed to get snap name"); goto out; - } + } - if ((status_cmd != GF_SNAP_STATUS_TYPE_SNAP) && - (status_cmd != GF_SNAP_STATUS_TYPE_ITER)) { - dict_copy (dict, local->dict); + if (!volname) { + gf_log("cli", GF_LOG_ERROR, "Failed to get volume name"); goto out; - } + } + if (rsp->op_ret != 0) { + gf_event(EVENT_SNAPSHOT_CREATE_FAILED, + "snapshot_name=%s;volume_name=%s;error=%s", snap_name, + volname, + rsp->op_errstr ? rsp->op_errstr + : "Please check log file for details"); + ret = 0; + break; + } - ret = snprintf (key, sizeof (key), "status.snap0"); - if (ret < 0) { + if (!snap_uuid) { + gf_log("cli", GF_LOG_ERROR, "Failed to get snap uuid"); goto out; - } + } - if (global_state->mode & GLUSTER_MODE_XML) { - ret = cli_xml_snapshot_status_single_snap (local, dict, key); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Failed to create " - "xml output for snapshot status"); - goto out; - } - } else { - ret = cli_get_single_snap_status (dict, key); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Could not fetch " - "status of snap"); - goto out; - } - } + gf_event(EVENT_SNAPSHOT_CREATED, + "snapshot_name=%s;" + "volume_name=%s;snapshot_uuid=%s", + snap_name, volname, snap_uuid); - ret = 0; -out: - return ret; -} - -int -gf_cli_generate_snapshot_event (gf_cli_rsp *rsp, dict_t *dict, - int32_t type, char *snap_name, - char *volname, char *snap_uuid, - char *clone_name) -{ - int ret = -1; - int config_command = 0; - int32_t delete_cmd = -1; - uint64_t hard_limit = 0; - uint64_t soft_limit = 0; - char *auto_delete = NULL; - char *snap_activate = NULL; - char msg[PATH_MAX] = {0, }; - char option[512] = {0, }; - - GF_VALIDATE_OR_GOTO ("cli", dict, out); - GF_VALIDATE_OR_GOTO ("cli", rsp, out); - - switch (type) { - case GF_SNAP_OPTION_TYPE_CREATE: - if (!snap_name) { - gf_log ("cli", GF_LOG_ERROR, - "Failed to get snap name"); - goto out; - } - - if (!volname) { - gf_log ("cli", GF_LOG_ERROR, - "Failed to get volume name"); - goto out; - } - - if (rsp->op_ret != 0) { - gf_event (EVENT_SNAPSHOT_CREATE_FAILED, - "snapshot_name=%s;volume_name=%s;error=%s", - snap_name, volname, - rsp->op_errstr ? rsp->op_errstr : - "Please check log file for details"); - ret = 0; - break; - } - - if (!snap_uuid) { - gf_log ("cli", GF_LOG_ERROR, "Failed to get snap uuid"); - goto out; - } + ret = 0; + break; - gf_event (EVENT_SNAPSHOT_CREATED, "snapshot_name=%s;" - "volume_name=%s;snapshot_uuid=%s", snap_name, - volname, snap_uuid); + case GF_SNAP_OPTION_TYPE_ACTIVATE: + if (!snap_name) { + gf_log("cli", GF_LOG_ERROR, "Failed to get snap name"); + goto out; + } + if (rsp->op_ret != 0) { + gf_event(EVENT_SNAPSHOT_ACTIVATE_FAILED, + "snapshot_name=%s;error=%s", snap_name, + rsp->op_errstr ? rsp->op_errstr + : "Please check log file for details"); ret = 0; break; + } - case GF_SNAP_OPTION_TYPE_ACTIVATE: - if (!snap_name) { - gf_log ("cli", GF_LOG_ERROR, - "Failed to get snap name"); - goto out; - } + if (!snap_uuid) { + gf_log("cli", GF_LOG_ERROR, "Failed to get snap uuid"); + goto out; + } - if (rsp->op_ret != 0) { - gf_event (EVENT_SNAPSHOT_ACTIVATE_FAILED, - "snapshot_name=%s;error=%s", snap_name, - rsp->op_errstr ? rsp->op_errstr : - "Please check log file for details"); - ret = 0; - break; - } + gf_event(EVENT_SNAPSHOT_ACTIVATED, + "snapshot_name=%s;" + "snapshot_uuid=%s", + snap_name, snap_uuid); - if (!snap_uuid) { - gf_log ("cli", GF_LOG_ERROR, "Failed to get snap uuid"); - goto out; - } + ret = 0; + break; - gf_event (EVENT_SNAPSHOT_ACTIVATED, "snapshot_name=%s;" - "snapshot_uuid=%s", snap_name, snap_uuid); + case GF_SNAP_OPTION_TYPE_DEACTIVATE: + if (!snap_name) { + gf_log("cli", GF_LOG_ERROR, "Failed to get snap name"); + goto out; + } + if (rsp->op_ret != 0) { + gf_event(EVENT_SNAPSHOT_DEACTIVATE_FAILED, + "snapshot_name=%s;error=%s", snap_name, + rsp->op_errstr ? rsp->op_errstr + : "Please check log file for details"); ret = 0; break; + } - case GF_SNAP_OPTION_TYPE_DEACTIVATE: - if (!snap_name) { - gf_log ("cli", GF_LOG_ERROR, - "Failed to get snap name"); - goto out; - } + if (!snap_uuid) { + gf_log("cli", GF_LOG_ERROR, "Failed to get snap uuid"); + goto out; + } - if (rsp->op_ret != 0) { - gf_event (EVENT_SNAPSHOT_DEACTIVATE_FAILED, - "snapshot_name=%s;error=%s", snap_name, - rsp->op_errstr ? rsp->op_errstr : - "Please check log file for details"); - ret = 0; - break; - } + gf_event(EVENT_SNAPSHOT_DEACTIVATED, + "snapshot_name=%s;" + "snapshot_uuid=%s", + snap_name, snap_uuid); - if (!snap_uuid) { - gf_log ("cli", GF_LOG_ERROR, "Failed to get snap uuid"); - goto out; - } + ret = 0; + break; - gf_event (EVENT_SNAPSHOT_DEACTIVATED, "snapshot_name=%s;" - "snapshot_uuid=%s", snap_name, snap_uuid); + case GF_SNAP_OPTION_TYPE_RESTORE: + if (!snap_name) { + gf_log("cli", GF_LOG_ERROR, "Failed to get snap name"); + goto out; + } + if (rsp->op_ret != 0) { + gf_event(EVENT_SNAPSHOT_RESTORE_FAILED, + "snapshot_name=%s;error=%s", snap_name, + rsp->op_errstr ? rsp->op_errstr + : "Please check log file for details"); ret = 0; break; + } - case GF_SNAP_OPTION_TYPE_RESTORE: - if (!snap_name) { - gf_log ("cli", GF_LOG_ERROR, - "Failed to get snap name"); - goto out; - } + if (!snap_uuid) { + gf_log("cli", GF_LOG_ERROR, "Failed to get snap uuid"); + goto out; + } - if (rsp->op_ret != 0) { - gf_event (EVENT_SNAPSHOT_RESTORE_FAILED, - "snapshot_name=%s;error=%s", snap_name, - rsp->op_errstr ? rsp->op_errstr : - "Please check log file for details"); - ret = 0; - break; - } + if (!volname) { + gf_log("cli", GF_LOG_ERROR, "Failed to get volname"); + goto out; + } - if (!snap_uuid) { - gf_log ("cli", GF_LOG_ERROR, "Failed to get snap uuid"); - goto out; - } + gf_event(EVENT_SNAPSHOT_RESTORED, + "snapshot_name=%s;" + "snapshot_uuid=%s;volume_name=%s", + snap_name, snap_uuid, volname); - if (!volname) { - gf_log ("cli", GF_LOG_ERROR, "Failed to get volname"); - goto out; - } + ret = 0; + break; - gf_event (EVENT_SNAPSHOT_RESTORED, "snapshot_name=%s;" - "snapshot_uuid=%s;volume_name=%s", - snap_name, snap_uuid, volname); + case GF_SNAP_OPTION_TYPE_DELETE: + ret = dict_get_int32(dict, "sub-cmd", &delete_cmd); + if (ret) { + gf_log("cli", GF_LOG_ERROR, "Could not get sub-cmd"); + goto out; + } + /* + * Need not generate any event (success or failure) for delete * + * all, as it will trigger individual delete for all snapshots * + */ + if (delete_cmd == GF_SNAP_DELETE_TYPE_ALL) { ret = 0; break; + } - case GF_SNAP_OPTION_TYPE_DELETE: - ret = dict_get_int32 (dict, "sub-cmd", &delete_cmd); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Could not get sub-cmd"); - goto out; - } + if (!snap_name) { + gf_log("cli", GF_LOG_ERROR, "Failed to get snap name"); + goto out; + } - /* - * Need not generate any event (success or failure) for delete * - * all, as it will trigger individual delete for all snapshots * - */ - if (delete_cmd == GF_SNAP_DELETE_TYPE_ALL) { - ret = 0; - break; - } + if (rsp->op_ret != 0) { + gf_event(EVENT_SNAPSHOT_DELETE_FAILED, + "snapshot_name=%s;error=%s", snap_name, + rsp->op_errstr ? rsp->op_errstr + : "Please check log file for details"); + ret = 0; + break; + } - if (!snap_name) { - gf_log ("cli", GF_LOG_ERROR, - "Failed to get snap name"); - goto out; - } + if (!snap_uuid) { + gf_log("cli", GF_LOG_ERROR, "Failed to get snap uuid"); + goto out; + } - if (rsp->op_ret != 0) { - gf_event (EVENT_SNAPSHOT_DELETE_FAILED, - "snapshot_name=%s;error=%s", snap_name, - rsp->op_errstr ? rsp->op_errstr : - "Please check log file for details"); - ret = 0; - break; - } + gf_event(EVENT_SNAPSHOT_DELETED, + "snapshot_name=%s;" + "snapshot_uuid=%s", + snap_name, snap_uuid); - if (!snap_uuid) { - gf_log ("cli", GF_LOG_ERROR, "Failed to get snap uuid"); - goto out; - } + ret = 0; + break; - gf_event (EVENT_SNAPSHOT_DELETED, "snapshot_name=%s;" - "snapshot_uuid=%s", snap_name, snap_uuid); + case GF_SNAP_OPTION_TYPE_CLONE: + if (!clone_name) { + gf_log("cli", GF_LOG_ERROR, "Failed to get clone name"); + goto out; + } + + if (!snap_name) { + gf_log("cli", GF_LOG_ERROR, "Failed to get snapname name"); + goto out; + } + if (rsp->op_ret != 0) { + gf_event(EVENT_SNAPSHOT_CLONE_FAILED, + "snapshot_name=%s;clone_name=%s;" + "error=%s", + snap_name, clone_name, + rsp->op_errstr ? rsp->op_errstr + : "Please check log file for details"); ret = 0; break; + } - case GF_SNAP_OPTION_TYPE_CLONE: - if (!clone_name) { - gf_log ("cli", GF_LOG_ERROR, - "Failed to get clone name"); - goto out; - } + if (!snap_uuid) { + gf_log("cli", GF_LOG_ERROR, "Failed to get snap uuid"); + goto out; + } - if (!snap_name) { - gf_log ("cli", GF_LOG_ERROR, - "Failed to get snapname name"); - goto out; - } + gf_event(EVENT_SNAPSHOT_CLONED, + "snapshot_name=%s;" + "clone_name=%s;clone_uuid=%s", + snap_name, clone_name, snap_uuid); - if (rsp->op_ret != 0) { - gf_event (EVENT_SNAPSHOT_CLONE_FAILED, - "snapshot_name=%s;clone_name=%s;" - "error=%s", snap_name, clone_name, - rsp->op_errstr ? rsp->op_errstr : - "Please check log file for details"); - ret = 0; - break; - } + ret = 0; + break; - if (!snap_uuid) { - gf_log ("cli", GF_LOG_ERROR, "Failed to get snap uuid"); - goto out; - } + case GF_SNAP_OPTION_TYPE_CONFIG: + if (rsp->op_ret != 0) { + gf_event(EVENT_SNAPSHOT_CONFIG_UPDATE_FAILED, "error=%s", + rsp->op_errstr ? rsp->op_errstr + : "Please check log file for details"); + ret = 0; + break; + } - gf_event (EVENT_SNAPSHOT_CLONED, "snapshot_name=%s;" - "clone_name=%s;clone_uuid=%s", - snap_name, clone_name, snap_uuid); + ret = dict_get_int32(dict, "config-command", &config_command); + if (ret) { + gf_log("cli", GF_LOG_ERROR, "Could not fetch config type"); + goto out; + } + if (config_command == GF_SNAP_CONFIG_DISPLAY) { ret = 0; break; + } - case GF_SNAP_OPTION_TYPE_CONFIG: - if (rsp->op_ret != 0) { - gf_event (EVENT_SNAPSHOT_CONFIG_UPDATE_FAILED, - "error=%s", - rsp->op_errstr ? rsp->op_errstr : - "Please check log file for details"); - ret = 0; - break; - } + /* These are optional parameters therefore ignore the error */ + ret = dict_get_uint64(dict, "snap-max-hard-limit", &hard_limit); + ret = dict_get_uint64(dict, "snap-max-soft-limit", &soft_limit); + ret = dict_get_str(dict, "auto-delete", &auto_delete); + ret = dict_get_str(dict, "snap-activate-on-create", &snap_activate); - ret = dict_get_int32 (dict, "config-command", &config_command); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, - "Could not fetch config type"); - goto out; - } - - if (config_command == GF_SNAP_CONFIG_DISPLAY) { - ret = 0; - break; - } + if (!hard_limit && !soft_limit && !auto_delete && !snap_activate) { + ret = -1; + gf_log("cli", GF_LOG_ERROR, + "At least one option from " + "snap-max-hard-limit, snap-max-soft-limit, " + "auto-delete and snap-activate-on-create " + "should be set"); + goto out; + } - /* These are optional parameters therefore ignore the error */ - ret = dict_get_uint64 (dict, "snap-max-hard-limit", - &hard_limit); - ret = dict_get_uint64 (dict, "snap-max-soft-limit", - &soft_limit); - ret = dict_get_str (dict, "auto-delete", - &auto_delete); - ret = dict_get_str (dict, "snap-activate-on-create", - &snap_activate); - - if (!hard_limit && !soft_limit && - !auto_delete && !snap_activate) { - ret = -1; - gf_log ("cli", GF_LOG_ERROR, "At least one option from " - "snap-max-hard-limit, snap-max-soft-limit, " - "auto-delete and snap-activate-on-create " - "should be set"); - goto out; - } + volname = NULL; + ret = dict_get_str(dict, "volname", &volname); - volname = NULL; - ret = dict_get_str (dict, "volname", &volname); - - if (hard_limit || soft_limit) { - snprintf (option, sizeof(option), "%s=%"PRIu64, - hard_limit ? "hard_limit" : "soft_limit", - hard_limit ? hard_limit:soft_limit); - } else if (auto_delete || snap_activate) { - snprintf (option, sizeof(option), "%s=%s", - auto_delete ? "auto-delete" : "snap-activate", - auto_delete ? auto_delete:snap_activate); - } + if (hard_limit || soft_limit) { + snprintf(option, sizeof(option), "%s=%" PRIu64, + hard_limit ? "hard_limit" : "soft_limit", + hard_limit ? hard_limit : soft_limit); + } else if (auto_delete || snap_activate) { + snprintf(option, sizeof(option), "%s=%s", + auto_delete ? "auto-delete" : "snap-activate", + auto_delete ? auto_delete : snap_activate); + } - snprintf (msg, sizeof(msg), "config_type=%s;%s", - volname?"volume_config":"system_config", option); + snprintf(msg, sizeof(msg), "config_type=%s;%s", + volname ? "volume_config" : "system_config", option); - gf_event (EVENT_SNAPSHOT_CONFIG_UPDATED, "%s", msg); + gf_event(EVENT_SNAPSHOT_CONFIG_UPDATED, "%s", msg); - ret = 0; - break; + ret = 0; + break; default: - gf_log ("cli", GF_LOG_WARNING, - "Cannot generate event for unknown type."); - ret = 0; - goto out; - } + gf_log("cli", GF_LOG_WARNING, + "Cannot generate event for unknown type."); + ret = 0; + goto out; + } out: - return ret; + return ret; } /* @@ -10766,1415 +10684,1463 @@ out: * repeating the same code again and again. * */ int -gf_cli_snapshot_get_data_from_dict (dict_t *dict, char **snap_name, - char **volname, char **snap_uuid, - int8_t *soft_limit_flag, - char **clone_name) +gf_cli_snapshot_get_data_from_dict(dict_t *dict, char **snap_name, + char **volname, char **snap_uuid, + int8_t *soft_limit_flag, char **clone_name) { - int ret = -1; + int ret = -1; - GF_VALIDATE_OR_GOTO ("cli", dict, out); + GF_VALIDATE_OR_GOTO("cli", dict, out); - if (snap_name) { - ret = dict_get_str (dict, "snapname", snap_name); - if (ret) { - gf_log ("cli", GF_LOG_DEBUG, - "failed to get snapname from dict"); - } + if (snap_name) { + ret = dict_get_str(dict, "snapname", snap_name); + if (ret) { + gf_log("cli", GF_LOG_DEBUG, "failed to get snapname from dict"); } + } - if (volname) { - ret = dict_get_str (dict, "volname1", volname); - if (ret) { - gf_log ("cli", GF_LOG_DEBUG, - "failed to get volname1 from dict"); - } + if (volname) { + ret = dict_get_str(dict, "volname1", volname); + if (ret) { + gf_log("cli", GF_LOG_DEBUG, "failed to get volname1 from dict"); } + } - if (snap_uuid) { - ret = dict_get_str (dict, "snapuuid", snap_uuid); - if (ret) { - gf_log ("cli", GF_LOG_DEBUG, - "failed to get snapuuid from dict"); - } + if (snap_uuid) { + ret = dict_get_str(dict, "snapuuid", snap_uuid); + if (ret) { + gf_log("cli", GF_LOG_DEBUG, "failed to get snapuuid from dict"); } + } - if (soft_limit_flag) { - ret = dict_get_int8 (dict, "soft-limit-reach", - soft_limit_flag); - if (ret) { - gf_log ("cli", GF_LOG_DEBUG, - "failed to get soft-limit-reach from dict"); - } + if (soft_limit_flag) { + ret = dict_get_int8(dict, "soft-limit-reach", soft_limit_flag); + if (ret) { + gf_log("cli", GF_LOG_DEBUG, + "failed to get soft-limit-reach from dict"); } + } - if (clone_name) { - ret = dict_get_str (dict, "clonename", clone_name); - if (ret) { - gf_log ("cli", GF_LOG_DEBUG, - "failed to get clonename from dict"); - } + if (clone_name) { + ret = dict_get_str(dict, "clonename", clone_name); + if (ret) { + gf_log("cli", GF_LOG_DEBUG, "failed to get clonename from dict"); } + } - ret = 0; + ret = 0; out: - return ret; + return ret; } int -gf_cli_snapshot_cbk (struct rpc_req *req, struct iovec *iov, - int count, void *myframe) -{ - int ret = -1; - gf_cli_rsp rsp = {0, }; - dict_t *dict = NULL; - char *snap_name = NULL; - char *clone_name = NULL; - int32_t type = 0; - call_frame_t *frame = NULL; - gf_boolean_t snap_driven = _gf_false; - int8_t soft_limit_flag = -1; - char *volname = NULL; - char *snap_uuid = NULL; - - GF_ASSERT (myframe); - - if (req->rpc_status == -1) { - goto out; - } - - frame = myframe; - - ret = xdr_to_generic (*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); - if (ret < 0) { - gf_log (frame->this->name, GF_LOG_ERROR, - "Failed to decode xdr response"); - goto out; - } +gf_cli_snapshot_cbk(struct rpc_req *req, struct iovec *iov, int count, + void *myframe) +{ + int ret = -1; + gf_cli_rsp rsp = { + 0, + }; + dict_t *dict = NULL; + char *snap_name = NULL; + char *clone_name = NULL; + int32_t type = 0; + call_frame_t *frame = NULL; + gf_boolean_t snap_driven = _gf_false; + int8_t soft_limit_flag = -1; + char *volname = NULL; + char *snap_uuid = NULL; + + GF_ASSERT(myframe); + + if (req->rpc_status == -1) { + goto out; + } + + frame = myframe; + + ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); + if (ret < 0) { + gf_log(frame->this->name, GF_LOG_ERROR, + "Failed to decode xdr response"); + goto out; + } + + dict = dict_new(); + + if (!dict) { + ret = -1; + goto out; + } + + ret = dict_unserialize(rsp.dict.dict_val, rsp.dict.dict_len, &dict); + + if (ret) + goto out; + + ret = dict_get_int32(dict, "type", &type); + if (ret) { + gf_log(frame->this->name, GF_LOG_ERROR, "failed to get type"); + goto out; + } + + ret = gf_cli_snapshot_get_data_from_dict( + dict, &snap_name, &volname, &snap_uuid, &soft_limit_flag, &clone_name); + if (ret) { + gf_log("cli", GF_LOG_ERROR, "Failed to fetch data from dict."); + goto out; + } - dict = dict_new (); - - if (!dict) { - ret = -1; - goto out; - } - - ret = dict_unserialize (rsp.dict.dict_val, rsp.dict.dict_len, &dict); - - if (ret) - goto out; +#if (USE_EVENTS) + ret = gf_cli_generate_snapshot_event(&rsp, dict, type, snap_name, volname, + snap_uuid, clone_name); + if (ret) { + gf_log("cli", GF_LOG_ERROR, "Failed to generate snapshot event"); + goto out; + } +#endif - ret = dict_get_int32 (dict, "type", &type); + /* Snapshot status and delete command is handled separately */ + if (global_state->mode & GLUSTER_MODE_XML && + GF_SNAP_OPTION_TYPE_STATUS != type && + GF_SNAP_OPTION_TYPE_DELETE != type) { + ret = cli_xml_output_snapshot(type, dict, rsp.op_ret, rsp.op_errno, + rsp.op_errstr); if (ret) { - gf_log (frame->this->name, GF_LOG_ERROR, "failed to get type"); - goto out; + gf_log("cli", GF_LOG_ERROR, "Error outputting to xml"); } - ret = gf_cli_snapshot_get_data_from_dict (dict, &snap_name, &volname, - &snap_uuid, &soft_limit_flag, - &clone_name); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Failed to fetch data from dict."); - goto out; - } + goto out; + } -#if (USE_EVENTS) - ret = gf_cli_generate_snapshot_event (&rsp, dict, type, snap_name, - volname, snap_uuid, clone_name); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, - "Failed to generate snapshot event"); + switch (type) { + case GF_SNAP_OPTION_TYPE_CREATE: + if (rsp.op_ret) { + cli_err("snapshot create: failed: %s", + rsp.op_errstr ? rsp.op_errstr + : "Please check log file for details"); + ret = rsp.op_ret; goto out; - } -#endif - - /* Snapshot status and delete command is handled separately */ - if (global_state->mode & GLUSTER_MODE_XML && - GF_SNAP_OPTION_TYPE_STATUS != type && - GF_SNAP_OPTION_TYPE_DELETE != type) { - ret = cli_xml_output_snapshot (type, dict, rsp.op_ret, - rsp.op_errno, rsp.op_errstr); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Error outputting to xml"); - } + } + if (!snap_name) { + gf_log("cli", GF_LOG_ERROR, "Failed to get snap name"); goto out; - } + } - switch (type) { - case GF_SNAP_OPTION_TYPE_CREATE: - if (rsp.op_ret) { - cli_err("snapshot create: failed: %s", - rsp.op_errstr ? rsp.op_errstr : - "Please check log file for details"); - ret = rsp.op_ret; - goto out; - } - - if (!snap_name) { - gf_log ("cli", GF_LOG_ERROR, - "Failed to get snap name"); - goto out; - } - - if (!volname) { - gf_log ("cli", GF_LOG_ERROR, - "Failed to get volume name"); - goto out; - } + if (!volname) { + gf_log("cli", GF_LOG_ERROR, "Failed to get volume name"); + goto out; + } - cli_out ("snapshot create: success: Snap %s created " - "successfully", snap_name); + cli_out( + "snapshot create: success: Snap %s created " + "successfully", + snap_name); - if (soft_limit_flag == 1) { - cli_out ("Warning: Soft-limit of volume (%s) is " - "reached. Snapshot creation is not possible " - "once hard-limit is reached.", volname); - } - ret = 0; - break; + if (soft_limit_flag == 1) { + cli_out( + "Warning: Soft-limit of volume (%s) is " + "reached. Snapshot creation is not possible " + "once hard-limit is reached.", + volname); + } + ret = 0; + break; case GF_SNAP_OPTION_TYPE_CLONE: - if (rsp.op_ret) { - cli_err("snapshot clone: failed: %s", - rsp.op_errstr ? rsp.op_errstr : - "Please check log file for details"); - ret = rsp.op_ret; - goto out; - } + if (rsp.op_ret) { + cli_err("snapshot clone: failed: %s", + rsp.op_errstr ? rsp.op_errstr + : "Please check log file for details"); + ret = rsp.op_ret; + goto out; + } - if (!clone_name) { - gf_log ("cli", GF_LOG_ERROR, - "Failed to get clone name"); - goto out; - } + if (!clone_name) { + gf_log("cli", GF_LOG_ERROR, "Failed to get clone name"); + goto out; + } - if (!snap_name) { - gf_log ("cli", GF_LOG_ERROR, - "Failed to get snapname name"); - goto out; - } + if (!snap_name) { + gf_log("cli", GF_LOG_ERROR, "Failed to get snapname name"); + goto out; + } - cli_out ("snapshot clone: success: Clone %s created " - "successfully", clone_name); + cli_out( + "snapshot clone: success: Clone %s created " + "successfully", + clone_name); - ret = 0; - break; + ret = 0; + break; case GF_SNAP_OPTION_TYPE_RESTORE: - if (rsp.op_ret) { - cli_err("snapshot restore: failed: %s", - rsp.op_errstr ? rsp.op_errstr : - "Please check log file for details"); - ret = rsp.op_ret; - goto out; - } + if (rsp.op_ret) { + cli_err("snapshot restore: failed: %s", + rsp.op_errstr ? rsp.op_errstr + : "Please check log file for details"); + ret = rsp.op_ret; + goto out; + } - if (!snap_name) { - gf_log ("cli", GF_LOG_ERROR, - "Failed to get snap name"); - goto out; - } + if (!snap_name) { + gf_log("cli", GF_LOG_ERROR, "Failed to get snap name"); + goto out; + } - cli_out ("Snapshot restore: %s: Snap restored " - "successfully", snap_name); + cli_out( + "Snapshot restore: %s: Snap restored " + "successfully", + snap_name); - ret = 0; - break; + ret = 0; + break; case GF_SNAP_OPTION_TYPE_ACTIVATE: - if (rsp.op_ret) { - cli_err("snapshot activate: failed: %s", - rsp.op_errstr ? rsp.op_errstr : - "Please check log file for details"); - ret = rsp.op_ret; - goto out; - } + if (rsp.op_ret) { + cli_err("snapshot activate: failed: %s", + rsp.op_errstr ? rsp.op_errstr + : "Please check log file for details"); + ret = rsp.op_ret; + goto out; + } - if (!snap_name) { - gf_log ("cli", GF_LOG_ERROR, - "Failed to get snap name"); - goto out; - } + if (!snap_name) { + gf_log("cli", GF_LOG_ERROR, "Failed to get snap name"); + goto out; + } - cli_out ("Snapshot activate: %s: Snap activated " - "successfully", snap_name); + cli_out( + "Snapshot activate: %s: Snap activated " + "successfully", + snap_name); - ret = 0; - break; + ret = 0; + break; case GF_SNAP_OPTION_TYPE_DEACTIVATE: - if (rsp.op_ret) { - cli_err("snapshot deactivate: failed: %s", - rsp.op_errstr ? rsp.op_errstr : - "Please check log file for details"); - ret = rsp.op_ret; - goto out; - } + if (rsp.op_ret) { + cli_err("snapshot deactivate: failed: %s", + rsp.op_errstr ? rsp.op_errstr + : "Please check log file for details"); + ret = rsp.op_ret; + goto out; + } - if (!snap_name) { - gf_log ("cli", GF_LOG_ERROR, - "Failed to get snap name"); - goto out; - } + if (!snap_name) { + gf_log("cli", GF_LOG_ERROR, "Failed to get snap name"); + goto out; + } - cli_out ("Snapshot deactivate: %s: Snap deactivated " - "successfully", snap_name); + cli_out( + "Snapshot deactivate: %s: Snap deactivated " + "successfully", + snap_name); - ret = 0; - break; + ret = 0; + break; case GF_SNAP_OPTION_TYPE_INFO: - if (rsp.op_ret) { - cli_err ("Snapshot info : failed: %s", - rsp.op_errstr ? rsp.op_errstr : - "Please check log file for details"); - ret = rsp.op_ret; - goto out; - } + if (rsp.op_ret) { + cli_err("Snapshot info : failed: %s", + rsp.op_errstr ? rsp.op_errstr + : "Please check log file for details"); + ret = rsp.op_ret; + goto out; + } - snap_driven = dict_get_str_boolean (dict, "snap-driven", - _gf_false); - if (snap_driven == _gf_true) { - ret = cli_call_snapshot_info (dict, snap_driven); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, - "Snapshot info failed"); - goto out; - } - } else if (snap_driven == _gf_false) { - ret = cli_get_snaps_in_volume (dict); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, - "Snapshot info failed"); - goto out; - } + snap_driven = dict_get_str_boolean(dict, "snap-driven", _gf_false); + if (snap_driven == _gf_true) { + ret = cli_call_snapshot_info(dict, snap_driven); + if (ret) { + gf_log("cli", GF_LOG_ERROR, "Snapshot info failed"); + goto out; } - break; - - case GF_SNAP_OPTION_TYPE_CONFIG: - ret = cli_snapshot_config_display (dict, &rsp); + } else if (snap_driven == _gf_false) { + ret = cli_get_snaps_in_volume(dict); if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Failed to display " - "snapshot config output."); - goto out; + gf_log("cli", GF_LOG_ERROR, "Snapshot info failed"); + goto out; } - break; + } + break; + + case GF_SNAP_OPTION_TYPE_CONFIG: + ret = cli_snapshot_config_display(dict, &rsp); + if (ret) { + gf_log("cli", GF_LOG_ERROR, + "Failed to display " + "snapshot config output."); + goto out; + } + break; case GF_SNAP_OPTION_TYPE_LIST: - if (rsp.op_ret) { - cli_err ("Snapshot list : failed: %s", - rsp.op_errstr ? rsp.op_errstr : - "Please check log file for details"); - ret = rsp.op_ret; - goto out; - } + if (rsp.op_ret) { + cli_err("Snapshot list : failed: %s", + rsp.op_errstr ? rsp.op_errstr + : "Please check log file for details"); + ret = rsp.op_ret; + goto out; + } - ret = cli_snapshot_list (dict); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Failed to display " - "snapshot list"); - goto out; - } - break; + ret = cli_snapshot_list(dict); + if (ret) { + gf_log("cli", GF_LOG_ERROR, + "Failed to display " + "snapshot list"); + goto out; + } + break; case GF_SNAP_OPTION_TYPE_DELETE: - ret = cli_snapshot_remove_reply (&rsp, dict, frame); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, - "Failed to delete snap"); - goto out; - } - break; + ret = cli_snapshot_remove_reply(&rsp, dict, frame); + if (ret) { + gf_log("cli", GF_LOG_ERROR, "Failed to delete snap"); + goto out; + } + break; case GF_SNAP_OPTION_TYPE_STATUS: - ret = cli_snapshot_status (dict, &rsp, frame); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Failed to display " - "snapshot status output."); - goto out; - } - break; + ret = cli_snapshot_status(dict, &rsp, frame); + if (ret) { + gf_log("cli", GF_LOG_ERROR, + "Failed to display " + "snapshot status output."); + goto out; + } + break; default: - cli_err ("Unknown command executed"); - ret = -1; - goto out; - } + cli_err("Unknown command executed"); + ret = -1; + goto out; + } out: - if (dict) - dict_unref (dict); - cli_cmd_broadcast_response (ret); + if (dict) + dict_unref(dict); + cli_cmd_broadcast_response(ret); - free (rsp.dict.dict_val); - free (rsp.op_errstr); + free(rsp.dict.dict_val); + free(rsp.op_errstr); - return ret; + return ret; } int32_t -gf_cli_snapshot_for_delete (call_frame_t *frame, xlator_t *this, - void *data) -{ - gf_cli_req req = {{0,}}; - int32_t ret = -1; - int32_t cmd = -1; - cli_local_t *local = NULL; - dict_t *snap_dict = NULL; - int32_t snapcount = 0; - int i = 0; - char question[PATH_MAX] = ""; - char *volname = NULL; - gf_answer_t answer = GF_ANSWER_NO; - - GF_VALIDATE_OR_GOTO ("cli", frame, out); - GF_VALIDATE_OR_GOTO ("cli", frame->local, out); - GF_VALIDATE_OR_GOTO ("cli", this, out); - GF_VALIDATE_OR_GOTO ("cli", data, out); - - local = frame->local; - - ret = dict_get_int32 (local->dict, "sub-cmd", &cmd); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Failed to get " - "sub-cmd"); - goto out; - } - - /* No need multiple RPCs for individual snapshot delete*/ - if (cmd == GF_SNAP_DELETE_TYPE_SNAP) { - ret = 0; - goto out; - } - - ret = dict_get_int32 (local->dict, "snapcount", - &snapcount); +gf_cli_snapshot_for_delete(call_frame_t *frame, xlator_t *this, void *data) +{ + gf_cli_req req = {{ + 0, + }}; + int32_t ret = -1; + int32_t cmd = -1; + cli_local_t *local = NULL; + dict_t *snap_dict = NULL; + int32_t snapcount = 0; + int i = 0; + char question[PATH_MAX] = ""; + char *volname = NULL; + gf_answer_t answer = GF_ANSWER_NO; + + GF_VALIDATE_OR_GOTO("cli", frame, out); + GF_VALIDATE_OR_GOTO("cli", frame->local, out); + GF_VALIDATE_OR_GOTO("cli", this, out); + GF_VALIDATE_OR_GOTO("cli", data, out); + + local = frame->local; + + ret = dict_get_int32(local->dict, "sub-cmd", &cmd); + if (ret) { + gf_log("cli", GF_LOG_ERROR, + "Failed to get " + "sub-cmd"); + goto out; + } + + /* No need multiple RPCs for individual snapshot delete*/ + if (cmd == GF_SNAP_DELETE_TYPE_SNAP) { + ret = 0; + goto out; + } + + ret = dict_get_int32(local->dict, "snapcount", &snapcount); + if (ret) { + gf_log("cli", GF_LOG_ERROR, + "Could not get " + "snapcount"); + goto out; + } + + if (global_state->mode & GLUSTER_MODE_XML) { +#ifdef HAVE_LIB_XML + ret = xmlTextWriterWriteFormatElement( + local->writer, (xmlChar *)"snapCount", "%d", snapcount); if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Could not get " - "snapcount"); - goto out; + gf_log("cli", GF_LOG_ERROR, + "Failed to write " + "xml element \"snapCount\""); + goto out; } - - if (global_state->mode & GLUSTER_MODE_XML) { -#ifdef HAVE_LIB_XML - ret = xmlTextWriterWriteFormatElement (local->writer, - (xmlChar *)"snapCount", - "%d", snapcount); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Failed to write " - "xml element \"snapCount\""); - goto out; - } #endif /* HAVE_LIB_XML */ - } else if (snapcount == 0) { - cli_out ("No snapshots present"); - goto out; + } else if (snapcount == 0) { + cli_out("No snapshots present"); + goto out; + } + + if (cmd == GF_SNAP_DELETE_TYPE_ALL) { + snprintf(question, sizeof(question), + "System contains %d " + "snapshot(s).\nDo you still " + "want to continue and delete them? ", + snapcount); + } else { + ret = dict_get_str(local->dict, "volname", &volname); + if (ret) { + gf_log("cli", GF_LOG_ERROR, + "Failed to fetch " + "volname from local dictionary"); + goto out; } - if (cmd == GF_SNAP_DELETE_TYPE_ALL) { - snprintf (question, sizeof (question), "System contains %d " - "snapshot(s).\nDo you still " - "want to continue and delete them? ", - snapcount); - } else { - ret = dict_get_str (local->dict, "volname", &volname); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Failed to fetch " - "volname from local dictionary"); - goto out; - } + snprintf(question, sizeof(question), + "Volume (%s) contains " + "%d snapshot(s).\nDo you still want to " + "continue and delete them? ", + volname, snapcount); + } - snprintf (question, sizeof (question), "Volume (%s) contains " - "%d snapshot(s).\nDo you still want to " - "continue and delete them? ", volname, - snapcount); - } - - answer = cli_cmd_get_confirmation (global_state, question); - if (GF_ANSWER_NO == answer) { - ret = 0; - gf_log ("cli", GF_LOG_DEBUG, "User cancelled " - "snapshot delete operation for snap delete"); - goto out; - } + answer = cli_cmd_get_confirmation(global_state, question); + if (GF_ANSWER_NO == answer) { + ret = 0; + gf_log("cli", GF_LOG_DEBUG, + "User cancelled " + "snapshot delete operation for snap delete"); + goto out; + } - for (i = 1 ; i <= snapcount ; i++) { - ret = -1; + for (i = 1; i <= snapcount; i++) { + ret = -1; - snap_dict = dict_new(); - if (!snap_dict) - goto out; + snap_dict = dict_new(); + if (!snap_dict) + goto out; - ret = cli_populate_req_dict_for_delete (snap_dict, - local->dict, i); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Could not " - "populate snap request dictionary"); - goto out; - } + ret = cli_populate_req_dict_for_delete(snap_dict, local->dict, i); + if (ret) { + gf_log("cli", GF_LOG_ERROR, + "Could not " + "populate snap request dictionary"); + goto out; + } - ret = cli_to_glusterd (&req, frame, - gf_cli_snapshot_cbk, - (xdrproc_t) xdr_gf_cli_req, snap_dict, - GLUSTER_CLI_SNAP, this, cli_rpc_prog, - NULL); - if (ret) { - /* Fail the operation if deleting one of the - * snapshots is failed - */ - gf_log ("cli", GF_LOG_ERROR, "cli_to_glusterd " - "for snapshot delete failed"); - goto out; - } - dict_unref (snap_dict); - snap_dict = NULL; + ret = cli_to_glusterd(&req, frame, gf_cli_snapshot_cbk, + (xdrproc_t)xdr_gf_cli_req, snap_dict, + GLUSTER_CLI_SNAP, this, cli_rpc_prog, NULL); + if (ret) { + /* Fail the operation if deleting one of the + * snapshots is failed + */ + gf_log("cli", GF_LOG_ERROR, + "cli_to_glusterd " + "for snapshot delete failed"); + goto out; } + dict_unref(snap_dict); + snap_dict = NULL; + } out: - if (snap_dict) - dict_unref (snap_dict); + if (snap_dict) + dict_unref(snap_dict); - return ret; + return ret; } int32_t -gf_cli_snapshot_for_status (call_frame_t *frame, xlator_t *this, - void *data) -{ - - gf_cli_req req = {{0,}}; - int ret = -1; - int32_t cmd = -1; - cli_local_t *local = NULL; - dict_t *snap_dict = NULL; - int snapcount = 0; - int i = 0; +gf_cli_snapshot_for_status(call_frame_t *frame, xlator_t *this, void *data) +{ + gf_cli_req req = {{ + 0, + }}; + int ret = -1; + int32_t cmd = -1; + cli_local_t *local = NULL; + dict_t *snap_dict = NULL; + int snapcount = 0; + int i = 0; + + GF_VALIDATE_OR_GOTO("cli", frame, out); + GF_VALIDATE_OR_GOTO("cli", frame->local, out); + GF_VALIDATE_OR_GOTO("cli", this, out); + GF_VALIDATE_OR_GOTO("cli", data, out); + + local = frame->local; + + ret = dict_get_int32(local->dict, "sub-cmd", &cmd); + if (ret) { + gf_log("cli", GF_LOG_ERROR, "Failed to get sub-cmd"); + goto out; + } + + /* Snapshot status of single snap (i.e. GF_SNAP_STATUS_TYPE_SNAP) + * is already handled. Therefore we can return from here. + * If want to get status of all snaps in the system or volume then + * we should get them one by one.*/ + if ((cmd == GF_SNAP_STATUS_TYPE_SNAP) || + (cmd == GF_SNAP_STATUS_TYPE_ITER)) { + ret = 0; + goto out; + } - GF_VALIDATE_OR_GOTO ("cli", frame, out); - GF_VALIDATE_OR_GOTO ("cli", frame->local, out); - GF_VALIDATE_OR_GOTO ("cli", this, out); - GF_VALIDATE_OR_GOTO ("cli", data, out); + ret = dict_get_int32(local->dict, "status.snapcount", &snapcount); + if (ret) { + gf_log("cli", GF_LOG_ERROR, "Could not get snapcount"); + goto out; + } - local = frame->local; + if (snapcount == 0 && !(global_state->mode & GLUSTER_MODE_XML)) { + cli_out("No snapshots present"); + } - ret = dict_get_int32 (local->dict, "sub-cmd", &cmd); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Failed to get sub-cmd"); - goto out; - } + for (i = 0; i < snapcount; i++) { + ret = -1; - /* Snapshot status of single snap (i.e. GF_SNAP_STATUS_TYPE_SNAP) - * is already handled. Therefore we can return from here. - * If want to get status of all snaps in the system or volume then - * we should get them one by one.*/ - if ((cmd == GF_SNAP_STATUS_TYPE_SNAP) || - (cmd == GF_SNAP_STATUS_TYPE_ITER)) { - ret = 0; - goto out; - } + snap_dict = dict_new(); + if (!snap_dict) + goto out; - ret = dict_get_int32 (local->dict, "status.snapcount", &snapcount); + ret = cli_populate_req_dict_for_status(snap_dict, local->dict, i); if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Could not get snapcount"); - goto out; + gf_log("cli", GF_LOG_ERROR, + "Could not " + "populate snap request dictionary"); + goto out; } - if (snapcount == 0 && !(global_state->mode & GLUSTER_MODE_XML)) { - cli_out ("No snapshots present"); - } + ret = cli_to_glusterd(&req, frame, gf_cli_snapshot_cbk, + (xdrproc_t)xdr_gf_cli_req, snap_dict, + GLUSTER_CLI_SNAP, this, cli_rpc_prog, NULL); - for (i = 0 ; i < snapcount; i++) { - ret = -1; - - snap_dict = dict_new(); - if (!snap_dict) - goto out; - - ret = cli_populate_req_dict_for_status (snap_dict, - local->dict, i); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Could not " - "populate snap request dictionary"); - goto out; - } - - ret = cli_to_glusterd (&req, frame, - gf_cli_snapshot_cbk, - (xdrproc_t) xdr_gf_cli_req, snap_dict, - GLUSTER_CLI_SNAP, this, cli_rpc_prog, - NULL); - - /* Ignore the return value and error for snapshot - * status of type "ALL" or "VOL" - * - * Scenario : There might be case where status command - * and delete command might be issued at the same time. - * In that case when status tried to fetch detail of - * snap which has been deleted by concurrent command, - * then it will show snapshot not present. Which will - * not be appropriate. - */ - if (ret && - (cmd != GF_SNAP_STATUS_TYPE_ALL && - cmd != GF_SNAP_STATUS_TYPE_VOL)) { - gf_log ("cli", GF_LOG_ERROR, - "cli_to_glusterd for snapshot status failed"); - goto out; - } - dict_unref (snap_dict); - snap_dict = NULL; + /* Ignore the return value and error for snapshot + * status of type "ALL" or "VOL" + * + * Scenario : There might be case where status command + * and delete command might be issued at the same time. + * In that case when status tried to fetch detail of + * snap which has been deleted by concurrent command, + * then it will show snapshot not present. Which will + * not be appropriate. + */ + if (ret && (cmd != GF_SNAP_STATUS_TYPE_ALL && + cmd != GF_SNAP_STATUS_TYPE_VOL)) { + gf_log("cli", GF_LOG_ERROR, + "cli_to_glusterd for snapshot status failed"); + goto out; } + dict_unref(snap_dict); + snap_dict = NULL; + } - ret = 0; + ret = 0; out: - if (snap_dict) - dict_unref (snap_dict); + if (snap_dict) + dict_unref(snap_dict); - return ret; + return ret; } int32_t -gf_cli_snapshot (call_frame_t *frame, xlator_t *this, - void *data) -{ - gf_cli_req req = {{0,}}; - dict_t *options = NULL; - int ret = -1; - int tmp_ret = -1; - cli_local_t *local = NULL; - char *err_str = NULL; - int type = -1; - - if (!frame || !this || !data) - goto out; - - if (!frame->local) - goto out; +gf_cli_snapshot(call_frame_t *frame, xlator_t *this, void *data) +{ + gf_cli_req req = {{ + 0, + }}; + dict_t *options = NULL; + int ret = -1; + int tmp_ret = -1; + cli_local_t *local = NULL; + char *err_str = NULL; + int type = -1; - local = frame->local; + if (!frame || !this || !data) + goto out; - options = data; + if (!frame->local) + goto out; - ret = dict_get_int32 (local->dict, "type", &type); + local = frame->local; - if (global_state->mode & GLUSTER_MODE_XML) { - ret = cli_xml_snapshot_begin_composite_op (local); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Failed to begin " - "snapshot xml composite op"); - goto out; - } - } + options = data; + ret = dict_get_int32(local->dict, "type", &type); - ret = cli_to_glusterd (&req, frame, gf_cli_snapshot_cbk, - (xdrproc_t) xdr_gf_cli_req, options, - GLUSTER_CLI_SNAP, this, cli_rpc_prog, - NULL); + if (global_state->mode & GLUSTER_MODE_XML) { + ret = cli_xml_snapshot_begin_composite_op(local); if (ret) { - gf_log ("cli", GF_LOG_ERROR, "cli_to_glusterd for " - "snapshot failed"); - goto xmlend; + gf_log("cli", GF_LOG_ERROR, + "Failed to begin " + "snapshot xml composite op"); + goto out; + } + } + + ret = cli_to_glusterd(&req, frame, gf_cli_snapshot_cbk, + (xdrproc_t)xdr_gf_cli_req, options, GLUSTER_CLI_SNAP, + this, cli_rpc_prog, NULL); + if (ret) { + gf_log("cli", GF_LOG_ERROR, + "cli_to_glusterd for " + "snapshot failed"); + goto xmlend; + } + + if (GF_SNAP_OPTION_TYPE_STATUS == type) { + ret = gf_cli_snapshot_for_status(frame, this, data); + if (ret) { + gf_log("cli", GF_LOG_ERROR, + "cli to glusterd " + "for snapshot status command failed"); } - if (GF_SNAP_OPTION_TYPE_STATUS == type) { - ret = gf_cli_snapshot_for_status (frame, this, data); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, "cli to glusterd " - "for snapshot status command failed"); - } + goto xmlend; + } - goto xmlend; + if (GF_SNAP_OPTION_TYPE_DELETE == type) { + ret = gf_cli_snapshot_for_delete(frame, this, data); + if (ret) { + gf_log("cli", GF_LOG_ERROR, + "cli to glusterd " + "for snapshot delete command failed"); } - if (GF_SNAP_OPTION_TYPE_DELETE == type) { - ret = gf_cli_snapshot_for_delete (frame, this, data); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, "cli to glusterd " - "for snapshot delete command failed"); - } + goto xmlend; + } - goto xmlend; - } - - ret = 0; + ret = 0; xmlend: - if (global_state->mode & GLUSTER_MODE_XML) { - ret = cli_xml_snapshot_end_composite_op (local); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Failed to end " - "snapshot xml composite op"); - goto out; - } - } -out: - if (ret && local && GF_SNAP_OPTION_TYPE_STATUS == type) { - tmp_ret = dict_get_str (local->dict, "op_err_str", &err_str); - if (tmp_ret || !err_str) { - cli_err ("Snapshot Status : failed: %s", "Please " - "check log file for details"); - } else { - cli_err ("Snapshot Status : failed: %s", err_str); - dict_del (local->dict, "op_err_str"); - } + if (global_state->mode & GLUSTER_MODE_XML) { + ret = cli_xml_snapshot_end_composite_op(local); + if (ret) { + gf_log("cli", GF_LOG_ERROR, + "Failed to end " + "snapshot xml composite op"); + goto out; + } + } +out: + if (ret && local && GF_SNAP_OPTION_TYPE_STATUS == type) { + tmp_ret = dict_get_str(local->dict, "op_err_str", &err_str); + if (tmp_ret || !err_str) { + cli_err("Snapshot Status : failed: %s", + "Please " + "check log file for details"); + } else { + cli_err("Snapshot Status : failed: %s", err_str); + dict_del(local->dict, "op_err_str"); } + } - gf_log ("cli", GF_LOG_DEBUG, "Returning %d", ret); + gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret); - GF_FREE (req.dict.dict_val); + GF_FREE(req.dict.dict_val); - if (global_state->mode & GLUSTER_MODE_XML) { - /* XML mode handles its own error */ - ret = 0; - } - return ret; + if (global_state->mode & GLUSTER_MODE_XML) { + /* XML mode handles its own error */ + ret = 0; + } + return ret; } int32_t -gf_cli_barrier_volume_cbk (struct rpc_req *req, struct iovec *iov, - int count, void *myframe) -{ - gf_cli_rsp rsp = {0,}; - int ret = -1; - - GF_ASSERT (myframe); - - if (-1 == req->rpc_status) - goto out; - - ret = xdr_to_generic (*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); - if (ret < 0) { - gf_log (((call_frame_t *) myframe)->this->name, GF_LOG_ERROR, - "Failed to decode xdr response"); - goto out; - } - gf_log ("cli", GF_LOG_DEBUG, "Received response to barrier"); - - if (rsp.op_ret) { - if (rsp.op_errstr && (strlen (rsp.op_errstr) > 1)) { - cli_err ("volume barrier: command unsuccessful : %s", - rsp.op_errstr); - } else { - cli_err ("volume barrier: command unsuccessful"); - } +gf_cli_barrier_volume_cbk(struct rpc_req *req, struct iovec *iov, int count, + void *myframe) +{ + gf_cli_rsp rsp = { + 0, + }; + int ret = -1; + + GF_ASSERT(myframe); + + if (-1 == req->rpc_status) + goto out; + + ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); + if (ret < 0) { + gf_log(((call_frame_t *)myframe)->this->name, GF_LOG_ERROR, + "Failed to decode xdr response"); + goto out; + } + gf_log("cli", GF_LOG_DEBUG, "Received response to barrier"); + + if (rsp.op_ret) { + if (rsp.op_errstr && (strlen(rsp.op_errstr) > 1)) { + cli_err("volume barrier: command unsuccessful : %s", rsp.op_errstr); } else { - cli_out ("volume barrier: command successful"); + cli_err("volume barrier: command unsuccessful"); } - ret = rsp.op_ret; + } else { + cli_out("volume barrier: command successful"); + } + ret = rsp.op_ret; out: - free (rsp.op_errstr); - free (rsp.dict.dict_val); - cli_cmd_broadcast_response (ret); - return ret; + free(rsp.op_errstr); + free(rsp.dict.dict_val); + cli_cmd_broadcast_response(ret); + return ret; } int -gf_cli_barrier_volume (call_frame_t *frame, xlator_t *this, void *data) +gf_cli_barrier_volume(call_frame_t *frame, xlator_t *this, void *data) { - gf_cli_req req = {{0,}}; - dict_t *options = NULL; - int ret = -1; + gf_cli_req req = {{ + 0, + }}; + dict_t *options = NULL; + int ret = -1; - if (!frame || !this || !data) - goto out; + if (!frame || !this || !data) + goto out; - options = data; + options = data; - ret = cli_to_glusterd (&req, frame, gf_cli_barrier_volume_cbk, - (xdrproc_t) xdr_gf_cli_req, options, - GLUSTER_CLI_BARRIER_VOLUME, this, cli_rpc_prog, - NULL); + ret = cli_to_glusterd(&req, frame, gf_cli_barrier_volume_cbk, + (xdrproc_t)xdr_gf_cli_req, options, + GLUSTER_CLI_BARRIER_VOLUME, this, cli_rpc_prog, NULL); out: - gf_log ("cli", GF_LOG_DEBUG, "Returning %d", ret); + gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret); - GF_FREE (req.dict.dict_val); - return ret; + GF_FREE(req.dict.dict_val); + return ret; } int32_t -gf_cli_get_vol_opt_cbk (struct rpc_req *req, struct iovec *iov, int count, - void *myframe) -{ - gf_cli_rsp rsp = {0,}; - int ret = -1; - dict_t *dict = NULL; - char *key = NULL; - char *value = NULL; - char msg[1024] = {0,}; - int i = 0; - char dict_key[50] = {0,}; - - GF_ASSERT (myframe); - - if (-1 == req->rpc_status) - goto out; - - ret = xdr_to_generic (*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); - if (ret < 0) { - gf_log (((call_frame_t *) myframe)->this->name, GF_LOG_ERROR, - "Failed to decode xdr response"); - goto out; - } - gf_log ("cli", GF_LOG_DEBUG, "Received response to get volume option"); - - if (rsp.op_ret) { - if (strcmp (rsp.op_errstr, "")) - snprintf (msg, sizeof (msg), "volume get option: " - "failed: %s", rsp.op_errstr); - else - snprintf (msg, sizeof (msg), "volume get option: " - "failed"); - - if (global_state->mode & GLUSTER_MODE_XML) { - ret = cli_xml_output_str ("volGetopts", msg, rsp.op_ret, - rsp.op_errno, rsp.op_errstr); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, - "Error outputting to xml"); - } - } else { - cli_err ("%s", msg); - } - ret = rsp.op_ret; - goto out_nolog; - } - dict = dict_new (); - - if (!dict) { - ret = -1; - goto out; - } - - ret = dict_unserialize (rsp.dict.dict_val, rsp.dict.dict_len, &dict); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, - "Failed rsp_dict unserialization"); - goto out; - } +gf_cli_get_vol_opt_cbk(struct rpc_req *req, struct iovec *iov, int count, + void *myframe) +{ + gf_cli_rsp rsp = { + 0, + }; + int ret = -1; + dict_t *dict = NULL; + char *key = NULL; + char *value = NULL; + char msg[1024] = { + 0, + }; + int i = 0; + char dict_key[50] = { + 0, + }; + + GF_ASSERT(myframe); + + if (-1 == req->rpc_status) + goto out; + + ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); + if (ret < 0) { + gf_log(((call_frame_t *)myframe)->this->name, GF_LOG_ERROR, + "Failed to decode xdr response"); + goto out; + } + gf_log("cli", GF_LOG_DEBUG, "Received response to get volume option"); + + if (rsp.op_ret) { + if (strcmp(rsp.op_errstr, "")) + snprintf(msg, sizeof(msg), + "volume get option: " + "failed: %s", + rsp.op_errstr); + else + snprintf(msg, sizeof(msg), + "volume get option: " + "failed"); if (global_state->mode & GLUSTER_MODE_XML) { - ret = cli_xml_output_vol_getopts (dict, rsp.op_ret, - rsp.op_errno, - rsp.op_errstr); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, "xml output generation " - "failed"); - ret = 0; - } - goto out; - } - - ret = dict_get_str (dict, "warning", &value); - if (!ret) { - cli_out ("%s", value); + ret = cli_xml_output_str("volGetopts", msg, rsp.op_ret, + rsp.op_errno, rsp.op_errstr); + if (ret) { + gf_log("cli", GF_LOG_ERROR, "Error outputting to xml"); + } + } else { + cli_err("%s", msg); } - - ret = dict_get_int32 (dict, "count", &count); + ret = rsp.op_ret; + goto out_nolog; + } + dict = dict_new(); + + if (!dict) { + ret = -1; + goto out; + } + + ret = dict_unserialize(rsp.dict.dict_val, rsp.dict.dict_len, &dict); + if (ret) { + gf_log("cli", GF_LOG_ERROR, "Failed rsp_dict unserialization"); + goto out; + } + + if (global_state->mode & GLUSTER_MODE_XML) { + ret = cli_xml_output_vol_getopts(dict, rsp.op_ret, rsp.op_errno, + rsp.op_errstr); if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Failed to retrieve count " - "from the dictionary"); - goto out; - } - - if (count <= 0) { - gf_log ("cli", GF_LOG_ERROR, "Value of count :%d is " - "invalid", count); - ret = -1; - goto out; - } - - cli_out ("%-40s%-40s", "Option", "Value"); - cli_out ("%-40s%-40s", "------", "-----"); - for (i=1; i<=count; i++) { - snprintf (dict_key, sizeof dict_key, "key%d", i); - ret = dict_get_str (dict, dict_key, &key); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Failed to" - " retrieve %s from the " - "dictionary", dict_key); - goto out; - } - snprintf (dict_key, sizeof dict_key, "value%d", i); - ret = dict_get_str (dict, dict_key, &value); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Failed to " - "retrieve key value for %s from" - "the dictionary", dict_key); - goto out; - } - cli_out ("%-40s%-40s", key, value); + gf_log("cli", GF_LOG_ERROR, + "xml output generation " + "failed"); + ret = 0; + } + goto out; + } + + ret = dict_get_str(dict, "warning", &value); + if (!ret) { + cli_out("%s", value); + } + + ret = dict_get_int32(dict, "count", &count); + if (ret) { + gf_log("cli", GF_LOG_ERROR, + "Failed to retrieve count " + "from the dictionary"); + goto out; + } + + if (count <= 0) { + gf_log("cli", GF_LOG_ERROR, + "Value of count :%d is " + "invalid", + count); + ret = -1; + goto out; + } + + cli_out("%-40s%-40s", "Option", "Value"); + cli_out("%-40s%-40s", "------", "-----"); + for (i = 1; i <= count; i++) { + snprintf(dict_key, sizeof dict_key, "key%d", i); + ret = dict_get_str(dict, dict_key, &key); + if (ret) { + gf_log("cli", GF_LOG_ERROR, + "Failed to" + " retrieve %s from the " + "dictionary", + dict_key); + goto out; + } + snprintf(dict_key, sizeof dict_key, "value%d", i); + ret = dict_get_str(dict, dict_key, &value); + if (ret) { + gf_log("cli", GF_LOG_ERROR, + "Failed to " + "retrieve key value for %s from" + "the dictionary", + dict_key); + goto out; } + cli_out("%-40s%-40s", key, value); + } out: - if (ret) { - cli_out ("volume get option failed. Check the cli/glusterd log " - "file for more details"); - } + if (ret) { + cli_out( + "volume get option failed. Check the cli/glusterd log " + "file for more details"); + } out_nolog: - if (dict) - dict_unref (dict); - free (rsp.op_errstr); - free (rsp.dict.dict_val); - cli_cmd_broadcast_response (ret); - return ret; + if (dict) + dict_unref(dict); + free(rsp.op_errstr); + free(rsp.dict.dict_val); + cli_cmd_broadcast_response(ret); + return ret; } int -gf_cli_get_vol_opt (call_frame_t *frame, xlator_t *this, void *data) +gf_cli_get_vol_opt(call_frame_t *frame, xlator_t *this, void *data) { - gf_cli_req req = {{0,}}; - dict_t *options = NULL; - int ret = -1; + gf_cli_req req = {{ + 0, + }}; + dict_t *options = NULL; + int ret = -1; - if (!frame || !this || !data) - goto out; + if (!frame || !this || !data) + goto out; - options = data; + options = data; - ret = cli_to_glusterd (&req, frame, gf_cli_get_vol_opt_cbk, - (xdrproc_t)xdr_gf_cli_req, options, - GLUSTER_CLI_GET_VOL_OPT, this, cli_rpc_prog, - NULL); + ret = cli_to_glusterd(&req, frame, gf_cli_get_vol_opt_cbk, + (xdrproc_t)xdr_gf_cli_req, options, + GLUSTER_CLI_GET_VOL_OPT, this, cli_rpc_prog, NULL); out: - gf_log ("cli", GF_LOG_DEBUG, "Returning %d", ret); + gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret); - GF_FREE (req.dict.dict_val); - return ret; + GF_FREE(req.dict.dict_val); + return ret; } int -add_cli_cmd_timeout_to_dict (dict_t *dict) +add_cli_cmd_timeout_to_dict(dict_t *dict) { - int ret = 0; + int ret = 0; - if (cli_default_conn_timeout > 120) { - ret = dict_set_uint32 (dict, "timeout", cli_default_conn_timeout); - if (ret) { - gf_log ("cli", GF_LOG_INFO, "Failed to save" - "timeout to dict"); - } + if (cli_default_conn_timeout > 120) { + ret = dict_set_uint32(dict, "timeout", cli_default_conn_timeout); + if (ret) { + gf_log("cli", GF_LOG_INFO, + "Failed to save" + "timeout to dict"); } - return ret; + } + return ret; } int -cli_to_glusterd (gf_cli_req *req, call_frame_t *frame, - fop_cbk_fn_t cbkfn, xdrproc_t xdrproc, dict_t *dict, - int procnum, xlator_t *this, rpc_clnt_prog_t *prog, - struct iobref *iobref) -{ - int ret = 0; - size_t len = 0; - char *cmd = NULL; - int i = 0; - const char **words = NULL; - cli_local_t *local = NULL; - - if (!this || !frame || !dict) { - ret = -1; - goto out; - } +cli_to_glusterd(gf_cli_req *req, call_frame_t *frame, fop_cbk_fn_t cbkfn, + xdrproc_t xdrproc, dict_t *dict, int procnum, xlator_t *this, + rpc_clnt_prog_t *prog, struct iobref *iobref) +{ + int ret = 0; + size_t len = 0; + char *cmd = NULL; + int i = 0; + const char **words = NULL; + cli_local_t *local = NULL; - if (!frame->local) { - ret = -1; - goto out; - } + if (!this || !frame || !dict) { + ret = -1; + goto out; + } - local = frame->local; + if (!frame->local) { + ret = -1; + goto out; + } - if (!local->words) { - ret = -1; - goto out; - } + local = frame->local; - words = local->words; + if (!local->words) { + ret = -1; + goto out; + } - while (words[i]) - len += strlen (words[i++]) + 1; + words = local->words; - cmd = GF_CALLOC (1, len, gf_common_mt_char); + while (words[i]) + len += strlen(words[i++]) + 1; - if (!cmd) { - ret = -1; - goto out; - } + cmd = GF_CALLOC(1, len, gf_common_mt_char); - for (i = 0; words[i]; i++) { - strncat (cmd, words[i], len - 1); - if (words[i+1] != NULL) - strncat (cmd, " ", len - 1); - } + if (!cmd) { + ret = -1; + goto out; + } - cmd [len - 1] = '\0'; + for (i = 0; words[i]; i++) { + strncat(cmd, words[i], len - 1); + if (words[i + 1] != NULL) + strncat(cmd, " ", len - 1); + } - ret = dict_set_dynstr (dict, "cmd-str", cmd); - if (ret) - goto out; + cmd[len - 1] = '\0'; - ret = add_cli_cmd_timeout_to_dict (dict); + ret = dict_set_dynstr(dict, "cmd-str", cmd); + if (ret) + goto out; - ret = dict_allocate_and_serialize (dict, &(req->dict).dict_val, - &(req->dict).dict_len); + ret = add_cli_cmd_timeout_to_dict(dict); - if (ret < 0) { - gf_log (this->name, GF_LOG_DEBUG, - "failed to get serialized length of dict"); - goto out; - } + ret = dict_allocate_and_serialize(dict, &(req->dict).dict_val, + &(req->dict).dict_len); - ret = cli_cmd_submit (NULL, req, frame, prog, procnum, iobref, this, - cbkfn, (xdrproc_t) xdrproc); -out: - return ret; + if (ret < 0) { + gf_log(this->name, GF_LOG_DEBUG, + "failed to get serialized length of dict"); + goto out; + } + ret = cli_cmd_submit(NULL, req, frame, prog, procnum, iobref, this, cbkfn, + (xdrproc_t)xdrproc); +out: + return ret; } int -gf_cli_print_bitrot_scrub_status (dict_t *dict) -{ - int i = 1; - int j = 0; - int ret = -1; - int count = 0; - char key[256] = {0,}; - char *volname = NULL; - char *node_name = NULL; - char *scrub_freq = NULL; - char *state_scrub = NULL; - char *scrub_impact = NULL; - char *bad_file_str = NULL; - char *scrub_log_file = NULL; - char *bitrot_log_file = NULL; - uint64_t scrub_files = 0; - uint64_t unsigned_files = 0; - uint64_t scrub_time = 0; - uint64_t days = 0; - uint64_t hours = 0; - uint64_t minutes = 0; - uint64_t seconds = 0; - char *last_scrub = NULL; - uint64_t error_count = 0; - int8_t scrub_running = 0; - char *scrub_state_op = NULL; - - - ret = dict_get_str (dict, "volname", &volname); +gf_cli_print_bitrot_scrub_status(dict_t *dict) +{ + int i = 1; + int j = 0; + int ret = -1; + int count = 0; + char key[256] = { + 0, + }; + char *volname = NULL; + char *node_name = NULL; + char *scrub_freq = NULL; + char *state_scrub = NULL; + char *scrub_impact = NULL; + char *bad_file_str = NULL; + char *scrub_log_file = NULL; + char *bitrot_log_file = NULL; + uint64_t scrub_files = 0; + uint64_t unsigned_files = 0; + uint64_t scrub_time = 0; + uint64_t days = 0; + uint64_t hours = 0; + uint64_t minutes = 0; + uint64_t seconds = 0; + char *last_scrub = NULL; + uint64_t error_count = 0; + int8_t scrub_running = 0; + char *scrub_state_op = NULL; + + ret = dict_get_str(dict, "volname", &volname); + if (ret) + gf_log("cli", GF_LOG_TRACE, "failed to get volume name"); + + ret = dict_get_str(dict, "features.scrub", &state_scrub); + if (ret) + gf_log("cli", GF_LOG_TRACE, "failed to get scrub state value"); + + ret = dict_get_str(dict, "features.scrub-throttle", &scrub_impact); + if (ret) + gf_log("cli", GF_LOG_TRACE, + "failed to get scrub impact " + "value"); + + ret = dict_get_str(dict, "features.scrub-freq", &scrub_freq); + if (ret) + gf_log("cli", GF_LOG_TRACE, "failed to get scrub -freq value"); + + ret = dict_get_str(dict, "bitrot_log_file", &bitrot_log_file); + if (ret) + gf_log("cli", GF_LOG_TRACE, + "failed to get bitrot log file " + "location"); + + ret = dict_get_str(dict, "scrub_log_file", &scrub_log_file); + if (ret) + gf_log("cli", GF_LOG_TRACE, + "failed to get scrubber log file " + "location"); + + ret = dict_get_int32(dict, "count", &count); + if (ret) { + gf_log("cli", GF_LOG_ERROR, + "count not get count value from" + " dictionary"); + goto out; + } + + for (i = 1; i <= count; i++) { + snprintf(key, 256, "scrub-running-%d", i); + ret = dict_get_int8(dict, key, &scrub_running); if (ret) - gf_log ("cli", GF_LOG_TRACE, "failed to get volume name"); - - ret = dict_get_str (dict, "features.scrub", &state_scrub); - if (ret) - gf_log ("cli", GF_LOG_TRACE, "failed to get scrub state value"); - - ret = dict_get_str (dict, "features.scrub-throttle", &scrub_impact); - if (ret) - gf_log ("cli", GF_LOG_TRACE, "failed to get scrub impact " - "value"); - - ret = dict_get_str (dict, "features.scrub-freq", &scrub_freq); - if (ret) - gf_log ("cli", GF_LOG_TRACE, "failed to get scrub -freq value"); - - ret = dict_get_str (dict, "bitrot_log_file", &bitrot_log_file); - if (ret) - gf_log ("cli", GF_LOG_TRACE, "failed to get bitrot log file " - "location"); - - ret = dict_get_str (dict, "scrub_log_file", &scrub_log_file); - if (ret) - gf_log ("cli", GF_LOG_TRACE, "failed to get scrubber log file " - "location"); + gf_log("cli", GF_LOG_TRACE, + "failed to get scrubbed " + "files"); + if (scrub_running) + break; + } - ret = dict_get_int32 (dict, "count", &count); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, "count not get count value from" - " dictionary"); - goto out; - } + if (scrub_running) + gf_asprintf(&scrub_state_op, "%s (In Progress)", state_scrub); + else + gf_asprintf(&scrub_state_op, "%s (Idle)", state_scrub); - for (i = 1; i <= count; i++) { - snprintf (key, 256, "scrub-running-%d", i); - ret = dict_get_int8 (dict, key, &scrub_running); - if (ret) - gf_log ("cli", GF_LOG_TRACE, "failed to get scrubbed " - "files"); - if (scrub_running) - break; - } + cli_out("\n%s: %s\n", "Volume name ", volname); - if (scrub_running) - gf_asprintf (&scrub_state_op, "%s (In Progress)", state_scrub); - else - gf_asprintf (&scrub_state_op, "%s (Idle)", state_scrub); + cli_out("%s: %s\n", "State of scrub", scrub_state_op); - cli_out ("\n%s: %s\n", "Volume name ", volname); + cli_out("%s: %s\n", "Scrub impact", scrub_impact); - cli_out ("%s: %s\n", "State of scrub", scrub_state_op); + cli_out("%s: %s\n", "Scrub frequency", scrub_freq); - cli_out ("%s: %s\n", "Scrub impact", scrub_impact); + cli_out("%s: %s\n", "Bitrot error log location", bitrot_log_file); - cli_out ("%s: %s\n", "Scrub frequency", scrub_freq); + cli_out("%s: %s\n", "Scrubber error log location", scrub_log_file); - cli_out ("%s: %s\n", "Bitrot error log location", bitrot_log_file); + for (i = 1; i <= count; i++) { + /* Reset the variables to prevent carryover of values */ + node_name = NULL; + last_scrub = NULL; + scrub_time = 0; + days = 0; + hours = 0; + minutes = 0; + seconds = 0; + error_count = 0; + scrub_files = 0; + unsigned_files = 0; - cli_out ("%s: %s\n", "Scrubber error log location", scrub_log_file); + snprintf(key, 256, "node-name-%d", i); + ret = dict_get_str(dict, key, &node_name); + if (ret) + gf_log("cli", GF_LOG_TRACE, "failed to get node-name"); + snprintf(key, 256, "scrubbed-files-%d", i); + ret = dict_get_uint64(dict, key, &scrub_files); + if (ret) + gf_log("cli", GF_LOG_TRACE, + "failed to get scrubbed " + "files"); - for (i = 1; i <= count; i++) { - /* Reset the variables to prevent carryover of values */ - node_name = NULL; - last_scrub = NULL; - scrub_time = 0; - days = 0; - hours = 0; - minutes = 0; - seconds = 0; - error_count = 0; - scrub_files = 0; - unsigned_files = 0; + snprintf(key, 256, "unsigned-files-%d", i); + ret = dict_get_uint64(dict, key, &unsigned_files); + if (ret) + gf_log("cli", GF_LOG_TRACE, + "failed to get unsigned " + "files"); - snprintf (key, 256, "node-name-%d", i); - ret = dict_get_str (dict, key, &node_name); - if (ret) - gf_log ("cli", GF_LOG_TRACE, "failed to get node-name"); + snprintf(key, 256, "scrub-duration-%d", i); + ret = dict_get_uint64(dict, key, &scrub_time); + if (ret) + gf_log("cli", GF_LOG_TRACE, + "failed to get last scrub " + "duration"); - snprintf (key, 256, "scrubbed-files-%d", i); - ret = dict_get_uint64 (dict, key, &scrub_files); - if (ret) - gf_log ("cli", GF_LOG_TRACE, "failed to get scrubbed " - "files"); + snprintf(key, 256, "last-scrub-time-%d", i); + ret = dict_get_str(dict, key, &last_scrub); + if (ret) + gf_log("cli", GF_LOG_TRACE, + "failed to get last scrub" + " time"); + snprintf(key, 256, "error-count-%d", i); + ret = dict_get_uint64(dict, key, &error_count); + if (ret) + gf_log("cli", GF_LOG_TRACE, + "failed to get error " + "count"); - snprintf (key, 256, "unsigned-files-%d", i); - ret = dict_get_uint64 (dict, key, &unsigned_files); - if (ret) - gf_log ("cli", GF_LOG_TRACE, "failed to get unsigned " - "files"); + cli_out("\n%s\n", + "==========================================" + "==============="); - snprintf (key, 256, "scrub-duration-%d", i); - ret = dict_get_uint64 (dict, key, &scrub_time); - if (ret) - gf_log ("cli", GF_LOG_TRACE, "failed to get last scrub " - "duration"); + cli_out("%s: %s\n", "Node", node_name); - snprintf (key, 256, "last-scrub-time-%d", i); - ret = dict_get_str (dict, key, &last_scrub); - if (ret) - gf_log ("cli", GF_LOG_TRACE, "failed to get last scrub" - " time"); - snprintf (key, 256, "error-count-%d", i); - ret = dict_get_uint64 (dict, key, &error_count); - if (ret) - gf_log ("cli", GF_LOG_TRACE, "failed to get error " - "count"); + cli_out("%s: %" PRIu64 "\n", "Number of Scrubbed files", scrub_files); - cli_out ("\n%s\n", "==========================================" - "==============="); + cli_out("%s: %" PRIu64 "\n", "Number of Skipped files", unsigned_files); - cli_out ("%s: %s\n", "Node", node_name); + if ((!last_scrub) || !strcmp(last_scrub, "")) + cli_out("%s: %s\n", "Last completed scrub time", + "Scrubber pending to complete."); + else + cli_out("%s: %s\n", "Last completed scrub time", last_scrub); - cli_out ("%s: %"PRIu64 "\n", "Number of Scrubbed files", - scrub_files); + /* Printing last scrub duration time in human readable form*/ + seconds = scrub_time % 60; + minutes = (scrub_time / 60) % 60; + hours = (scrub_time / 3600) % 24; + days = scrub_time / 86400; + cli_out("%s: %" PRIu64 ":%" PRIu64 ":%" PRIu64 ":%" PRIu64 "\n", + "Duration of last scrub (D:M:H:M:S)", days, hours, minutes, + seconds); - cli_out ("%s: %"PRIu64 "\n", "Number of Skipped files", - unsigned_files); + cli_out("%s: %" PRIu64 "\n", "Error count", error_count); - if ((!last_scrub) || !strcmp (last_scrub, "")) - cli_out ("%s: %s\n", "Last completed scrub time", - "Scrubber pending to complete."); - else - cli_out ("%s: %s\n", "Last completed scrub time", - last_scrub); - - /* Printing last scrub duration time in human readable form*/ - seconds = scrub_time%60; - minutes = (scrub_time/60)%60; - hours = (scrub_time/3600)%24; - days = scrub_time/86400; - cli_out ("%s: %"PRIu64 ":%"PRIu64 ":%"PRIu64 ":%"PRIu64 "\n", - "Duration of last scrub (D:M:H:M:S)", - days, hours, minutes, seconds); - - cli_out ("%s: %"PRIu64 "\n", "Error count", error_count); - - if (error_count) { - cli_out ("%s:\n", "Corrupted object's [GFID]"); - /* Printing list of bad file's (Corrupted object's)*/ - for (j = 0; j < error_count; j++) { - snprintf (key, 256, "quarantine-%d-%d", j, i); - ret = dict_get_str (dict, key, &bad_file_str); - if (!ret) { - cli_out ("%s\n", bad_file_str); - } - } + if (error_count) { + cli_out("%s:\n", "Corrupted object's [GFID]"); + /* Printing list of bad file's (Corrupted object's)*/ + for (j = 0; j < error_count; j++) { + snprintf(key, 256, "quarantine-%d-%d", j, i); + ret = dict_get_str(dict, key, &bad_file_str); + if (!ret) { + cli_out("%s\n", bad_file_str); } + } } - cli_out ("%s\n", "==========================================" - "==============="); + } + cli_out("%s\n", + "==========================================" + "==============="); out: - GF_FREE (scrub_state_op); - return 0; + GF_FREE(scrub_state_op); + return 0; } int -gf_cli_bitrot_cbk (struct rpc_req *req, struct iovec *iov, - int count, void *myframe) -{ - int ret = -1; - int type = 0; - gf_cli_rsp rsp = {0, }; - dict_t *dict = NULL; - char *scrub_cmd = NULL; - char *volname = NULL; - char *cmd_str = NULL; - char *cmd_op = NULL; - - GF_ASSERT (myframe); - - if (req->rpc_status == -1) { - goto out; - } - - ret = xdr_to_generic (*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); - if (ret < 0) { - gf_log (((call_frame_t *) myframe)->this->name, GF_LOG_ERROR, - "Failed to decode xdr response"); - goto out; - } - - if (rsp.op_ret) { - ret = -1; - if (global_state->mode & GLUSTER_MODE_XML) - goto xml_output; - - if (strcmp (rsp.op_errstr, "")) - cli_err ("Bitrot command failed : %s", rsp.op_errstr); - else - cli_err ("Bitrot command : failed"); - - goto out; - } +gf_cli_bitrot_cbk(struct rpc_req *req, struct iovec *iov, int count, + void *myframe) +{ + int ret = -1; + int type = 0; + gf_cli_rsp rsp = { + 0, + }; + dict_t *dict = NULL; + char *scrub_cmd = NULL; + char *volname = NULL; + char *cmd_str = NULL; + char *cmd_op = NULL; + + GF_ASSERT(myframe); + + if (req->rpc_status == -1) { + goto out; + } + + ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp); + if (ret < 0) { + gf_log(((call_frame_t *)myframe)->this->name, GF_LOG_ERROR, + "Failed to decode xdr response"); + goto out; + } + + if (rsp.op_ret) { + ret = -1; + if (global_state->mode & GLUSTER_MODE_XML) + goto xml_output; - if (rsp.dict.dict_len) { - /* Unserialize the dictionary */ - dict = dict_new (); + if (strcmp(rsp.op_errstr, "")) + cli_err("Bitrot command failed : %s", rsp.op_errstr); + else + cli_err("Bitrot command : failed"); - if (!dict) { - ret = -1; - goto out; - } + goto out; + } - ret = dict_unserialize (rsp.dict.dict_val, - rsp.dict.dict_len, - &dict); + if (rsp.dict.dict_len) { + /* Unserialize the dictionary */ + dict = dict_new(); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, "failed to unserialize " - "req-buffer to dictionary"); - goto out; - } + if (!dict) { + ret = -1; + goto out; } - gf_log ("cli", GF_LOG_DEBUG, "Received resp to bit rot command"); + ret = dict_unserialize(rsp.dict.dict_val, rsp.dict.dict_len, &dict); - ret = dict_get_int32 (dict, "type", &type); if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Failed to get command type"); - goto out; + gf_log("cli", GF_LOG_ERROR, + "failed to unserialize " + "req-buffer to dictionary"); + goto out; } + } - /* Ignoring the error, as using dict val for cli output only */ - ret = dict_get_str (dict, "scrub-value", &scrub_cmd); - if (ret) - gf_log ("cli", GF_LOG_TRACE, "Failed to get scrub command"); + gf_log("cli", GF_LOG_DEBUG, "Received resp to bit rot command"); - ret = dict_get_str (dict, "volname", &volname); - if (ret) - gf_log ("cli", GF_LOG_TRACE, "failed to get volume name"); + ret = dict_get_int32(dict, "type", &type); + if (ret) { + gf_log("cli", GF_LOG_ERROR, "Failed to get command type"); + goto out; + } - ret = dict_get_str (dict, "cmd-str", &cmd_str); - if (ret) - gf_log ("cli", GF_LOG_TRACE, "failed to get command string"); + /* Ignoring the error, as using dict val for cli output only */ + ret = dict_get_str(dict, "scrub-value", &scrub_cmd); + if (ret) + gf_log("cli", GF_LOG_TRACE, "Failed to get scrub command"); - if (cmd_str) - cmd_op = strrchr(cmd_str, ' ') + 1; + ret = dict_get_str(dict, "volname", &volname); + if (ret) + gf_log("cli", GF_LOG_TRACE, "failed to get volume name"); - if ((type == GF_BITROT_CMD_SCRUB_STATUS) && - !(global_state->mode & GLUSTER_MODE_XML)) { - ret = gf_cli_print_bitrot_scrub_status (dict); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, "Failed to print bitrot " - "scrub status"); - } - goto out; - } + ret = dict_get_str(dict, "cmd-str", &cmd_str); + if (ret) + gf_log("cli", GF_LOG_TRACE, "failed to get command string"); - switch (type) { - case GF_BITROT_OPTION_TYPE_ENABLE: - cli_out("volume bitrot: success bitrot enabled " - "for volume %s", volname); - ret = 0; - goto out; + if (cmd_str) + cmd_op = strrchr(cmd_str, ' ') + 1; + + if ((type == GF_BITROT_CMD_SCRUB_STATUS) && + !(global_state->mode & GLUSTER_MODE_XML)) { + ret = gf_cli_print_bitrot_scrub_status(dict); + if (ret) { + gf_log("cli", GF_LOG_ERROR, + "Failed to print bitrot " + "scrub status"); + } + goto out; + } + + switch (type) { + case GF_BITROT_OPTION_TYPE_ENABLE: + cli_out( + "volume bitrot: success bitrot enabled " + "for volume %s", + volname); + ret = 0; + goto out; case GF_BITROT_OPTION_TYPE_DISABLE: - cli_out("volume bitrot: success bitrot disabled " - "for volume %s", volname); - ret = 0; - goto out; + cli_out( + "volume bitrot: success bitrot disabled " + "for volume %s", + volname); + ret = 0; + goto out; case GF_BITROT_CMD_SCRUB_ONDEMAND: - cli_out("volume bitrot: scrubber started ondemand " - "for volume %s", volname); - ret = 0; - goto out; + cli_out( + "volume bitrot: scrubber started ondemand " + "for volume %s", + volname); + ret = 0; + goto out; case GF_BITROT_OPTION_TYPE_SCRUB: - if (!strncmp ("pause", scrub_cmd, sizeof ("pause"))) - cli_out("volume bitrot: scrubber paused " - "for volume %s", volname); - if (!strncmp ("resume", scrub_cmd, sizeof ("resume"))) - cli_out("volume bitrot: scrubber resumed " - "for volume %s", volname); - ret = 0; - goto out; + if (!strncmp("pause", scrub_cmd, sizeof("pause"))) + cli_out( + "volume bitrot: scrubber paused " + "for volume %s", + volname); + if (!strncmp("resume", scrub_cmd, sizeof("resume"))) + cli_out( + "volume bitrot: scrubber resumed " + "for volume %s", + volname); + ret = 0; + goto out; case GF_BITROT_OPTION_TYPE_SCRUB_FREQ: - cli_out("volume bitrot: scrub-frequency is set to %s " - "successfully for volume %s", cmd_op, volname); - ret = 0; - goto out; + cli_out( + "volume bitrot: scrub-frequency is set to %s " + "successfully for volume %s", + cmd_op, volname); + ret = 0; + goto out; case GF_BITROT_OPTION_TYPE_SCRUB_THROTTLE: - cli_out("volume bitrot: scrub-throttle is set to %s " - "successfully for volume %s", cmd_op, volname); - ret = 0; - goto out; - } + cli_out( + "volume bitrot: scrub-throttle is set to %s " + "successfully for volume %s", + cmd_op, volname); + ret = 0; + goto out; + } xml_output: - if (global_state->mode & GLUSTER_MODE_XML) { - ret = cli_xml_output_vol_profile (dict, rsp.op_ret, - rsp.op_errno, - rsp.op_errstr); - if (ret) - gf_log ("cli", GF_LOG_ERROR, - "Error outputting to xml"); - goto out; - } + if (global_state->mode & GLUSTER_MODE_XML) { + ret = cli_xml_output_vol_profile(dict, rsp.op_ret, rsp.op_errno, + rsp.op_errstr); + if (ret) + gf_log("cli", GF_LOG_ERROR, "Error outputting to xml"); + goto out; + } - if (!rsp.op_ret) - cli_out ("volume bitrot: success"); + if (!rsp.op_ret) + cli_out("volume bitrot: success"); - ret = rsp.op_ret; + ret = rsp.op_ret; out: - if (dict) - dict_unref (dict); + if (dict) + dict_unref(dict); - free (rsp.dict.dict_val); - free (rsp.op_errstr); + free(rsp.dict.dict_val); + free(rsp.op_errstr); - cli_cmd_broadcast_response (ret); + cli_cmd_broadcast_response(ret); - return ret; + return ret; } int32_t -gf_cli_bitrot (call_frame_t *frame, xlator_t *this, void *data) +gf_cli_bitrot(call_frame_t *frame, xlator_t *this, void *data) { - gf_cli_req req = { {0,} }; - dict_t *options = NULL; - int ret = -1; + gf_cli_req req = {{ + 0, + }}; + dict_t *options = NULL; + int ret = -1; - if (!frame || !this || !data) - goto out; + if (!frame || !this || !data) + goto out; - options = data; + options = data; - ret = cli_to_glusterd (&req, frame, gf_cli_bitrot_cbk, - (xdrproc_t) xdr_gf_cli_req, options, - GLUSTER_CLI_BITROT, this, cli_rpc_prog, - NULL); - if (ret) { - gf_log ("cli", GF_LOG_ERROR, "cli_to_glusterd for " - "bitrot failed"); - goto out; - } + ret = cli_to_glusterd(&req, frame, gf_cli_bitrot_cbk, + (xdrproc_t)xdr_gf_cli_req, options, + GLUSTER_CLI_BITROT, this, cli_rpc_prog, NULL); + if (ret) { + gf_log("cli", GF_LOG_ERROR, + "cli_to_glusterd for " + "bitrot failed"); + goto out; + } out: - gf_log ("cli", GF_LOG_DEBUG, "Returning %d", ret); + gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret); - GF_FREE (req.dict.dict_val); + GF_FREE(req.dict.dict_val); - return ret; + return ret; } struct rpc_clnt_procedure gluster_cli_actors[GLUSTER_CLI_MAXVALUE] = { - [GLUSTER_CLI_NULL] = {"NULL", NULL }, - [GLUSTER_CLI_PROBE] = {"PROBE_QUERY", gf_cli_probe}, - [GLUSTER_CLI_DEPROBE] = {"DEPROBE_QUERY", gf_cli_deprobe}, - [GLUSTER_CLI_LIST_FRIENDS] = {"LIST_FRIENDS", gf_cli_list_friends}, - [GLUSTER_CLI_UUID_RESET] = {"UUID_RESET", gf_cli3_1_uuid_reset}, - [GLUSTER_CLI_UUID_GET] = {"UUID_GET", gf_cli3_1_uuid_get}, - [GLUSTER_CLI_CREATE_VOLUME] = {"CREATE_VOLUME", gf_cli_create_volume}, - [GLUSTER_CLI_DELETE_VOLUME] = {"DELETE_VOLUME", gf_cli_delete_volume}, - [GLUSTER_CLI_START_VOLUME] = {"START_VOLUME", gf_cli_start_volume}, - [GLUSTER_CLI_STOP_VOLUME] = {"STOP_VOLUME", gf_cli_stop_volume}, - [GLUSTER_CLI_RENAME_VOLUME] = {"RENAME_VOLUME", gf_cli_rename_volume}, - [GLUSTER_CLI_DEFRAG_VOLUME] = {"DEFRAG_VOLUME", gf_cli_defrag_volume}, - [GLUSTER_CLI_GET_VOLUME] = {"GET_VOLUME", gf_cli_get_volume}, - [GLUSTER_CLI_GET_NEXT_VOLUME] = {"GET_NEXT_VOLUME", gf_cli_get_next_volume}, - [GLUSTER_CLI_SET_VOLUME] = {"SET_VOLUME", gf_cli_set_volume}, - [GLUSTER_CLI_ADD_BRICK] = {"ADD_BRICK", gf_cli_add_brick}, - [GLUSTER_CLI_REMOVE_BRICK] = {"REMOVE_BRICK", gf_cli_remove_brick}, - [GLUSTER_CLI_REPLACE_BRICK] = {"REPLACE_BRICK", gf_cli_replace_brick}, - [GLUSTER_CLI_LOG_ROTATE] = {"LOG ROTATE", gf_cli_log_rotate}, - [GLUSTER_CLI_GETSPEC] = {"GETSPEC", gf_cli_getspec}, - [GLUSTER_CLI_PMAP_PORTBYBRICK] = {"PMAP PORTBYBRICK", gf_cli_pmap_b2p}, - [GLUSTER_CLI_SYNC_VOLUME] = {"SYNC_VOLUME", gf_cli_sync_volume}, - [GLUSTER_CLI_RESET_VOLUME] = {"RESET_VOLUME", gf_cli_reset_volume}, - [GLUSTER_CLI_FSM_LOG] = {"FSM_LOG", gf_cli_fsm_log}, - [GLUSTER_CLI_GSYNC_SET] = {"GSYNC_SET", gf_cli_gsync_set}, - [GLUSTER_CLI_PROFILE_VOLUME] = {"PROFILE_VOLUME", gf_cli_profile_volume}, - [GLUSTER_CLI_QUOTA] = {"QUOTA", gf_cli_quota}, - [GLUSTER_CLI_TOP_VOLUME] = {"TOP_VOLUME", gf_cli_top_volume}, - [GLUSTER_CLI_GETWD] = {"GETWD", gf_cli_getwd}, - [GLUSTER_CLI_STATUS_VOLUME] = {"STATUS_VOLUME", gf_cli_status_volume}, - [GLUSTER_CLI_STATUS_ALL] = {"STATUS_ALL", gf_cli_status_volume_all}, - [GLUSTER_CLI_MOUNT] = {"MOUNT", gf_cli_mount}, - [GLUSTER_CLI_UMOUNT] = {"UMOUNT", gf_cli_umount}, - [GLUSTER_CLI_HEAL_VOLUME] = {"HEAL_VOLUME", gf_cli_heal_volume}, - [GLUSTER_CLI_STATEDUMP_VOLUME] = {"STATEDUMP_VOLUME", gf_cli_statedump_volume}, - [GLUSTER_CLI_LIST_VOLUME] = {"LIST_VOLUME", gf_cli_list_volume}, - [GLUSTER_CLI_CLRLOCKS_VOLUME] = {"CLEARLOCKS_VOLUME", gf_cli_clearlocks_volume}, - [GLUSTER_CLI_COPY_FILE] = {"COPY_FILE", gf_cli_copy_file}, - [GLUSTER_CLI_SYS_EXEC] = {"SYS_EXEC", gf_cli_sys_exec}, - [GLUSTER_CLI_SNAP] = {"SNAP", gf_cli_snapshot}, - [GLUSTER_CLI_BARRIER_VOLUME] = {"BARRIER VOLUME", gf_cli_barrier_volume}, - [GLUSTER_CLI_GET_VOL_OPT] = {"GET_VOL_OPT", gf_cli_get_vol_opt}, - [GLUSTER_CLI_BITROT] = {"BITROT", gf_cli_bitrot}, - [GLUSTER_CLI_ATTACH_TIER] = {"ATTACH_TIER", gf_cli_attach_tier}, - [GLUSTER_CLI_TIER] = {"TIER", gf_cli_tier}, - [GLUSTER_CLI_GET_STATE] = {"GET_STATE", gf_cli_get_state}, - [GLUSTER_CLI_RESET_BRICK] = {"RESET_BRICK", gf_cli_reset_brick}, - [GLUSTER_CLI_REMOVE_TIER_BRICK] = {"DETACH_TIER", gf_cli_remove_tier_brick}, - [GLUSTER_CLI_ADD_TIER_BRICK] = {"ADD_TIER_BRICK", gf_cli_add_tier_brick} -}; + [GLUSTER_CLI_NULL] = {"NULL", NULL}, + [GLUSTER_CLI_PROBE] = {"PROBE_QUERY", gf_cli_probe}, + [GLUSTER_CLI_DEPROBE] = {"DEPROBE_QUERY", gf_cli_deprobe}, + [GLUSTER_CLI_LIST_FRIENDS] = {"LIST_FRIENDS", gf_cli_list_friends}, + [GLUSTER_CLI_UUID_RESET] = {"UUID_RESET", gf_cli3_1_uuid_reset}, + [GLUSTER_CLI_UUID_GET] = {"UUID_GET", gf_cli3_1_uuid_get}, + [GLUSTER_CLI_CREATE_VOLUME] = {"CREATE_VOLUME", gf_cli_create_volume}, + [GLUSTER_CLI_DELETE_VOLUME] = {"DELETE_VOLUME", gf_cli_delete_volume}, + [GLUSTER_CLI_START_VOLUME] = {"START_VOLUME", gf_cli_start_volume}, + [GLUSTER_CLI_STOP_VOLUME] = {"STOP_VOLUME", gf_cli_stop_volume}, + [GLUSTER_CLI_RENAME_VOLUME] = {"RENAME_VOLUME", gf_cli_rename_volume}, + [GLUSTER_CLI_DEFRAG_VOLUME] = {"DEFRAG_VOLUME", gf_cli_defrag_volume}, + [GLUSTER_CLI_GET_VOLUME] = {"GET_VOLUME", gf_cli_get_volume}, + [GLUSTER_CLI_GET_NEXT_VOLUME] = {"GET_NEXT_VOLUME", gf_cli_get_next_volume}, + [GLUSTER_CLI_SET_VOLUME] = {"SET_VOLUME", gf_cli_set_volume}, + [GLUSTER_CLI_ADD_BRICK] = {"ADD_BRICK", gf_cli_add_brick}, + [GLUSTER_CLI_REMOVE_BRICK] = {"REMOVE_BRICK", gf_cli_remove_brick}, + [GLUSTER_CLI_REPLACE_BRICK] = {"REPLACE_BRICK", gf_cli_replace_brick}, + [GLUSTER_CLI_LOG_ROTATE] = {"LOG ROTATE", gf_cli_log_rotate}, + [GLUSTER_CLI_GETSPEC] = {"GETSPEC", gf_cli_getspec}, + [GLUSTER_CLI_PMAP_PORTBYBRICK] = {"PMAP PORTBYBRICK", gf_cli_pmap_b2p}, + [GLUSTER_CLI_SYNC_VOLUME] = {"SYNC_VOLUME", gf_cli_sync_volume}, + [GLUSTER_CLI_RESET_VOLUME] = {"RESET_VOLUME", gf_cli_reset_volume}, + [GLUSTER_CLI_FSM_LOG] = {"FSM_LOG", gf_cli_fsm_log}, + [GLUSTER_CLI_GSYNC_SET] = {"GSYNC_SET", gf_cli_gsync_set}, + [GLUSTER_CLI_PROFILE_VOLUME] = {"PROFILE_VOLUME", gf_cli_profile_volume}, + [GLUSTER_CLI_QUOTA] = {"QUOTA", gf_cli_quota}, + [GLUSTER_CLI_TOP_VOLUME] = {"TOP_VOLUME", gf_cli_top_volume}, + [GLUSTER_CLI_GETWD] = {"GETWD", gf_cli_getwd}, + [GLUSTER_CLI_STATUS_VOLUME] = {"STATUS_VOLUME", gf_cli_status_volume}, + [GLUSTER_CLI_STATUS_ALL] = {"STATUS_ALL", gf_cli_status_volume_all}, + [GLUSTER_CLI_MOUNT] = {"MOUNT", gf_cli_mount}, + [GLUSTER_CLI_UMOUNT] = {"UMOUNT", gf_cli_umount}, + [GLUSTER_CLI_HEAL_VOLUME] = {"HEAL_VOLUME", gf_cli_heal_volume}, + [GLUSTER_CLI_STATEDUMP_VOLUME] = {"STATEDUMP_VOLUME", + gf_cli_statedump_volume}, + [GLUSTER_CLI_LIST_VOLUME] = {"LIST_VOLUME", gf_cli_list_volume}, + [GLUSTER_CLI_CLRLOCKS_VOLUME] = {"CLEARLOCKS_VOLUME", + gf_cli_clearlocks_volume}, + [GLUSTER_CLI_COPY_FILE] = {"COPY_FILE", gf_cli_copy_file}, + [GLUSTER_CLI_SYS_EXEC] = {"SYS_EXEC", gf_cli_sys_exec}, + [GLUSTER_CLI_SNAP] = {"SNAP", gf_cli_snapshot}, + [GLUSTER_CLI_BARRIER_VOLUME] = {"BARRIER VOLUME", gf_cli_barrier_volume}, + [GLUSTER_CLI_GET_VOL_OPT] = {"GET_VOL_OPT", gf_cli_get_vol_opt}, + [GLUSTER_CLI_BITROT] = {"BITROT", gf_cli_bitrot}, + [GLUSTER_CLI_ATTACH_TIER] = {"ATTACH_TIER", gf_cli_attach_tier}, + [GLUSTER_CLI_TIER] = {"TIER", gf_cli_tier}, + [GLUSTER_CLI_GET_STATE] = {"GET_STATE", gf_cli_get_state}, + [GLUSTER_CLI_RESET_BRICK] = {"RESET_BRICK", gf_cli_reset_brick}, + [GLUSTER_CLI_REMOVE_TIER_BRICK] = {"DETACH_TIER", gf_cli_remove_tier_brick}, + [GLUSTER_CLI_ADD_TIER_BRICK] = {"ADD_TIER_BRICK", gf_cli_add_tier_brick}}; struct rpc_clnt_program cli_prog = { - .progname = "Gluster CLI", - .prognum = GLUSTER_CLI_PROGRAM, - .progver = GLUSTER_CLI_VERSION, - .numproc = GLUSTER_CLI_MAXVALUE, - .proctable = gluster_cli_actors, + .progname = "Gluster CLI", + .prognum = GLUSTER_CLI_PROGRAM, + .progver = GLUSTER_CLI_VERSION, + .numproc = GLUSTER_CLI_MAXVALUE, + .proctable = gluster_cli_actors, }; struct rpc_clnt_procedure cli_quotad_procs[GF_AGGREGATOR_MAXVALUE] = { - [GF_AGGREGATOR_NULL] = {"NULL", NULL}, - [GF_AGGREGATOR_LOOKUP] = {"LOOKUP", NULL}, - [GF_AGGREGATOR_GETLIMIT] = {"GETLIMIT", cli_quotad_getlimit}, + [GF_AGGREGATOR_NULL] = {"NULL", NULL}, + [GF_AGGREGATOR_LOOKUP] = {"LOOKUP", NULL}, + [GF_AGGREGATOR_GETLIMIT] = {"GETLIMIT", cli_quotad_getlimit}, }; struct rpc_clnt_program cli_quotad_clnt = { - .progname = "CLI Quotad client", - .prognum = GLUSTER_AGGREGATOR_PROGRAM, - .progver = GLUSTER_AGGREGATOR_VERSION, - .numproc = GF_AGGREGATOR_MAXVALUE, - .proctable = cli_quotad_procs, + .progname = "CLI Quotad client", + .prognum = GLUSTER_AGGREGATOR_PROGRAM, + .progver = GLUSTER_AGGREGATOR_VERSION, + .numproc = GF_AGGREGATOR_MAXVALUE, + .proctable = cli_quotad_procs, }; |