diff options
Diffstat (limited to 'xlators/mgmt/glusterd/src/glusterd-handler.c')
-rw-r--r-- | xlators/mgmt/glusterd/src/glusterd-handler.c | 45 |
1 files changed, 23 insertions, 22 deletions
diff --git a/xlators/mgmt/glusterd/src/glusterd-handler.c b/xlators/mgmt/glusterd/src/glusterd-handler.c index 8ba16b7a804..eaa05969656 100644 --- a/xlators/mgmt/glusterd/src/glusterd-handler.c +++ b/xlators/mgmt/glusterd/src/glusterd-handler.c @@ -465,7 +465,7 @@ glusterd_add_volume_detail_to_dict (glusterd_volinfo_t *volinfo, } #endif - list_for_each_entry (brickinfo, &volinfo->bricks, brick_list) { + cds_list_for_each_entry (brickinfo, &volinfo->bricks, brick_list) { char brick[1024] = {0,}; char brick_uuid[64] = {0,}; snprintf (key, 256, "volume%d.brick%d", count, i); @@ -609,7 +609,7 @@ glusterd_op_txn_begin (rpcsvc_request_t *req, glusterd_op_t op, void *ctx, local_locking_done: - INIT_LIST_HEAD (&priv->xaction_peers); + CDS_INIT_LIST_HEAD (&priv->xaction_peers); npeers = gd_build_peers_list (&priv->peers, &priv->xaction_peers, op); @@ -1162,8 +1162,8 @@ __glusterd_handle_cli_deprobe (rpcsvc_request_t *req) * detached. It's not a problem if a volume contains none or all * of its bricks on the peer being detached */ - list_for_each_entry_safe (volinfo, tmp, &priv->volumes, - vol_list) { + cds_list_for_each_entry_safe (volinfo, tmp, &priv->volumes, + vol_list) { ret = glusterd_friend_contains_vol_bricks (volinfo, uuid); if (ret == 1) { @@ -1379,7 +1379,7 @@ __glusterd_handle_cli_uuid_reset (rpcsvc_request_t *req) */ ret = -1; // Do not allow peer reset if there are any volumes in the cluster - if (!list_empty (&priv->volumes)) { + if (!cds_list_empty (&priv->volumes)) { snprintf (msg_str, sizeof (msg_str), "volumes are already " "present in the cluster. Resetting uuid is not " "allowed"); @@ -1388,7 +1388,7 @@ __glusterd_handle_cli_uuid_reset (rpcsvc_request_t *req) } // Do not allow peer reset if trusted storage pool is already formed - if (!list_empty (&priv->peers)) { + if (!cds_list_empty (&priv->peers)) { snprintf (msg_str, sizeof (msg_str),"trusted storage pool " "has been already formed. Please detach this peer " "from the pool and reset its uuid."); @@ -1550,7 +1550,7 @@ __glusterd_handle_cli_list_volume (rpcsvc_request_t *req) if (!dict) goto out; - list_for_each_entry (volinfo, &priv->volumes, vol_list) { + cds_list_for_each_entry (volinfo, &priv->volumes, vol_list) { memset (key, 0, sizeof (key)); snprintf (key, sizeof (key), "volume%d", count); ret = dict_set_str (dict, key, volinfo->volname); @@ -2550,7 +2550,7 @@ __glusterd_handle_probe_query (rpcsvc_request_t *req) goto out; } peerinfo = glusterd_peerinfo_find (probe_req.uuid, remote_hostname); - if ((peerinfo == NULL) && (!list_empty (&conf->peers))) { + if ((peerinfo == NULL) && (!cds_list_empty (&conf->peers))) { rsp.op_ret = -1; rsp.op_errno = GF_PROBE_ANOTHER_CLUSTER; } else if (peerinfo == NULL) { @@ -3079,7 +3079,7 @@ glusterd_friend_add (const char *hoststr, int port, * invalid peer name). That would mean we're adding something that had * just been free, and we're likely to crash later. */ - list_add_tail (&(*friend)->uuid_list, &conf->peers); + cds_list_add_tail (&(*friend)->uuid_list, &conf->peers); //restore needs to first create the list of peers, then create rpcs //to keep track of quorum in race-free manner. In restore for each peer @@ -3132,7 +3132,7 @@ glusterd_friend_add_from_peerinfo (glusterd_peerinfo_t *friend, * invalid peer name). That would mean we're adding something that had * just been free, and we're likely to crash later. */ - list_add_tail (&friend->uuid_list, &conf->peers); + cds_list_add_tail (&friend->uuid_list, &conf->peers); //restore needs to first create the list of peers, then create rpcs //to keep track of quorum in race-free manner. In restore for each peer @@ -3590,8 +3590,8 @@ glusterd_list_friends (rpcsvc_request_t *req, dict_t *dict, int32_t flags) gf_log ("", GF_LOG_WARNING, "Out of Memory"); goto out; } - if (!list_empty (&priv->peers)) { - list_for_each_entry (entry, &priv->peers, uuid_list) { + if (!cds_list_empty (&priv->peers)) { + cds_list_for_each_entry (entry, &priv->peers, uuid_list) { count++; ret = gd_add_peer_detail_to_dict (entry, friends, count); @@ -3665,13 +3665,13 @@ glusterd_get_volumes (rpcsvc_request_t *req, dict_t *dict, int32_t flags) goto out; } - if (list_empty (&priv->volumes)) { + if (cds_list_empty (&priv->volumes)) { ret = 0; goto respond; } if (flags == GF_CLI_GET_VOLUME_ALL) { - list_for_each_entry (entry, &priv->volumes, vol_list) { + cds_list_for_each_entry (entry, &priv->volumes, vol_list) { ret = glusterd_add_volume_detail_to_dict (entry, volumes, count); if (ret) @@ -3686,17 +3686,17 @@ glusterd_get_volumes (rpcsvc_request_t *req, dict_t *dict, int32_t flags) if (ret) { if (priv->volumes.next) { - entry = list_entry (priv->volumes.next, - typeof (*entry), - vol_list); + entry = cds_list_entry (priv->volumes.next, + typeof (*entry), + vol_list); } } else { ret = glusterd_volinfo_find (volname, &entry); if (ret) goto respond; - entry = list_entry (entry->vol_list.next, - typeof (*entry), - vol_list); + entry = cds_list_entry (entry->vol_list.next, + typeof (*entry), + vol_list); } if (&entry->vol_list == &priv->volumes) { @@ -4508,8 +4508,9 @@ __glusterd_peer_rpc_notify (struct rpc_clnt *rpc, void *mydata, !uuid_compare (peerinfo->uuid, uuid)) glusterd_unlock (peerinfo->uuid); } else { - list_for_each_entry (volinfo, &conf->volumes, - vol_list) { + cds_list_for_each_entry (volinfo, + &conf->volumes, + vol_list) { ret = glusterd_mgmt_v3_unlock (volinfo->volname, peerinfo->uuid, |