<feed xmlns='http://www.w3.org/2005/Atom'>
<title>glusterfs.git/rpc, branch v3.3.0qa12</title>
<subtitle></subtitle>
<link rel='alternate' type='text/html' href='http://dev.gluster.org/cgit/glusterfs.git/'/>
<entry>
<title>build: Solaris build fixes</title>
<updated>2011-09-27T12:44:41+00:00</updated>
<author>
<name>Kaushik BV</name>
<email>kaushikbv@gluster.com</email>
</author>
<published>2011-09-26T09:08:58+00:00</published>
<link rel='alternate' type='text/html' href='http://dev.gluster.org/cgit/glusterfs.git/commit/?id=16b7e3bf201686ca03f8c35c20295e05abe52df8'/>
<id>16b7e3bf201686ca03f8c35c20295e05abe52df8</id>
<content type='text'>
Change-Id: Ic47d07a199ecbc42a141e391bbc75096e270e404
BUG: 3446
Reviewed-on: http://review.gluster.com/503
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Amar Tumballi &lt;amar@gluster.com&gt;
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
Change-Id: Ic47d07a199ecbc42a141e391bbc75096e270e404
BUG: 3446
Reviewed-on: http://review.gluster.com/503
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Amar Tumballi &lt;amar@gluster.com&gt;
</pre>
</div>
</content>
</entry>
<entry>
<title>rpc: fix wrong wiping of state machine's state</title>
<updated>2011-09-25T13:15:12+00:00</updated>
<author>
<name>Anand Avati</name>
<email>avati@gluster.com</email>
</author>
<published>2011-09-25T13:04:14+00:00</published>
<link rel='alternate' type='text/html' href='http://dev.gluster.org/cgit/glusterfs.git/commit/?id=b894944857a7ce61a3bd31566d6169ed8d19ce8b'/>
<id>b894944857a7ce61a3bd31566d6169ed8d19ce8b</id>
<content type='text'>
The macro __socket_proto_read() is used to read one sub-header in a fragment
after another. It also has to 'get out' of the state machine's
switch/case construct in two situations -

a) finished reading the full header

b) after a partial read, (and return in the same state at next pollin)

The 'reset' of the state should happen only after a full read of the header
and not if there was a partial read of the header.

Change-Id: I3650a83e1fae0f317cfd1b549835c72e39dc5253
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
The macro __socket_proto_read() is used to read one sub-header in a fragment
after another. It also has to 'get out' of the state machine's
switch/case construct in two situations -

a) finished reading the full header

b) after a partial read, (and return in the same state at next pollin)

The 'reset' of the state should happen only after a full read of the header
and not if there was a partial read of the header.

Change-Id: I3650a83e1fae0f317cfd1b549835c72e39dc5253
</pre>
</div>
</content>
</entry>
<entry>
<title>rpc: No timeouts for lock fops.</title>
<updated>2011-09-23T13:11:32+00:00</updated>
<author>
<name>Krishnan Parthasarathi</name>
<email>kp@gluster.com</email>
</author>
<published>2011-09-21T06:23:45+00:00</published>
<link rel='alternate' type='text/html' href='http://dev.gluster.org/cgit/glusterfs.git/commit/?id=1fdcbb61872b0a849490d1f0de4fdb545e2865aa'/>
<id>1fdcbb61872b0a849490d1f0de4fdb545e2865aa</id>
<content type='text'>
As of today, all fops bail out after 30 mins by default. This is not desirable
with lock fops since it could be 'rightfully blocked' for a really long time.
But the client would assume that the lock fop has 'expired' after 30 mins and
clean up its references to the locks.  Later when the locks xlator decides to
grant it, we are left with an orphan (stale) lock .

This fix exempts lock fops from timeouts. Only on network disruptions both
client and server decide to 'clean up' the locks held.

Change-Id: If1d74ba21113650b976728e9b764c551a0a49e59
BUG: 3617
Reviewed-on: http://review.gluster.com/478
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Amar Tumballi &lt;amar@gluster.com&gt;
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
As of today, all fops bail out after 30 mins by default. This is not desirable
with lock fops since it could be 'rightfully blocked' for a really long time.
But the client would assume that the lock fop has 'expired' after 30 mins and
clean up its references to the locks.  Later when the locks xlator decides to
grant it, we are left with an orphan (stale) lock .

This fix exempts lock fops from timeouts. Only on network disruptions both
client and server decide to 'clean up' the locks held.

Change-Id: If1d74ba21113650b976728e9b764c551a0a49e59
BUG: 3617
Reviewed-on: http://review.gluster.com/478
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Amar Tumballi &lt;amar@gluster.com&gt;
</pre>
</div>
</content>
</entry>
<entry>
<title>glusterd: Implemented cmd to trigger self-heal on a replicate volume.</title>
<updated>2011-09-22T16:43:25+00:00</updated>
<author>
<name>Krishnan Parthasarathi</name>
<email>kp@gluster.com</email>
</author>
<published>2011-09-16T05:10:32+00:00</published>
<link rel='alternate' type='text/html' href='http://dev.gluster.org/cgit/glusterfs.git/commit/?id=4765dd1a1c51c67ab86687fbd871c89156680c34'/>
<id>4765dd1a1c51c67ab86687fbd871c89156680c34</id>
<content type='text'>
This cmd is used in the context of proactive self-heal for replicated
volumes. User invokes the following cmd when (s)he suspects that self-heal
needs to be done on a particular volume,
        gluster volume heal &lt;VOLNAME&gt;.

Change-Id: I3954353b53488c28b70406e261808239b44997f3
BUG: 3602
Reviewed-on: http://review.gluster.com/454
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Vijay Bellur &lt;vijay@gluster.com&gt;
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
This cmd is used in the context of proactive self-heal for replicated
volumes. User invokes the following cmd when (s)he suspects that self-heal
needs to be done on a particular volume,
        gluster volume heal &lt;VOLNAME&gt;.

Change-Id: I3954353b53488c28b70406e261808239b44997f3
BUG: 3602
Reviewed-on: http://review.gluster.com/454
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Vijay Bellur &lt;vijay@gluster.com&gt;
</pre>
</div>
</content>
</entry>
<entry>
<title>NFS : Posix compliance fixes.</title>
<updated>2011-09-22T11:51:22+00:00</updated>
<author>
<name>Gaurav</name>
<email>gaurav@gluster.com</email>
</author>
<published>2011-09-22T07:18:41+00:00</published>
<link rel='alternate' type='text/html' href='http://dev.gluster.org/cgit/glusterfs.git/commit/?id=2e61c86ba89d546c7cab522b53d6b9c7f164cd8c'/>
<id>2e61c86ba89d546c7cab522b53d6b9c7f164cd8c</id>
<content type='text'>
Change-Id: I4f1d8883a3eedee20557e306815674a6dc062bfe
BUG: 1337
Reviewed-on: http://review.gluster.com/487
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Amar Tumballi &lt;amar@gluster.com&gt;
Reviewed-by: Vijay Bellur &lt;vijay@gluster.com&gt;
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
Change-Id: I4f1d8883a3eedee20557e306815674a6dc062bfe
BUG: 1337
Reviewed-on: http://review.gluster.com/487
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Amar Tumballi &lt;amar@gluster.com&gt;
Reviewed-by: Vijay Bellur &lt;vijay@gluster.com&gt;
</pre>
</div>
</content>
</entry>
<entry>
<title>rpc:memcpy data of right size to auth params</title>
<updated>2011-09-20T05:00:08+00:00</updated>
<author>
<name>Vijay Bellur</name>
<email>vijay@gluster.com</email>
</author>
<published>2011-09-19T16:31:13+00:00</published>
<link rel='alternate' type='text/html' href='http://dev.gluster.org/cgit/glusterfs.git/commit/?id=e0178776546f06646c11bcb1ea345f82436d73aa'/>
<id>e0178776546f06646c11bcb1ea345f82436d73aa</id>
<content type='text'>
Change-Id: I6999e0a1463ee3475f7607990337326b3f008538
BUG: 3587
Reviewed-on: http://review.gluster.com/464
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Amar Tumballi &lt;amar@gluster.com&gt;
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
Change-Id: I6999e0a1463ee3475f7607990337326b3f008538
BUG: 3587
Reviewed-on: http://review.gluster.com/464
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Amar Tumballi &lt;amar@gluster.com&gt;
</pre>
</div>
</content>
</entry>
<entry>
<title>rpc: Use correct variable for volume-specific insecure ports check</title>
<updated>2011-09-19T05:17:26+00:00</updated>
<author>
<name>krishna</name>
<email>krishna@gluster.com</email>
</author>
<published>2011-09-14T10:43:27+00:00</published>
<link rel='alternate' type='text/html' href='http://dev.gluster.org/cgit/glusterfs.git/commit/?id=9eec7900639684e5a2fd80b273437fefe6b5b930'/>
<id>9eec7900639684e5a2fd80b273437fefe6b5b930</id>
<content type='text'>
Change-Id: Ica59a3154c4e7d980d3477dd3879c7cebe92ad87
BUG: 3515
Reviewed-on: http://review.gluster.com/418
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Vijay Bellur &lt;vijay@gluster.com&gt;
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
Change-Id: Ica59a3154c4e7d980d3477dd3879c7cebe92ad87
BUG: 3515
Reviewed-on: http://review.gluster.com/418
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Vijay Bellur &lt;vijay@gluster.com&gt;
</pre>
</div>
</content>
</entry>
<entry>
<title>rpc: fix up mountbroker xdr defs and regenerate headers</title>
<updated>2011-09-16T04:56:12+00:00</updated>
<author>
<name>Csaba Henk</name>
<email>csaba@gluster.com</email>
</author>
<published>2011-09-15T16:14:53+00:00</published>
<link rel='alternate' type='text/html' href='http://dev.gluster.org/cgit/glusterfs.git/commit/?id=77cce4c7df9a61b599a27522696720e747714ff2'/>
<id>77cce4c7df9a61b599a27522696720e747714ff2</id>
<content type='text'>
Change-Id: I8a88d2b9228c3614ee7cbaf48782a419e6aee0f6
BUG: 3482
Reported-by: Krishnan Parthasarathi &lt;kp@gluster.com&gt;
Reviewed-on: http://review.gluster.com/432
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Krishnan Parthasarathi &lt;kp@gluster.com&gt;
Reviewed-by: Vijay Bellur &lt;vijay@gluster.com&gt;
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
Change-Id: I8a88d2b9228c3614ee7cbaf48782a419e6aee0f6
BUG: 3482
Reported-by: Krishnan Parthasarathi &lt;kp@gluster.com&gt;
Reviewed-on: http://review.gluster.com/432
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Krishnan Parthasarathi &lt;kp@gluster.com&gt;
Reviewed-by: Vijay Bellur &lt;vijay@gluster.com&gt;
</pre>
</div>
</content>
</entry>
<entry>
<title>support for de-commissioning a node using 'remove-brick'</title>
<updated>2011-09-13T09:10:12+00:00</updated>
<author>
<name>Amar Tumballi</name>
<email>amar@gluster.com</email>
</author>
<published>2011-09-09T04:12:51+00:00</published>
<link rel='alternate' type='text/html' href='http://dev.gluster.org/cgit/glusterfs.git/commit/?id=25daa42911d2ff697880ee29c591cac5f2abebed'/>
<id>25daa42911d2ff697880ee29c591cac5f2abebed</id>
<content type='text'>
to achieve this, we now create volume-file with
'decommissioned-nodes' option in distribute volume, then just
perform the rebalance set of operations (with 'force' flag set).

now onwards, the 'remove-brick' (with 'start' option) operation tries
to migrate data from removed bricks to existing bricks.

'remove-brick' also supports similar options as of replace-brick.

* (no options) -&gt; works as 'force', will have the current behavior
         of remove-brick, ie., no data-migration, volume changes.

* start  (starts remove-brick with data-migration/draining process,
          which takes care of migrating data and once complete, will
          commit the changes to volume file)
* pause  (stop data migration, but keep the volume file intact with
          extra options whatever is set)
* abort  (stop data-migration, and fall back to old configuration)
* commit (if volume is stopped, commits the changes to volumefile)
* force  (stops the data-migration and commits the changes to
          volume file)

Change-Id: I3952bcfbe604a0952e68b6accace7014d5e401d3
BUG: 1952
Reviewed-on: http://review.gluster.com/118
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Vijay Bellur &lt;vijay@gluster.com&gt;
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
to achieve this, we now create volume-file with
'decommissioned-nodes' option in distribute volume, then just
perform the rebalance set of operations (with 'force' flag set).

now onwards, the 'remove-brick' (with 'start' option) operation tries
to migrate data from removed bricks to existing bricks.

'remove-brick' also supports similar options as of replace-brick.

* (no options) -&gt; works as 'force', will have the current behavior
         of remove-brick, ie., no data-migration, volume changes.

* start  (starts remove-brick with data-migration/draining process,
          which takes care of migrating data and once complete, will
          commit the changes to volume file)
* pause  (stop data migration, but keep the volume file intact with
          extra options whatever is set)
* abort  (stop data-migration, and fall back to old configuration)
* commit (if volume is stopped, commits the changes to volumefile)
* force  (stops the data-migration and commits the changes to
          volume file)

Change-Id: I3952bcfbe604a0952e68b6accace7014d5e401d3
BUG: 1952
Reviewed-on: http://review.gluster.com/118
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Vijay Bellur &lt;vijay@gluster.com&gt;
</pre>
</div>
</content>
</entry>
<entry>
<title>socket: provide proper arguments to getaddrinfo</title>
<updated>2011-09-13T07:04:33+00:00</updated>
<author>
<name>Amar Tumballi</name>
<email>amar@gluster.com</email>
</author>
<published>2011-09-13T06:41:52+00:00</published>
<link rel='alternate' type='text/html' href='http://dev.gluster.org/cgit/glusterfs.git/commit/?id=58c5e9feae1fdf618bb120f4c8ea318cc628dd06'/>
<id>58c5e9feae1fdf618bb120f4c8ea318cc628dd06</id>
<content type='text'>
-----
from 'man getaddrinfo' :

If hints.ai_flags includes the AI_ADDRCONFIG flag, then IPv4 addresses
are returned in the list pointed to by res only if the local system has
at least one IPv4 address  configured, and IPv6 addresses are only
returned if the local system has at least one IPv6 address configured.
-----

Change-Id: Ie30344daf1bb9d41ac58741b38e83af35cd8b5e9
BUG: 2456
Reviewed-on: http://review.gluster.com/405
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Vijay Bellur &lt;vijay@gluster.com&gt;
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
-----
from 'man getaddrinfo' :

If hints.ai_flags includes the AI_ADDRCONFIG flag, then IPv4 addresses
are returned in the list pointed to by res only if the local system has
at least one IPv4 address  configured, and IPv6 addresses are only
returned if the local system has at least one IPv6 address configured.
-----

Change-Id: Ie30344daf1bb9d41ac58741b38e83af35cd8b5e9
BUG: 2456
Reviewed-on: http://review.gluster.com/405
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Vijay Bellur &lt;vijay@gluster.com&gt;
</pre>
</div>
</content>
</entry>
</feed>
