<feed xmlns='http://www.w3.org/2005/Atom'>
<title>glusterfs.git/tests/bugs/glusterd, branch v3.7.0beta2</title>
<subtitle></subtitle>
<link rel='alternate' type='text/html' href='http://dev.gluster.org/cgit/glusterfs.git/'/>
<entry>
<title>dht: make lookup-unhashed=auto do something actually useful</title>
<updated>2015-05-10T04:55:09+00:00</updated>
<author>
<name>Jeff Darcy</name>
<email>jdarcy@redhat.com</email>
</author>
<published>2014-05-07T19:31:30+00:00</published>
<link rel='alternate' type='text/html' href='http://dev.gluster.org/cgit/glusterfs.git/commit/?id=243d61575c093c03b9beb014bf9d097646836e95'/>
<id>243d61575c093c03b9beb014bf9d097646836e95</id>
<content type='text'>
The key concept here is to determine whether a directory is "clean" by
comparing its last-known-good topology to the current one for the
volume.  These are stored as "commit hashes" on the directory and the
volume root respectively.  The volume's commit hash changes whenever a
brick is added or removed, and a fix-layout is done.  A directory's
commit hash changes only when a full rebalance (not just fix-layout)
is done on it.  If all bricks are present and have a directory
commit hash that matches the volume commit hash, then we can assume
that every file is in its "proper" place. Therefore, if we look for
a file in that proper place and don't find it, we can assume it's not
on any other subvolume and *safely* skip the global (broadcast to all)
lookup.

Change-Id: Id6ce4593ba1f7daffa74cfab591cb45960629ae3
BUG: 1220064
Reviewed-on-master: http://review.gluster.org/#/c/7702/
Signed-off-by: Jeff Darcy &lt;jdarcy@redhat.com&gt;
Signed-off-by: Shyam &lt;srangana@redhat.com&gt;
Reviewed-on: http://review.gluster.org/10729
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Krishnan Parthasarathi &lt;kparthas@redhat.com&gt;
Reviewed-by: Vijay Bellur &lt;vbellur@redhat.com&gt;
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
The key concept here is to determine whether a directory is "clean" by
comparing its last-known-good topology to the current one for the
volume.  These are stored as "commit hashes" on the directory and the
volume root respectively.  The volume's commit hash changes whenever a
brick is added or removed, and a fix-layout is done.  A directory's
commit hash changes only when a full rebalance (not just fix-layout)
is done on it.  If all bricks are present and have a directory
commit hash that matches the volume commit hash, then we can assume
that every file is in its "proper" place. Therefore, if we look for
a file in that proper place and don't find it, we can assume it's not
on any other subvolume and *safely* skip the global (broadcast to all)
lookup.

Change-Id: Id6ce4593ba1f7daffa74cfab591cb45960629ae3
BUG: 1220064
Reviewed-on-master: http://review.gluster.org/#/c/7702/
Signed-off-by: Jeff Darcy &lt;jdarcy@redhat.com&gt;
Signed-off-by: Shyam &lt;srangana@redhat.com&gt;
Reviewed-on: http://review.gluster.org/10729
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Krishnan Parthasarathi &lt;kparthas@redhat.com&gt;
Reviewed-by: Vijay Bellur &lt;vbellur@redhat.com&gt;
</pre>
</div>
</content>
</entry>
<entry>
<title>glusterd: remove replace brick with data migration support form cli/glusterd</title>
<updated>2015-05-07T07:08:07+00:00</updated>
<author>
<name>Gaurav Kumar Garg</name>
<email>ggarg@redhat.com</email>
</author>
<published>2015-03-27T09:50:03+00:00</published>
<link rel='alternate' type='text/html' href='http://dev.gluster.org/cgit/glusterfs.git/commit/?id=f9ebf5ab3cbec423f75e64c25385125d4b65e31b'/>
<id>f9ebf5ab3cbec423f75e64c25385125d4b65e31b</id>
<content type='text'>
Replace-brick operation with data migration support have been
deprecated from gluster.

With this fix replace brick command will support only one commad

gluster volume replace-brick &lt;VOLNAME&gt; &lt;SOURCE-BRICK&gt; &lt;NEW-BRICK&gt; {commit force}

Change-Id: Ib81d49e5d8e7eaa4ccb5830cfec2bc081191b43b
BUG: 1218602
Signed-off-by: Gaurav Kumar Garg &lt;ggarg@redhat.com&gt;
Reviewed-on: http://review.gluster.org/10577
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Tested-by: NetBSD Build System
Reviewed-by: Kaushal M &lt;kaushal@redhat.com&gt;
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
Replace-brick operation with data migration support have been
deprecated from gluster.

With this fix replace brick command will support only one commad

gluster volume replace-brick &lt;VOLNAME&gt; &lt;SOURCE-BRICK&gt; &lt;NEW-BRICK&gt; {commit force}

Change-Id: Ib81d49e5d8e7eaa4ccb5830cfec2bc081191b43b
BUG: 1218602
Signed-off-by: Gaurav Kumar Garg &lt;ggarg@redhat.com&gt;
Reviewed-on: http://review.gluster.org/10577
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Tested-by: NetBSD Build System
Reviewed-by: Kaushal M &lt;kaushal@redhat.com&gt;
</pre>
</div>
</content>
</entry>
<entry>
<title>tests: Check aux umount is unmounted for quota tests</title>
<updated>2015-05-05T17:42:18+00:00</updated>
<author>
<name>Pranith Kumar K</name>
<email>pkarampu@redhat.com</email>
</author>
<published>2015-05-01T05:31:15+00:00</published>
<link rel='alternate' type='text/html' href='http://dev.gluster.org/cgit/glusterfs.git/commit/?id=abb576afae304e487c010c5bc8e92c6ffae85332'/>
<id>abb576afae304e487c010c5bc8e92c6ffae85332</id>
<content type='text'>
Change-Id: If57d08f3446755ea41f66ca258efcc8ea5a89063
BUG: 1218593
Signed-off-by: Pranith Kumar K &lt;pkarampu@redhat.com&gt;
Reviewed-on: http://review.gluster.org/10480
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Vijay Bellur &lt;vbellur@redhat.com&gt;
Signed-off-by: Sachin Pandit &lt;spandit@redhat.com&gt;
Reviewed-on: http://review.gluster.org/10575
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
Change-Id: If57d08f3446755ea41f66ca258efcc8ea5a89063
BUG: 1218593
Signed-off-by: Pranith Kumar K &lt;pkarampu@redhat.com&gt;
Reviewed-on: http://review.gluster.org/10480
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Vijay Bellur &lt;vbellur@redhat.com&gt;
Signed-off-by: Sachin Pandit &lt;spandit@redhat.com&gt;
Reviewed-on: http://review.gluster.org/10575
</pre>
</div>
</content>
</entry>
<entry>
<title>glusterd: initialize snapd svc at volume restore path</title>
<updated>2015-04-28T08:53:35+00:00</updated>
<author>
<name>Atin Mukherjee</name>
<email>amukherj@redhat.com</email>
</author>
<published>2015-04-20T12:07:21+00:00</published>
<link rel='alternate' type='text/html' href='http://dev.gluster.org/cgit/glusterfs.git/commit/?id=018a0a5b846ed903d5d2545c2c353281e1e9949d'/>
<id>018a0a5b846ed903d5d2545c2c353281e1e9949d</id>
<content type='text'>
In restore path snapd svc was not initialized because of which any glusterd
instance which went down and came back may have uninitialized snapd svc. The
reason I used 'may' is because depending on the nodes in the cluster. In a
single node cluster this wouldn't be a problem since glusterd_spawn_daemon takes
care of initializing it.

Backport of http://review.gluster.org/10304

Change-Id: I2da1e419a0506d3b2742c1cf39a3b9416eb3c305
BUG: 1215518
Signed-off-by: Atin Mukherjee &lt;amukherj@redhat.com&gt;
Reviewed-on: http://review.gluster.org/10304
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Tested-by: NetBSD Build System
Reviewed-by: Kaushal M &lt;kaushal@redhat.com&gt;
(cherry picked from commit 18fd2fdd60839d737ab0ac64f33a444b54bdeee4)
Reviewed-on: http://review.gluster.org/10397
Reviewed-by: Krishnan Parthasarathi &lt;kparthas@redhat.com&gt;
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
In restore path snapd svc was not initialized because of which any glusterd
instance which went down and came back may have uninitialized snapd svc. The
reason I used 'may' is because depending on the nodes in the cluster. In a
single node cluster this wouldn't be a problem since glusterd_spawn_daemon takes
care of initializing it.

Backport of http://review.gluster.org/10304

Change-Id: I2da1e419a0506d3b2742c1cf39a3b9416eb3c305
BUG: 1215518
Signed-off-by: Atin Mukherjee &lt;amukherj@redhat.com&gt;
Reviewed-on: http://review.gluster.org/10304
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Tested-by: NetBSD Build System
Reviewed-by: Kaushal M &lt;kaushal@redhat.com&gt;
(cherry picked from commit 18fd2fdd60839d737ab0ac64f33a444b54bdeee4)
Reviewed-on: http://review.gluster.org/10397
Reviewed-by: Krishnan Parthasarathi &lt;kparthas@redhat.com&gt;
</pre>
</div>
</content>
</entry>
<entry>
<title>glusterd: remove-brick status/stop should not show output for non-existing brick</title>
<updated>2015-04-09T17:21:06+00:00</updated>
<author>
<name>Gaurav Kumar Garg</name>
<email>ggarg@redhat.com</email>
</author>
<published>2015-02-15T13:52:13+00:00</published>
<link rel='alternate' type='text/html' href='http://dev.gluster.org/cgit/glusterfs.git/commit/?id=2788ddd3a0afa98f78128247cca89427a323b090'/>
<id>2788ddd3a0afa98f78128247cca89427a323b090</id>
<content type='text'>
Previously when user start remove-brick operation on a volume then by
giving non-existing brick for remove-brick status/stop command it was
showing remove-brick status/stoping remove-brick operation on a volume.

With this fix it will validate bricks which user have given for
remove-brick status/stop command and if bricks are part of volume then
it will show statistics of remove-brick operation otherwise it will show
error "Incorrect brick &lt;brick_name&gt; for &lt;volume_name&gt;".

Change-Id: I151284ef78c25f52d1b39cdbd71ebfb9eb4b8471
BUG: 1121584
Signed-off-by: Gaurav Kumar Garg &lt;ggarg@redhat.com&gt;
Reviewed-on: http://review.gluster.org/9681
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Krishnan Parthasarathi &lt;kparthas@redhat.com&gt;
Tested-by: Krishnan Parthasarathi &lt;kparthas@redhat.com&gt;
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
Previously when user start remove-brick operation on a volume then by
giving non-existing brick for remove-brick status/stop command it was
showing remove-brick status/stoping remove-brick operation on a volume.

With this fix it will validate bricks which user have given for
remove-brick status/stop command and if bricks are part of volume then
it will show statistics of remove-brick operation otherwise it will show
error "Incorrect brick &lt;brick_name&gt; for &lt;volume_name&gt;".

Change-Id: I151284ef78c25f52d1b39cdbd71ebfb9eb4b8471
BUG: 1121584
Signed-off-by: Gaurav Kumar Garg &lt;ggarg@redhat.com&gt;
Reviewed-on: http://review.gluster.org/9681
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Krishnan Parthasarathi &lt;kparthas@redhat.com&gt;
Tested-by: Krishnan Parthasarathi &lt;kparthas@redhat.com&gt;
</pre>
</div>
</content>
</entry>
<entry>
<title>tests: remove tests for clear-locks</title>
<updated>2015-04-09T09:49:32+00:00</updated>
<author>
<name>Jeff Darcy</name>
<email>jdarcy@redhat.com</email>
</author>
<published>2015-04-08T21:17:13+00:00</published>
<link rel='alternate' type='text/html' href='http://dev.gluster.org/cgit/glusterfs.git/commit/?id=0086a55bb7de1ef5dc7a24583f5fc2b560e835fd'/>
<id>0086a55bb7de1ef5dc7a24583f5fc2b560e835fd</id>
<content type='text'>
These are suspected of causing core dumps during regression tests,
leading to spurious failures.  Per email conversation, since this
isn't a supported feature anyway, the tests are being removed to
facilitate testing of features we do support.

Change-Id: I7fd5c76d26dd6c3ffa91f89fc10469ae3a63afdf
BUG: 1195415
Signed-off-by: Jeff Darcy &lt;jdarcy@redhat.com&gt;
Reviewed-on: http://review.gluster.org/10167
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Kaleb KEITHLEY &lt;kkeithle@redhat.com&gt;
Reviewed-by: Vijay Bellur &lt;vbellur@redhat.com&gt;
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
These are suspected of causing core dumps during regression tests,
leading to spurious failures.  Per email conversation, since this
isn't a supported feature anyway, the tests are being removed to
facilitate testing of features we do support.

Change-Id: I7fd5c76d26dd6c3ffa91f89fc10469ae3a63afdf
BUG: 1195415
Signed-off-by: Jeff Darcy &lt;jdarcy@redhat.com&gt;
Reviewed-on: http://review.gluster.org/10167
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Kaleb KEITHLEY &lt;kkeithle@redhat.com&gt;
Reviewed-by: Vijay Bellur &lt;vbellur@redhat.com&gt;
</pre>
</div>
</content>
</entry>
<entry>
<title>glusterd: gluster command should retrieve current op-version of the NODE</title>
<updated>2015-04-01T12:50:44+00:00</updated>
<author>
<name>Gaurav Kumar Garg</name>
<email>ggarg@redhat.com</email>
</author>
<published>2015-03-06T10:41:16+00:00</published>
<link rel='alternate' type='text/html' href='http://dev.gluster.org/cgit/glusterfs.git/commit/?id=2b949eb89e8ca22d2928c05d549b6f722adf1544'/>
<id>2b949eb89e8ca22d2928c05d549b6f722adf1544</id>
<content type='text'>
Problem:
glusterd was failing to get some specific volume option. for eg:
gluster volume get &lt;vol-name&gt; cluster.op-version

Fix:
glusterd should set count value in dictionary while retrieving specific volume
option.

Change-Id: Iada768ea3d8a0006895525eca2c2dcc40432a4ea
BUG: 1199451
Signed-off-by: Gaurav Kumar Garg &lt;ggarg@redhat.com&gt;
Reviewed-on: http://review.gluster.org/9821
Reviewed-by: Atin Mukherjee &lt;amukherj@redhat.com&gt;
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Kaushal M &lt;kaushal@redhat.com&gt;
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
Problem:
glusterd was failing to get some specific volume option. for eg:
gluster volume get &lt;vol-name&gt; cluster.op-version

Fix:
glusterd should set count value in dictionary while retrieving specific volume
option.

Change-Id: Iada768ea3d8a0006895525eca2c2dcc40432a4ea
BUG: 1199451
Signed-off-by: Gaurav Kumar Garg &lt;ggarg@redhat.com&gt;
Reviewed-on: http://review.gluster.org/9821
Reviewed-by: Atin Mukherjee &lt;amukherj@redhat.com&gt;
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Kaushal M &lt;kaushal@redhat.com&gt;
</pre>
</div>
</content>
</entry>
<entry>
<title>tests: fix online_brick_count</title>
<updated>2015-03-31T06:41:29+00:00</updated>
<author>
<name>Jeff Darcy</name>
<email>jdarcy@redhat.com</email>
</author>
<published>2015-03-19T19:15:07+00:00</published>
<link rel='alternate' type='text/html' href='http://dev.gluster.org/cgit/glusterfs.git/commit/?id=d3eacb0d83834db485061d875d95b4f6af41f30a'/>
<id>d3eacb0d83834db485061d875d95b4f6af41f30a</id>
<content type='text'>
It turns out that "pidof" is unreliable on some platforms (e.g. Fedora
21) because it will show spurious entries for processes using the same
inode under a different name.  Use "pgrep" instead because it's
name-based and doesn't get confused by glusterd/glusterfs being links
to glusterfsd.

Also changed bug-913555.t because it had the same mistake in its own
version of the same function.  Now it uses the common version.

Change-Id: I5d70edd5655faa5470e0f378b8c16a6adacbd4b4
BUG: 1163543
Signed-off-by: Jeff Darcy &lt;jdarcy@redhat.com&gt;
Reviewed-on: http://review.gluster.org/9948
Reviewed-by: Niels de Vos &lt;ndevos@redhat.com&gt;
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Atin Mukherjee &lt;amukherj@redhat.com&gt;
Reviewed-by: Vijay Bellur &lt;vbellur@redhat.com&gt;
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
It turns out that "pidof" is unreliable on some platforms (e.g. Fedora
21) because it will show spurious entries for processes using the same
inode under a different name.  Use "pgrep" instead because it's
name-based and doesn't get confused by glusterd/glusterfs being links
to glusterfsd.

Also changed bug-913555.t because it had the same mistake in its own
version of the same function.  Now it uses the common version.

Change-Id: I5d70edd5655faa5470e0f378b8c16a6adacbd4b4
BUG: 1163543
Signed-off-by: Jeff Darcy &lt;jdarcy@redhat.com&gt;
Reviewed-on: http://review.gluster.org/9948
Reviewed-by: Niels de Vos &lt;ndevos@redhat.com&gt;
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Atin Mukherjee &lt;amukherj@redhat.com&gt;
Reviewed-by: Vijay Bellur &lt;vbellur@redhat.com&gt;
</pre>
</div>
</content>
</entry>
<entry>
<title>cli: volume status for tcp,rdma type volume display only tcp port</title>
<updated>2015-02-18T13:02:12+00:00</updated>
<author>
<name>Mohammed Rafi KC</name>
<email>rkavunga@redhat.com</email>
</author>
<published>2014-11-24T11:37:02+00:00</published>
<link rel='alternate' type='text/html' href='http://dev.gluster.org/cgit/glusterfs.git/commit/?id=692dd7e83cc8a1c85581dda3f752d5ea3b184f8c'/>
<id>692dd7e83cc8a1c85581dda3f752d5ea3b184f8c</id>
<content type='text'>
For tcp,rdma type voumes, there will be two ports, one for tcp
and one for rdma. But volume status command only display tcp port.
By this change, adding an extra column for rdma port and changing
the port to tcp port.

Eg:

&gt;gluster volume status pathy

&gt;For tcp,rdma type volume

Status of volume: patchy
Gluster process                             TCP Port RDMA Port  Online  Pid
------------------------------------------------------------------------------
Brick brickname                             49152    49153      Y       14158

&gt;For rdma type volume

Status of volume: patchy
Gluster process                             TCP Port RDMA Port  Online  Pid
------------------------------------------------------------------------------
Brick brickname                             0        49153      Y       14158

For tcp type volume

Status of volume: patchy
Gluster process                             TCP Port RDMA Port  Online  Pid
------------------------------------------------------------------------------
Brick brickname                             49152    0          Y       14158

&gt;gluster volume status patchy detail

Status of volume: xcube2
------------------------------------------------------------------------------
Brick                : Brick brickname
TCP Port             : 49152
RDMA Port            : 49153
Online               : Y
Pid                  : 14158
File System          : ext4
Device               :
/dev/mapper/luks-2099dd4a-0050-4cae-ad7b-c6a0498c4e88
Mount Options        : rw,seclabel,relatime,data=ordered
Inode Size           : 256
Disk Space Free      : 31.1GB
Total Disk Space     : 47.9GB
Inode Count          : 3203072
Free Inodes          : 2926789

&gt;gluster volume status xcube --xml

&lt;?xml version="1.0" encoding="UTF-8" standalone="yes"?&gt;
&lt;cliOutput&gt;
  &lt;opRet&gt;0&lt;/opRet&gt;
  &lt;opErrno&gt;0&lt;/opErrno&gt;
  &lt;opErrstr&gt;(null)&lt;/opErrstr&gt;
  &lt;volStatus&gt;
    &lt;volumes&gt;
      &lt;volume&gt;
        &lt;volName&gt;xcube&lt;/volName&gt;
        &lt;nodeCount&gt;2&lt;/nodeCount&gt;
        &lt;node&gt;
          &lt;hostname&gt;hostname&lt;/hostname&gt;
          &lt;path&gt;/home/brick1&lt;/path&gt;
          &lt;peerid&gt;2d7bcb95-3d26-4d4f-b3c6-e2ee01b71662&lt;/peerid&gt;
          &lt;status&gt;1&lt;/status&gt;
          &lt;port&gt;49152&lt;/port&gt;
          &lt;ports&gt;
            &lt;tcp&gt;49152&lt;/tcp&gt;
            &lt;rdma&gt;N/A&lt;/rdma&gt;
          &lt;/ports&gt;
          &lt;pid&gt;5657&lt;/pid&gt;
        &lt;/node&gt;
        &lt;node&gt;
          &lt;hostname&gt;NFS Server&lt;/hostname&gt;
          &lt;path&gt;localhost&lt;/path&gt;
          &lt;peerid&gt;2d7bcb95-3d26-4d4f-b3c6-e2ee01b71662&lt;/peerid&gt;
          &lt;status&gt;1&lt;/status&gt;
          &lt;port&gt;2049&lt;/port&gt;
          &lt;ports&gt;
            &lt;tcp&gt;2049&lt;/tcp&gt;
            &lt;rdma&gt;N/A&lt;/rdma&gt;
          &lt;/ports&gt;
          &lt;pid&gt;5665&lt;/pid&gt;
        &lt;/node&gt;
        &lt;tasks/&gt;
      &lt;/volume&gt;
    &lt;/volumes&gt;
  &lt;/volStatus&gt;
&lt;/cliOutput&gt;

Change-Id: I81aab226edbd400d29cd3f510af4f344dd99ba51
BUG: 1164079
Signed-off-by: Mohammed Rafi KC &lt;rkavunga@redhat.com&gt;
Reviewed-on: http://review.gluster.org/9191
Reviewed-by: Atin Mukherjee &lt;amukherj@redhat.com&gt;
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Kaushal M &lt;kaushal@redhat.com&gt;
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
For tcp,rdma type voumes, there will be two ports, one for tcp
and one for rdma. But volume status command only display tcp port.
By this change, adding an extra column for rdma port and changing
the port to tcp port.

Eg:

&gt;gluster volume status pathy

&gt;For tcp,rdma type volume

Status of volume: patchy
Gluster process                             TCP Port RDMA Port  Online  Pid
------------------------------------------------------------------------------
Brick brickname                             49152    49153      Y       14158

&gt;For rdma type volume

Status of volume: patchy
Gluster process                             TCP Port RDMA Port  Online  Pid
------------------------------------------------------------------------------
Brick brickname                             0        49153      Y       14158

For tcp type volume

Status of volume: patchy
Gluster process                             TCP Port RDMA Port  Online  Pid
------------------------------------------------------------------------------
Brick brickname                             49152    0          Y       14158

&gt;gluster volume status patchy detail

Status of volume: xcube2
------------------------------------------------------------------------------
Brick                : Brick brickname
TCP Port             : 49152
RDMA Port            : 49153
Online               : Y
Pid                  : 14158
File System          : ext4
Device               :
/dev/mapper/luks-2099dd4a-0050-4cae-ad7b-c6a0498c4e88
Mount Options        : rw,seclabel,relatime,data=ordered
Inode Size           : 256
Disk Space Free      : 31.1GB
Total Disk Space     : 47.9GB
Inode Count          : 3203072
Free Inodes          : 2926789

&gt;gluster volume status xcube --xml

&lt;?xml version="1.0" encoding="UTF-8" standalone="yes"?&gt;
&lt;cliOutput&gt;
  &lt;opRet&gt;0&lt;/opRet&gt;
  &lt;opErrno&gt;0&lt;/opErrno&gt;
  &lt;opErrstr&gt;(null)&lt;/opErrstr&gt;
  &lt;volStatus&gt;
    &lt;volumes&gt;
      &lt;volume&gt;
        &lt;volName&gt;xcube&lt;/volName&gt;
        &lt;nodeCount&gt;2&lt;/nodeCount&gt;
        &lt;node&gt;
          &lt;hostname&gt;hostname&lt;/hostname&gt;
          &lt;path&gt;/home/brick1&lt;/path&gt;
          &lt;peerid&gt;2d7bcb95-3d26-4d4f-b3c6-e2ee01b71662&lt;/peerid&gt;
          &lt;status&gt;1&lt;/status&gt;
          &lt;port&gt;49152&lt;/port&gt;
          &lt;ports&gt;
            &lt;tcp&gt;49152&lt;/tcp&gt;
            &lt;rdma&gt;N/A&lt;/rdma&gt;
          &lt;/ports&gt;
          &lt;pid&gt;5657&lt;/pid&gt;
        &lt;/node&gt;
        &lt;node&gt;
          &lt;hostname&gt;NFS Server&lt;/hostname&gt;
          &lt;path&gt;localhost&lt;/path&gt;
          &lt;peerid&gt;2d7bcb95-3d26-4d4f-b3c6-e2ee01b71662&lt;/peerid&gt;
          &lt;status&gt;1&lt;/status&gt;
          &lt;port&gt;2049&lt;/port&gt;
          &lt;ports&gt;
            &lt;tcp&gt;2049&lt;/tcp&gt;
            &lt;rdma&gt;N/A&lt;/rdma&gt;
          &lt;/ports&gt;
          &lt;pid&gt;5665&lt;/pid&gt;
        &lt;/node&gt;
        &lt;tasks/&gt;
      &lt;/volume&gt;
    &lt;/volumes&gt;
  &lt;/volStatus&gt;
&lt;/cliOutput&gt;

Change-Id: I81aab226edbd400d29cd3f510af4f344dd99ba51
BUG: 1164079
Signed-off-by: Mohammed Rafi KC &lt;rkavunga@redhat.com&gt;
Reviewed-on: http://review.gluster.org/9191
Reviewed-by: Atin Mukherjee &lt;amukherj@redhat.com&gt;
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Kaushal M &lt;kaushal@redhat.com&gt;
</pre>
</div>
</content>
</entry>
<entry>
<title>cluster/afr: Enable auto-quorum for replicate with odd number of bricks</title>
<updated>2015-02-09T14:03:43+00:00</updated>
<author>
<name>Pranith Kumar K</name>
<email>pkarampu@redhat.com</email>
</author>
<published>2015-01-31T10:37:03+00:00</published>
<link rel='alternate' type='text/html' href='http://dev.gluster.org/cgit/glusterfs.git/commit/?id=efd93f0b2b42f8a8b151bf694fdd340fbcb22ec0'/>
<id>efd93f0b2b42f8a8b151bf694fdd340fbcb22ec0</id>
<content type='text'>
Change-Id: I908934f1f22cf7d2d0ceccc0dedf28a69861997f
BUG: 1187885
Signed-off-by: Pranith Kumar K &lt;pkarampu@redhat.com&gt;
Reviewed-on: http://review.gluster.org/9517
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Krutika Dhananjay &lt;kdhananj@redhat.com&gt;
Reviewed-by: Anuradha Talur &lt;atalur@redhat.com&gt;
Reviewed-by: Ravishankar N &lt;ravishankar@redhat.com&gt;
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
Change-Id: I908934f1f22cf7d2d0ceccc0dedf28a69861997f
BUG: 1187885
Signed-off-by: Pranith Kumar K &lt;pkarampu@redhat.com&gt;
Reviewed-on: http://review.gluster.org/9517
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Krutika Dhananjay &lt;kdhananj@redhat.com&gt;
Reviewed-by: Anuradha Talur &lt;atalur@redhat.com&gt;
Reviewed-by: Ravishankar N &lt;ravishankar@redhat.com&gt;
</pre>
</div>
</content>
</entry>
</feed>
