<feed xmlns='http://www.w3.org/2005/Atom'>
<title>glusterfs.git/heal, branch v4.1.0rc0</title>
<subtitle></subtitle>
<link rel='alternate' type='text/html' href='http://dev.gluster.org/cgit/glusterfs.git/'/>
<entry>
<title>glfs heal binary change to accomodate socket file arguments</title>
<updated>2018-03-27T02:24:11+00:00</updated>
<author>
<name>Vishal Pandey</name>
<email>vpandey@redhat.com</email>
</author>
<published>2018-03-26T06:35:25+00:00</published>
<link rel='alternate' type='text/html' href='http://dev.gluster.org/cgit/glusterfs.git/commit/?id=25fc63912023919e4eaea918810bcb91ee07db46'/>
<id>25fc63912023919e4eaea918810bcb91ee07db46</id>
<content type='text'>
Change-Id: I755d6552decd015aec7859ad2cf99c76c8bee9dc
fixes: bz#1558380
BUG: 1558380
Signed-off-by: Vishal Pandey &lt;vpandey@redhat.com&gt;
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
Change-Id: I755d6552decd015aec7859ad2cf99c76c8bee9dc
fixes: bz#1558380
BUG: 1558380
Signed-off-by: Vishal Pandey &lt;vpandey@redhat.com&gt;
</pre>
</div>
</content>
</entry>
<entry>
<title>glusterd: TLS verification fails while using intermediate CA</title>
<updated>2018-03-19T19:00:03+00:00</updated>
<author>
<name>Mohit Agrawal</name>
<email>moagrawa@redhat.com</email>
</author>
<published>2018-03-14T04:07:52+00:00</published>
<link rel='alternate' type='text/html' href='http://dev.gluster.org/cgit/glusterfs.git/commit/?id=cf06dd544004701ef43fa81c5b7a95353d5c1d65'/>
<id>cf06dd544004701ef43fa81c5b7a95353d5c1d65</id>
<content type='text'>
Problem: TLS verification fails while using intermediate CA
         if mgmt SSL is enabled.

Solution: There are two main issue of TLS verification failing
          1) not calling ssl_api to set cert_depth
          2) The current code does not allow to set certificate depth
             while MGMT SSL is enabled.
          After apply this patch to set certificate depth user
          need to set parameter option transport.socket.ssl-cert-depth &lt;depth&gt;
          in /var/lib/glusterd/secure_acccess instead to set in
          /etc/glusterfs/glusterd.vol. At the time of set secure_mgmt in ctx
          we will check the value of cert-depth and save the value of cert-depth
          in ctx.If user does not provide any value in cert-depth in that case
          it will consider default value is 1

BUG: 1555154
Change-Id: I89e9a9e1026e37efb5c20f9ec62b1989ef644f35
Signed-off-by: Mohit Agrawal &lt;moagrawa@redhat.com&gt;
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
Problem: TLS verification fails while using intermediate CA
         if mgmt SSL is enabled.

Solution: There are two main issue of TLS verification failing
          1) not calling ssl_api to set cert_depth
          2) The current code does not allow to set certificate depth
             while MGMT SSL is enabled.
          After apply this patch to set certificate depth user
          need to set parameter option transport.socket.ssl-cert-depth &lt;depth&gt;
          in /var/lib/glusterd/secure_acccess instead to set in
          /etc/glusterfs/glusterd.vol. At the time of set secure_mgmt in ctx
          we will check the value of cert-depth and save the value of cert-depth
          in ctx.If user does not provide any value in cert-depth in that case
          it will consider default value is 1

BUG: 1555154
Change-Id: I89e9a9e1026e37efb5c20f9ec62b1989ef644f35
Signed-off-by: Mohit Agrawal &lt;moagrawa@redhat.com&gt;
</pre>
</div>
</content>
</entry>
<entry>
<title>build: add --without-server option</title>
<updated>2018-02-19T17:38:55+00:00</updated>
<author>
<name>Niels de Vos</name>
<email>ndevos@redhat.com</email>
</author>
<published>2018-01-31T15:38:34+00:00</published>
<link rel='alternate' type='text/html' href='http://dev.gluster.org/cgit/glusterfs.git/commit/?id=8d9961e63786cf8a9bb9e2f4140ab9a77f1ccbe3'/>
<id>8d9961e63786cf8a9bb9e2f4140ab9a77f1ccbe3</id>
<content type='text'>
With Gluster 4.0 we will not provide the server components for EL6 and
older. At one point Gluster 4.x will get GlusterD2, which requires
Golang tools in the distribution. EL6 does not contain these at the
moment.

With this change, it is possible to `./configure --without-server` which
prevents building glusterd and the xlators for the bricks. Building RPMs
can pass `--without server` and the glusterfs-server sub-package will
not be created.

Change-Id: I97f5ccf9f2c76e60d9af83915fc59fae57ad6d25
BUG: 1074947
Signed-off-by: Niels de Vos &lt;ndevos@redhat.com&gt;
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
With Gluster 4.0 we will not provide the server components for EL6 and
older. At one point Gluster 4.x will get GlusterD2, which requires
Golang tools in the distribution. EL6 does not contain these at the
moment.

With this change, it is possible to `./configure --without-server` which
prevents building glusterd and the xlators for the bricks. Building RPMs
can pass `--without server` and the glusterfs-server sub-package will
not be created.

Change-Id: I97f5ccf9f2c76e60d9af83915fc59fae57ad6d25
BUG: 1074947
Signed-off-by: Niels de Vos &lt;ndevos@redhat.com&gt;
</pre>
</div>
</content>
</entry>
<entry>
<title>heal: Fix coverity scan dead-code report</title>
<updated>2017-11-30T12:22:00+00:00</updated>
<author>
<name>Pranith Kumar K</name>
<email>pkarampu@redhat.com</email>
</author>
<published>2017-11-27T09:37:05+00:00</published>
<link rel='alternate' type='text/html' href='http://dev.gluster.org/cgit/glusterfs.git/commit/?id=7ae126dcd967a953a64102a96de8e0678331e548'/>
<id>7ae126dcd967a953a64102a96de8e0678331e548</id>
<content type='text'>
Got following kind of scan issues:

*** CID 1382469:    (DEADCODE)
/heal/src/glfs-heal.c: 1749 in main()
1744                             break;
1745             case GF_SHD_OP_SBRAIN_HEAL_FROM_BRICK:
1746                     ret = glfsh_heal_from_brick (fs, top_subvol, &amp;rootloc,
1747                                                  hostname, path, file);
1748                     break;
&gt;&gt;&gt;     CID 1382469:    (DEADCODE)
&gt;&gt;&gt;     Execution cannot reach this statement: "case GF_SHD_OP_INVALID:".

Change-Id: I184da785893cb3e281b21f2d4fcdd2e5452102e6
Signed-off-by: Pranith Kumar K &lt;pkarampu@redhat.com&gt;
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
Got following kind of scan issues:

*** CID 1382469:    (DEADCODE)
/heal/src/glfs-heal.c: 1749 in main()
1744                             break;
1745             case GF_SHD_OP_SBRAIN_HEAL_FROM_BRICK:
1746                     ret = glfsh_heal_from_brick (fs, top_subvol, &amp;rootloc,
1747                                                  hostname, path, file);
1748                     break;
&gt;&gt;&gt;     CID 1382469:    (DEADCODE)
&gt;&gt;&gt;     Execution cannot reach this statement: "case GF_SHD_OP_INVALID:".

Change-Id: I184da785893cb3e281b21f2d4fcdd2e5452102e6
Signed-off-by: Pranith Kumar K &lt;pkarampu@redhat.com&gt;
</pre>
</div>
</content>
</entry>
<entry>
<title>cluster/afr: Print heal info summary output in stream fashion</title>
<updated>2017-11-22T10:12:28+00:00</updated>
<author>
<name>karthik-us</name>
<email>ksubrahm@redhat.com</email>
</author>
<published>2017-11-21T13:31:36+00:00</published>
<link rel='alternate' type='text/html' href='http://dev.gluster.org/cgit/glusterfs.git/commit/?id=77e3bc671aab2fda68ada53f38ec368b20675f59'/>
<id>77e3bc671aab2fda68ada53f38ec368b20675f59</id>
<content type='text'>
Problem:
The heal info summary was printing the output at the end after
crawling for pending heal entries completes on all the bricks.

Fix:
Printing the output immediately after the crawl on individual brick
completes, so that it won't give the impression of CLI being hung.

Change-Id: Ieaf5718736a7ee6837bac02bd30a95836e605dab
BUG: 1506104
Signed-off-by: karthik-us &lt;ksubrahm@redhat.com&gt;
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
Problem:
The heal info summary was printing the output at the end after
crawling for pending heal entries completes on all the bricks.

Fix:
Printing the output immediately after the crawl on individual brick
completes, so that it won't give the impression of CLI being hung.

Change-Id: Ieaf5718736a7ee6837bac02bd30a95836e605dab
BUG: 1506104
Signed-off-by: karthik-us &lt;ksubrahm@redhat.com&gt;
</pre>
</div>
</content>
</entry>
<entry>
<title>cluster/afr: Print heal info split-brain output in stream fashion</title>
<updated>2017-11-13T03:44:12+00:00</updated>
<author>
<name>karthik-us</name>
<email>ksubrahm@redhat.com</email>
</author>
<published>2017-10-25T07:12:23+00:00</published>
<link rel='alternate' type='text/html' href='http://dev.gluster.org/cgit/glusterfs.git/commit/?id=05f9c13f4d69e4113f5a851f4097ef35ba3f33b2'/>
<id>05f9c13f4d69e4113f5a851f4097ef35ba3f33b2</id>
<content type='text'>
Problem:
When we trigger the heal info split-brain command the o/p is not
streamed as it is received, but dumped at the end for all the bricks
together. This gives a perception that the command is hung.

Fix:
When we get a split brain entry while crawling throught the pending
heal entries, flush that immediately so that it prints the output
in a stream fashion and doesn't look like the cli is hung.

Change-Id: I7547e86b83202d66616749b8b31d4d0dff0abf07
BUG: 1506104
Signed-off-by: karthik-us &lt;ksubrahm@redhat.com&gt;
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
Problem:
When we trigger the heal info split-brain command the o/p is not
streamed as it is received, but dumped at the end for all the bricks
together. This gives a perception that the command is hung.

Fix:
When we get a split brain entry while crawling throught the pending
heal entries, flush that immediately so that it prints the output
in a stream fashion and doesn't look like the cli is hung.

Change-Id: I7547e86b83202d66616749b8b31d4d0dff0abf07
BUG: 1506104
Signed-off-by: karthik-us &lt;ksubrahm@redhat.com&gt;
</pre>
</div>
</content>
</entry>
<entry>
<title>heal: Fix coverity errors</title>
<updated>2017-11-02T11:03:37+00:00</updated>
<author>
<name>Pranith Kumar K</name>
<email>pkarampu@redhat.com</email>
</author>
<published>2017-10-31T11:39:45+00:00</published>
<link rel='alternate' type='text/html' href='http://dev.gluster.org/cgit/glusterfs.git/commit/?id=84f4f68bfafd3f7fc14a1e83e0272579e5cc1519'/>
<id>84f4f68bfafd3f7fc14a1e83e0272579e5cc1519</id>
<content type='text'>
Change-Id: Idbebfc2e4cfa739a87ccaabbc6e2088362b6bbc1
Signed-off-by: Pranith Kumar K &lt;pkarampu@redhat.com&gt;
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
Change-Id: Idbebfc2e4cfa739a87ccaabbc6e2088362b6bbc1
Signed-off-by: Pranith Kumar K &lt;pkarampu@redhat.com&gt;
</pre>
</div>
</content>
</entry>
<entry>
<title>heal: New feature heal info summary to list the status of brick and count of entries to be healed</title>
<updated>2017-09-15T12:21:36+00:00</updated>
<author>
<name>Mohamed Ashiq Liyazudeen</name>
<email>mliyazud@redhat.com</email>
</author>
<published>2016-08-10T08:14:19+00:00</published>
<link rel='alternate' type='text/html' href='http://dev.gluster.org/cgit/glusterfs.git/commit/?id=dabfc3116ae59b331f967c090c0c51eb3f0355fd'/>
<id>dabfc3116ae59b331f967c090c0c51eb3f0355fd</id>
<content type='text'>
Command output:
Brick 192.168.2.8:/brick/1
Status: Connected
Total Number of entries: 363
Number of entries in heal pending: 362
Number of entries in split-brain: 0
Number of entries possibly healing: 1

&lt;?xml version="1.0" encoding="UTF-8" standalone="yes"?&gt;
&lt;cliOutput&gt;
  &lt;healInfo&gt;
    &lt;bricks&gt;
      &lt;brick hostUuid="9105dd4b-eca8-4fdb-85b2-b81cdf77eda3"&gt;
        &lt;name&gt;192.168.2.8:/brick/1&lt;/name&gt;
        &lt;status&gt;Connected&lt;/status&gt;
        &lt;totalNumberOfEntries&gt;363&lt;/numberOfEntries&gt;
        &lt;numberOfEntriesInHealPending&gt;362&lt;/numberOfEntriesInHealPending&gt;
        &lt;numberOfEntriesInSplitBrain&gt;0&lt;/numberOfEntriesInSplitBrain&gt;
        &lt;numberOfEntriesPossiblyHealing&gt;1&lt;/numberOfEntriesPossiblyHealing&gt;
      &lt;/brick&gt;
    &lt;/bricks&gt;
  &lt;/healInfo&gt;
  &lt;opRet&gt;0&lt;/opRet&gt;
  &lt;opErrno&gt;0&lt;/opErrno&gt;
  &lt;opErrstr/&gt;
&lt;/cliOutput&gt;

Change-Id: I40cb6f77a14131c9e41b292f4901b41a228863d7
BUG: 1261463
Signed-off-by: Mohamed Ashiq Liyazudeen &lt;mliyazud@redhat.com&gt;
Reviewed-on: https://review.gluster.org/12154
Smoke: Gluster Build System &lt;jenkins@build.gluster.org&gt;
Tested-by: Karthik U S &lt;ksubrahm@redhat.com&gt;
CentOS-regression: Gluster Build System &lt;jenkins@build.gluster.org&gt;
Reviewed-by: Ravishankar N &lt;ravishankar@redhat.com&gt;
Reviewed-by: Atin Mukherjee &lt;amukherj@redhat.com&gt;
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
Command output:
Brick 192.168.2.8:/brick/1
Status: Connected
Total Number of entries: 363
Number of entries in heal pending: 362
Number of entries in split-brain: 0
Number of entries possibly healing: 1

&lt;?xml version="1.0" encoding="UTF-8" standalone="yes"?&gt;
&lt;cliOutput&gt;
  &lt;healInfo&gt;
    &lt;bricks&gt;
      &lt;brick hostUuid="9105dd4b-eca8-4fdb-85b2-b81cdf77eda3"&gt;
        &lt;name&gt;192.168.2.8:/brick/1&lt;/name&gt;
        &lt;status&gt;Connected&lt;/status&gt;
        &lt;totalNumberOfEntries&gt;363&lt;/numberOfEntries&gt;
        &lt;numberOfEntriesInHealPending&gt;362&lt;/numberOfEntriesInHealPending&gt;
        &lt;numberOfEntriesInSplitBrain&gt;0&lt;/numberOfEntriesInSplitBrain&gt;
        &lt;numberOfEntriesPossiblyHealing&gt;1&lt;/numberOfEntriesPossiblyHealing&gt;
      &lt;/brick&gt;
    &lt;/bricks&gt;
  &lt;/healInfo&gt;
  &lt;opRet&gt;0&lt;/opRet&gt;
  &lt;opErrno&gt;0&lt;/opErrno&gt;
  &lt;opErrstr/&gt;
&lt;/cliOutput&gt;

Change-Id: I40cb6f77a14131c9e41b292f4901b41a228863d7
BUG: 1261463
Signed-off-by: Mohamed Ashiq Liyazudeen &lt;mliyazud@redhat.com&gt;
Reviewed-on: https://review.gluster.org/12154
Smoke: Gluster Build System &lt;jenkins@build.gluster.org&gt;
Tested-by: Karthik U S &lt;ksubrahm@redhat.com&gt;
CentOS-regression: Gluster Build System &lt;jenkins@build.gluster.org&gt;
Reviewed-by: Ravishankar N &lt;ravishankar@redhat.com&gt;
Reviewed-by: Atin Mukherjee &lt;amukherj@redhat.com&gt;
</pre>
</div>
</content>
</entry>
<entry>
<title>cluster/afr: GFID split-brain resolution with existing CLI</title>
<updated>2017-07-18T15:24:54+00:00</updated>
<author>
<name>karthik-us</name>
<email>ksubrahm@redhat.com</email>
</author>
<published>2017-06-07T10:26:13+00:00</published>
<link rel='alternate' type='text/html' href='http://dev.gluster.org/cgit/glusterfs.git/commit/?id=657d78dbad118e511e1fca8b1badb9f8ae7a6f60'/>
<id>657d78dbad118e511e1fca8b1badb9f8ae7a6f60</id>
<content type='text'>
Problem:
Currently there is no way for the admin from CLI to resolve gfid
split-brain based on some policy like choice of the brick, mtime
or size.

Fix:
With the existing CLI options based on size, mtime, and choice of
brick, we do lookup on the parent for the specified file. As
part of the lookup, if we find gfid mismatch, we resolve them
based on the policy and return. If the file is not in gfid split-
brain, then we check for the data and metadata split-brain in the
getxattr code path, and resolve if any.

This will work provided absolute path to the file with the CLI
and not with gfid of the file. Hence the source-brick policy
without any file path will also not resolve the gfid split-brain
since it uses the gfid of the files. But it can resolve any other
type of split-brains and skip the gfid mismatch resolution with
the usual error message.

Reverting the change https://review.gluster.org/17290. This patch
resolves the issue.

Fixes gluster/glusterfs#135

Change-Id: Iaeba6fc32f184a34255d03be87cda02773130a09
BUG: 1459530
Signed-off-by: karthik-us &lt;ksubrahm@redhat.com&gt;
Reviewed-on: https://review.gluster.org/17485
Reviewed-by: Ravishankar N &lt;ravishankar@redhat.com&gt;
Reviewed-by: Pranith Kumar Karampuri &lt;pkarampu@redhat.com&gt;
CentOS-regression: Gluster Build System &lt;jenkins@build.gluster.org&gt;
Smoke: Gluster Build System &lt;jenkins@build.gluster.org&gt;
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
Problem:
Currently there is no way for the admin from CLI to resolve gfid
split-brain based on some policy like choice of the brick, mtime
or size.

Fix:
With the existing CLI options based on size, mtime, and choice of
brick, we do lookup on the parent for the specified file. As
part of the lookup, if we find gfid mismatch, we resolve them
based on the policy and return. If the file is not in gfid split-
brain, then we check for the data and metadata split-brain in the
getxattr code path, and resolve if any.

This will work provided absolute path to the file with the CLI
and not with gfid of the file. Hence the source-brick policy
without any file path will also not resolve the gfid split-brain
since it uses the gfid of the files. But it can resolve any other
type of split-brains and skip the gfid mismatch resolution with
the usual error message.

Reverting the change https://review.gluster.org/17290. This patch
resolves the issue.

Fixes gluster/glusterfs#135

Change-Id: Iaeba6fc32f184a34255d03be87cda02773130a09
BUG: 1459530
Signed-off-by: karthik-us &lt;ksubrahm@redhat.com&gt;
Reviewed-on: https://review.gluster.org/17485
Reviewed-by: Ravishankar N &lt;ravishankar@redhat.com&gt;
Reviewed-by: Pranith Kumar Karampuri &lt;pkarampu@redhat.com&gt;
CentOS-regression: Gluster Build System &lt;jenkins@build.gluster.org&gt;
Smoke: Gluster Build System &lt;jenkins@build.gluster.org&gt;
</pre>
</div>
</content>
</entry>
<entry>
<title>glfsheal: prevent background self-heals</title>
<updated>2017-06-30T11:08:33+00:00</updated>
<author>
<name>Ravishankar N</name>
<email>ravishankar@redhat.com</email>
</author>
<published>2017-06-25T16:20:09+00:00</published>
<link rel='alternate' type='text/html' href='http://dev.gluster.org/cgit/glusterfs.git/commit/?id=b4db625d0ccb4fdc6537ed9f6e8ebeaffd1c4873'/>
<id>b4db625d0ccb4fdc6537ed9f6e8ebeaffd1c4873</id>
<content type='text'>
Problem:
For a file in gfid split-brain, the parent directory ('/' during
testing) was detected as possibly undergoing heal instead of split-brain
in `heal-info` output. Also, it was not being displayed in `info
split-brain` output for the same reason. The problem was that when `glfsheal`
was run, lookup on '/' triggered a background self-heal due to which processing
of '/' during `heal info` failed to acquire locks with errno=EAGAIN.

Fix:
Set background-self-heal-count to zero while launching glfsheal.

Change-Id: I153a7c75af71f213a4eefacf504a0f9806c528a5
BUG: 1318895
Signed-off-by: Ravishankar N &lt;ravishankar@redhat.com&gt;
Reviewed-on: https://review.gluster.org/13772
CentOS-regression: Gluster Build System &lt;jenkins@build.gluster.org&gt;
NetBSD-regression: NetBSD Build System &lt;jenkins@build.gluster.org&gt;
Smoke: Gluster Build System &lt;jenkins@build.gluster.org&gt;
Reviewed-by: Jeff Darcy &lt;jeff@pl.atyp.us&gt;
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
Problem:
For a file in gfid split-brain, the parent directory ('/' during
testing) was detected as possibly undergoing heal instead of split-brain
in `heal-info` output. Also, it was not being displayed in `info
split-brain` output for the same reason. The problem was that when `glfsheal`
was run, lookup on '/' triggered a background self-heal due to which processing
of '/' during `heal info` failed to acquire locks with errno=EAGAIN.

Fix:
Set background-self-heal-count to zero while launching glfsheal.

Change-Id: I153a7c75af71f213a4eefacf504a0f9806c528a5
BUG: 1318895
Signed-off-by: Ravishankar N &lt;ravishankar@redhat.com&gt;
Reviewed-on: https://review.gluster.org/13772
CentOS-regression: Gluster Build System &lt;jenkins@build.gluster.org&gt;
NetBSD-regression: NetBSD Build System &lt;jenkins@build.gluster.org&gt;
Smoke: Gluster Build System &lt;jenkins@build.gluster.org&gt;
Reviewed-by: Jeff Darcy &lt;jeff@pl.atyp.us&gt;
</pre>
</div>
</content>
</entry>
</feed>
