summaryrefslogtreecommitdiffstats
path: root/doc/release-notes/4.1.0.md
diff options
context:
space:
mode:
Diffstat (limited to 'doc/release-notes/4.1.0.md')
-rw-r--r--doc/release-notes/4.1.0.md459
1 files changed, 0 insertions, 459 deletions
diff --git a/doc/release-notes/4.1.0.md b/doc/release-notes/4.1.0.md
deleted file mode 100644
index f05b845f0ac..00000000000
--- a/doc/release-notes/4.1.0.md
+++ /dev/null
@@ -1,459 +0,0 @@
-# Release notes for Gluster 4.1.0
-
-This is a major release that includes a range of features enhancing management,
-performance, monitoring, and providing newer functionality like thin arbiters,
-cloud archival, time consistency. It also contains several bug fixes.
-
-A selection of the important features and changes are documented on this page.
-A full list of bugs that have been addressed is included further below.
-
-- [Announcements](#announcements)
-- [Major changes and features](#major-changes-and-features)
-- [Major issues](#major-issues)
-- [Bugs addressed in the release](#bugs-addressed)
-
-## Announcements
-
-1. As 4.0 was a short term maintenance release, features which have been
-included in that release are available with 4.1.0 as well. These features may
-be of interest to users upgrading to 4.1.0 from older than 4.0 releases. The 4.0
-[release notes](http://docs.gluster.org/en/latest/release-notes/) captures the list of features that were introduced with 4.0.
-
-**NOTE:** As 4.0 was a short term maintenance release, it will reach end of
-life (EOL) with the release of 4.1.0. ([reference](https://www.gluster.org/release-schedule/))
-
-2. Releases that receive maintenance updates post 4.1 release are, 3.12, and
-4.1 ([reference](https://www.gluster.org/release-schedule/))
-
-**NOTE:** 3.10 long term maintenance release, will reach end of life (EOL) with
-the release of 4.1.0. ([reference](https://www.gluster.org/release-schedule/))
-
-3. Continuing with this release, the CentOS storage SIG will not build server
-packages for CentOS6. Server packages will be available for CentOS7 only. For
-ease of migrations, client packages on CentOS6 will be published and maintained.
-
-**NOTE**: This change was announced [here](http://lists.gluster.org/pipermail/gluster-users/2018-January/033212.html)
-
-## Major changes and features
-
-Features are categorized into the following sections,
-
-- [Management](#management)
-- [Monitoring](#monitoring)
-- [Performance](#performance)
-- [Standalone](#standalone)
-- [Developer related](#developer-related)
-
-### Management
-
-#### GlusterD2
-
-> **IMP:** GlusterD2 in Gluster-4.1.0 is still considered a preview and is
-> experimental. It should not be considered for production use. Users should
-> still expect breaking changes to be possible, though efforts will be taken to
-> avoid such changes. As GD2 is still under heavy development, new features can
-> be expected throughout the 4.1 release.
-
-GD2 brings initial support for rebalance, snapshots, intelligent volume
-provisioning and a lot of other bug fixes and internal changes.
-
-##### Rebalance [#786](https://github.com/gluster/glusterd2/pull/786)
-
-GD2 supports running rebalance on volumes. Supported rebalance operations include,
-- rebalance start
- - rebalance start with fix-layout
-- rebalance stop
-- rebalance status
-
-Support only exists in the ReST API right now. CLI support will be introduced in subsequent releases.
-
-##### Snapshot [#533](https://github.com/gluster/glusterd2/pull/533)
-
-Initial support for volume snapshot has been introduced. At the moment, snapshots are supported only on Thin-LVM bricks.
-
-Support snapshot operations include,
-- create
-- activate/deactivate
-- list
-- info
-
-##### Intelligent volume provisioning (IVP) [#661](https://github.com/gluster/glusterd2/pull/661)
-
-GD2 brings very early preview for intelligent volume creation, similar to
-[Heketi](https://github.com/heketi/heketi).
-
-> **IMP:** This is considered experimental, and the API and implementation is
-> not final. It is very possible that both the API and the implementation will
-> change.
-
-IVP enables users to create volumes by just providing the expected volume type
-and a size, without providing the bricks layout. IVP is supported in CLI in the
-normal `volume create` command.
-
-More information on IVP can be found in the pull-request.
-
-To support IVP, support for adding and managing block devices, and basic support
-for zones is available. [#783](https://github.com/gluster/glusterd2/pull/783) [#785](https://github.com/gluster/glusterd2/pull/785)
-
-##### Other changes
-
-Other notable changes include,
-- Support for volume option levels (experimental, advanced, deprecated) [#591](https://github.com/gluster/glusterd2/pull/591)
-- Support for resetting volume options [#545](https://github.com/gluster/glusterd2/pull/545)
-- Option hooks for volume set [#708](https://github.com/gluster/glusterd2/pull/708)
-- Support for setting quota options [#583](https://github.com/gluster/glusterd2/pull/583)
-- Changes to transaction locking [#808](https://github.com/gluster/glusterd2/pull/808)
-- Support for setting metadata on peers and volume [#600](https://github.com/gluster/glusterd2/pull/600) [#689](https://github.com/gluster/glusterd2/pull/689) [#704](https://github.com/gluster/glusterd2/pull/704)
-- Thin arbiter support [#673](https://github.com/gluster/glusterd2/pull/673) [#702](https://github.com/gluster/glusterd2/pull/702)
-
-In addition to the above, a lot of smaller bug-fixes and enhancements to internal frameworks and tests have also been done.
-
-##### Known issues
-
-GD2 is still under heavy development and has lots of known bugs. For filing new bugs or tracking known bugs, please use the [GD2 github issue tracker](http://github.com/gluster/glusterd2/issues?q=is%3Aissue+is%3Aopen+label%3Abug).
-
-#### 2. Changes to gluster based smb.conf share management
-
-Previously Gluster used to delete the entire volume share section from smb.conf
-either after volume is stopped or while disabling user.cifs/user.smb volume set
-options. With this release those volume share sections, that were added by
-samba hook scripts inside smb.conf, will not get removed post a volume stop or
-on disabling user.cifs/user.smb volume set options. Instead we add the following
-share specific smb.conf parameter to the end of corresponding volume share
-section to make it unavailable for client access:
-
-```
-available = no
-```
-
-This will make sure that the additional smb.conf parameters configured
-externally are retained. For more details on the above parameter search under
-"available (S)" at [smb.conf(5)](https://www.samba.org/samba/docs/current/man-html/smb.conf.5.html) manual page.
-
-### Monitoring
-
-Various xlators are enhanced to provide additional metrics, that help in
-determining the effectiveness of the xlator in various workloads.
-
-These metrics can be dumped and visualized as detailed [here](https://docs.gluster.org/en/latest/release-notes/4.0.0/#monitoring).
-
-#### 1. Additional metrics added to negative lookup cache xlator
-
-Metrics added are:
- - negative_lookup_hit_count
- - negative_lookup_miss_count
- - get_real_filename_hit_count
- - get_real_filename_miss_count
- - nameless_lookup_count
- - inodes_with_positive_dentry_cache
- - inodes_with_negative_dentry_cache
- - dentry_invalidations_recieved
- - cache_limit
- - consumed_cache_size
- - inode_limit
- - consumed_inodes
-
-#### 2. Additional metrics added to md-cache xlator
-
-Metrics added are:
- - stat_cache_hit_count
- - stat_cache_miss_count
- - xattr_cache_hit_count
- - xattr_cache_miss_count
- - nameless_lookup_count
- - negative_lookup_count
- - stat_cache_invalidations_received
- - xattr_cache_invalidations_received
-
-#### 3. Additional metrics added to quick-read xlator
-
-Metrics added are:
- - total_files_cached
- - total_cache_used
- - cache-hit
- - cache-miss
- - cache-invalidations
-
-### Performance
-
-#### 1. Support for fuse writeback cache
-
-Gluster FUSE mounts support FUSE extension to leverage the kernel
-"writeback cache".
-
-For usage help see `man 8 glusterfs` and `man 8 mount.glusterfs`, specifically
-the options `-kernel-writeback-cache` and `-attr-times-granularity`.
-
-#### 2. Extended eager-lock to metadata transactions in replicate xlator
-
-Eager lock feature in replicate xlator is extended to support metadata
-transactions in addition to data transactions. This helps in improving the
-performance when there are frequent metadata updates in the workload. This is
-typically seen with sharded volumes by default, and in other workloads that
-incur a higher rate of metadata modifications to the same set of files.
-
-As a part of this feature, compounded FOPs feature in AFR is deprecated, volumes
-that are configured to leverage compounding will start disregarding the option
-`use-compound-fops`.
-
-**NOTE:** This is an internal change in AFR xlator and is not user controlled
-or configurable.
-
-#### 3. Support for multi-threaded fuse readers
-
-FUSE based mounts can specify number of FUSE request processing threads during
-a mount. For workloads that have high concurrency on a single client, this helps
-in processing FUSE requests in parallel, than the existing single reader model.
-
-This is provided as a mount time option named `reader-thread-count` and can be
-used as follows,
-```
-# mount -t glusterfs -o reader-thread-count=<n> <server>:<volname> <mntpoint>
-```
-
-#### 4. Configurable aggregate size for write-behind xlator
-
-Write-behind xlator provides the option `performance.aggregate-size` to enable
-configurable aggregate write sizes. This option enables write-behind xlator to
-aggregate writes till the specified value before the writes are sent to the
-bricks.
-
-Existing behaviour set this size to a maximum of 128KB per file. The
-configurable option provides the ability to tune this up or down based on the
-workload to improve performance of writes.
-
-Usage:
-```
-# gluster volume set <volname> performance.aggregate-size <size>
-```
-
-#### 5. Adaptive read replica selection based on queue length
-
-AFR xlator is enhanced with a newer value for the option `read-hash-mode`.
-Providing this option with a value of `3` will distribute reads across AFR
-subvolumes based on the subvol having the least outstanding read requests.
-
-This helps in better distributing and hence improving workload performance on
-reads, in replicate based volumes.
-
-### Standalone
-
-#### 1. Thin arbiter quorum for 2-way replication
-
-**NOTE:** This feature is available only with GlusterD2
-
-Documentation for the feature is provided [here](https://docs.gluster.org/en/latest/Administrator%20Guide/Thin-Arbiter-Volumes/).
-
-#### 2. Automatically configure backup volfile servers in clients
-
-**NOTE:** This feature is available only with GlusterD2
-
-Clients connecting and mounting a Gluster volume, will automatically fetch and
-configure backup volfile servers, for future volfile updates and fetches, when
-the initial server used to fetch the volfile and mount is down.
-
-When using glusterd, this is achieved using the FUSE mount option
-`backup-volfile-servers`, and when using GlusterD2 this is done automatically.
-
-#### 3. (c/m)time equivalence across replicate and disperse subvolumes
-
-Enabling the utime feature, enables Gluster to maintain consistent change and
-modification time stamps on files and directories across bricks.
-
-This feature is useful when applications are sensitive to time deltas between
-operations (for example tar may report "file changed as we read it"), to
-maintain and report equal time stamps on the file across the subvolumes.
-
-To enable the feature use,
-```
-# gluster volume set <volname> features.utime
-```
-
-**Limitations**:
-- Mounting gluster volume with time attribute options (noatime, realatime...)
-is not supported with this feature
-- Certain entry operations (with differing creation flags) would reflect an
-eventual consistency w.r.t the time attributes
-- This feature does not guarantee consistent time for directories if hashed
-sub-volume for the directory is down
-- readdirp (or directory listing) is not supported with this feature
-
-### Developer related
-
-#### 1. New API for acquiring leases and acting on lease recalls
-
-A new API to acquire a lease on an open file and also to receive callbacks when
-the lease is recalled, is provided with gfapi.
-
-Refer to the [header](https://github.com/gluster/glusterfs/blob/release-4.1/api/src/glfs.h#L1112) for details on how to use this API.
-
-#### 2. Extended language bindings for gfapi to include perl
-
-See, [libgfapi-perl](https://github.com/gluster/libgfapi-perl) - Libgfapi bindings for Perl using FFI
-
-## Major issues
-
-**None**
-
-## Bugs addressed
-
-Bugs addressed since release-4.0.0 are listed below.
-
-- [#1074947](https://bugzilla.redhat.com/1074947): add option to build rpm without server
-- [#1234873](https://bugzilla.redhat.com/1234873): glusterfs-resource-agents - volume - voldir is not properly set
-- [#1272030](https://bugzilla.redhat.com/1272030): Remove lock recovery logic from client and server protocol translators
-- [#1304962](https://bugzilla.redhat.com/1304962): Intermittent file creation fail,while doing concurrent writes on distributed volume has more than 40 bricks
-- [#1312830](https://bugzilla.redhat.com/1312830): tests fail because bug-924726.t depends on netstat
-- [#1319992](https://bugzilla.redhat.com/1319992): RFE: Lease support for gluster
-- [#1450546](https://bugzilla.redhat.com/1450546): Paths to some tools are hardcoded to /sbin or /usr/sbin
-- [#1450593](https://bugzilla.redhat.com/1450593): Gluster Python scripts do not check return value of find_library
-- [#1468483](https://bugzilla.redhat.com/1468483): Sharding sends all application sent fsyncs to the main shard file
-- [#1495153](https://bugzilla.redhat.com/1495153): xlator_t structure's 'client_latency' variable is not used
-- [#1500649](https://bugzilla.redhat.com/1500649): Shellcheck errors in hook scripts
-- [#1505355](https://bugzilla.redhat.com/1505355): quota: directories doesn't get heal on newly added bricks when quota is full on sub-directory
-- [#1506140](https://bugzilla.redhat.com/1506140): Add quorum checks in post-op
-- [#1507230](https://bugzilla.redhat.com/1507230): Man pages badly formatted
-- [#1512691](https://bugzilla.redhat.com/1512691): PostgreSQL DB Restore: unexpected data beyond EOF
-- [#1517260](https://bugzilla.redhat.com/1517260): Volume wrong size
-- [#1521030](https://bugzilla.redhat.com/1521030): rpc: unregister programs before registering them again
-- [#1523122](https://bugzilla.redhat.com/1523122): fix serval bugs found on testing protocol/client
-- [#1523219](https://bugzilla.redhat.com/1523219): fuse xlator uses block size and fragment size 128KB leading to rounding off in df output
-- [#1530905](https://bugzilla.redhat.com/1530905): Reducing regression time of glusterd test cases
-- [#1533342](https://bugzilla.redhat.com/1533342): Syntactical errors in hook scripts for managing SELinux context on bricks
-- [#1536024](https://bugzilla.redhat.com/1536024): Rebalance process is behaving differently for AFR and EC volume.
-- [#1536186](https://bugzilla.redhat.com/1536186): build: glibc has removed legacy rpc headers and rpcgen in Fedora28, use libtirpc
-- [#1537362](https://bugzilla.redhat.com/1537362): glustershd/glusterd is not using right port when connecting to glusterfsd process
-- [#1537364](https://bugzilla.redhat.com/1537364): [RFE] - get-state option should mark profiling enabled flag at volume level
-- [#1537457](https://bugzilla.redhat.com/1537457): DHT log messages: Found anomalies in (null) (gfid = 00000000-0000-0000-0000-000000000000). Holes=1 overlaps=0
-- [#1537602](https://bugzilla.redhat.com/1537602): Georeplication tests intermittently fail
-- [#1538258](https://bugzilla.redhat.com/1538258): build: python-ctypes only in RHEL <= 7
-- [#1538427](https://bugzilla.redhat.com/1538427): Seeing timer errors in the rebalance logs
-- [#1539023](https://bugzilla.redhat.com/1539023): Add ability to control verbosity settings while compiling
-- [#1539166](https://bugzilla.redhat.com/1539166): [bitrot] scrub ondemand reports it's start as success without additional detail
-- [#1539358](https://bugzilla.redhat.com/1539358): Changes to self-heal logic w.r.t. detecting of split-brains
-- [#1539510](https://bugzilla.redhat.com/1539510): Optimize glusterd_import_friend_volume code path
-- [#1539545](https://bugzilla.redhat.com/1539545): gsyncd is running gluster command to get config file path is not required
-- [#1539603](https://bugzilla.redhat.com/1539603): Glusterfs crash when doing statedump with memory accounting is disabled
-- [#1540338](https://bugzilla.redhat.com/1540338): Change op-version of master to 4.1.0 for future options that maybe added
-- [#1540607](https://bugzilla.redhat.com/1540607): glusterd fails to attach brick during restart of the node
-- [#1540669](https://bugzilla.redhat.com/1540669): Do lock conflict check correctly for wait-list
-- [#1541038](https://bugzilla.redhat.com/1541038): A down brick is incorrectly considered to be online and makes the volume to be started without any brick available
-- [#1541264](https://bugzilla.redhat.com/1541264): dht_layout_t leak in dht_populate_inode_for_dentry
-- [#1541916](https://bugzilla.redhat.com/1541916): The used space in the volume increases when the volume is expanded
-- [#1542318](https://bugzilla.redhat.com/1542318): dht_lookup_unlink_of_false_linkto_cbk fails with "Permission denied"
-- [#1542829](https://bugzilla.redhat.com/1542829): Too many log messages about dictionary and options
-- [#1543279](https://bugzilla.redhat.com/1543279): Moving multiple temporary files to the same destination concurrently causes ESTALE error
-- [#1544090](https://bugzilla.redhat.com/1544090): possible memleak in glusterfsd process with brick multiplexing on
-- [#1544600](https://bugzilla.redhat.com/1544600): 3.8 -> 3.10 rolling upgrade fails (same for 3.12 or 3.13) on Ubuntu 14
-- [#1544699](https://bugzilla.redhat.com/1544699): Rolling upgrade to 4.0 is broken
-- [#1544961](https://bugzilla.redhat.com/1544961): libgfrpc does not export IPv6 RPC methods even with --with-ipv6-default
-- [#1545048](https://bugzilla.redhat.com/1545048): [brick-mux] process termination race while killing glusterfsd on last brick detach
-- [#1545056](https://bugzilla.redhat.com/1545056): [CIOT] : Gluster CLI says "io-threads : enabled" on existing volumes post upgrade.
-- [#1545891](https://bugzilla.redhat.com/1545891): Provide a automated way to update bugzilla status with patch merge.
-- [#1546129](https://bugzilla.redhat.com/1546129): Geo-rep: glibc fix breaks geo-replication
-- [#1546620](https://bugzilla.redhat.com/1546620): DHT calls dht_lookup_everywhere for 1xn volumes
-- [#1546954](https://bugzilla.redhat.com/1546954): [Rebalance] "Migrate file failed: <filepath>: failed to get xattr [No data available]" warnings in rebalance logs
-- [#1547068](https://bugzilla.redhat.com/1547068): Bricks getting assigned to different pids depending on whether brick path is IP or hostname based
-- [#1547128](https://bugzilla.redhat.com/1547128): Typo error in __dht_check_free_space function log message
-- [#1547662](https://bugzilla.redhat.com/1547662): After a replace brick command, self-heal takes some time to start healing files on disperse volumes
-- [#1547888](https://bugzilla.redhat.com/1547888): [brick-mux] incorrect event-thread scaling in server_reconfigure()
-- [#1548361](https://bugzilla.redhat.com/1548361): Make afr_fsync a transaction
-- [#1549000](https://bugzilla.redhat.com/1549000): line-coverage tests not capturing details properly.
-- [#1549606](https://bugzilla.redhat.com/1549606): Eager lock should be present for both metadata and data transactions
-- [#1549915](https://bugzilla.redhat.com/1549915): [Fuse Sub-dir] After performing add-brick on volume,doing rm -rf * on subdir mount point fails with "Transport endpoint is not connected"
-- [#1550078](https://bugzilla.redhat.com/1550078): memory leak in pre-op in replicate volumes for every write
-- [#1550339](https://bugzilla.redhat.com/1550339): glusterd leaks memory when vol status is issued
-- [#1550895](https://bugzilla.redhat.com/1550895): GD2 fails to dlopen server xlator
-- [#1550936](https://bugzilla.redhat.com/1550936): Pause/Resume of geo-replication with wrong user specified returns success
-- [#1553129](https://bugzilla.redhat.com/1553129): Memory corruption is causing crashes, hangs and invalid answers
-- [#1553598](https://bugzilla.redhat.com/1553598): [Rebalance] ENOSPC errors on few files in rebalance logs
-- [#1553926](https://bugzilla.redhat.com/1553926): configure --without-ipv6-default has odd behaviour
-- [#1553938](https://bugzilla.redhat.com/1553938): configure summary TIRPC result is misleading
-- [#1554053](https://bugzilla.redhat.com/1554053): 4.0 clients may fail to convert iatt in dict when recieving the same from older (< 4.0) servers
-- [#1554743](https://bugzilla.redhat.com/1554743): [EC] Read performance of EC volume exported over gNFS is significantly lower than write performance
-- [#1555154](https://bugzilla.redhat.com/1555154): glusterd: TLS verification fails when using intermediate CA instead of self-signed certificates
-- [#1555167](https://bugzilla.redhat.com/1555167): namespace test failure
-- [#1557435](https://bugzilla.redhat.com/1557435): Enable lookup-optimize by default
-- [#1557876](https://bugzilla.redhat.com/1557876): Fuse mount crashed with only one VM running with its image on that volume
-- [#1557932](https://bugzilla.redhat.com/1557932): Shard replicate volumes don't use eager-lock affectively
-- [#1558016](https://bugzilla.redhat.com/1558016): test ./tests/bugs/ec/bug-1236065.t is generating crash on build
-- [#1558074](https://bugzilla.redhat.com/1558074): [disperse] Add tests for in-memory stripe cache for the non aligned write
-- [#1558380](https://bugzilla.redhat.com/1558380): Modify glfsheal binary to accept socket file path as an optional argument.
-- [#1559004](https://bugzilla.redhat.com/1559004): /var/log/glusterfs/bricks/export_vdb.log flooded with this error message "Not able to add to index [Too many links]"
-- [#1559075](https://bugzilla.redhat.com/1559075): enable ownthread feature for glusterfs4_0_fop_prog
-- [#1559126](https://bugzilla.redhat.com/1559126): Incorrect error message in /features/changelog/lib/src/gf-history-changelog.c
-- [#1559130](https://bugzilla.redhat.com/1559130): ssh stderr in glusterfind gets swallowed
-- [#1559235](https://bugzilla.redhat.com/1559235): Increase the inode table size on server when upcall enabled
-- [#1560319](https://bugzilla.redhat.com/1560319): NFS client gets "Invalid argument" when writing file through nfs-ganesha with quota
-- [#1560393](https://bugzilla.redhat.com/1560393): Fix regresssion failure for ./tests/basic/md-cache/bug-1418249.t
-- [#1560411](https://bugzilla.redhat.com/1560411): fallocate created data set is crossing storage reserve space limits resulting 100% brick full
-- [#1560441](https://bugzilla.redhat.com/1560441): volume stop in mgmt v3
-- [#1560589](https://bugzilla.redhat.com/1560589): nl-cache.t fails
-- [#1560957](https://bugzilla.redhat.com/1560957): After performing remove-brick followed by add-brick operation, brick went offline state
-- [#1561129](https://bugzilla.redhat.com/1561129): When storage reserve limit is reached, appending data to an existing file throws EROFS error
-- [#1561406](https://bugzilla.redhat.com/1561406): Rebalance failures on a dispersed volume with lookup-optimize enabled
-- [#1562052](https://bugzilla.redhat.com/1562052): build: revert configure --without-ipv6-default behaviour
-- [#1562717](https://bugzilla.redhat.com/1562717): SHD is not healing entries in halo replication
-- [#1562907](https://bugzilla.redhat.com/1562907): set mgmt_v3_timer->timer to NULL after mgmt_v3_timer is deleted
-- [#1563273](https://bugzilla.redhat.com/1563273): mark brick as online only when portmap registration is completed
-- [#1563334](https://bugzilla.redhat.com/1563334): Honour cluster.localtime-logging option for all the daemons
-- [#1563511](https://bugzilla.redhat.com/1563511): Redundant synchronization in rename codepath for a single subvolume DHT
-- [#1563945](https://bugzilla.redhat.com/1563945): [EC] Turn ON the stripe-cache option by default for ec volume
-- [#1564198](https://bugzilla.redhat.com/1564198): [Remove-brick] Many files were not migrated from the decommissioned bricks; commit results in data loss
-- [#1564235](https://bugzilla.redhat.com/1564235): gfapi: fix a couple of minor issues
-- [#1564600](https://bugzilla.redhat.com/1564600): Client can create denial of service (DOS) conditions on server
-- [#1566067](https://bugzilla.redhat.com/1566067): Volume status inode is broken with brickmux
-- [#1566207](https://bugzilla.redhat.com/1566207): Linux kernel untar failed with "xz: (stdin): Read error: Invalid argument" immediate after add-brick
-- [#1566303](https://bugzilla.redhat.com/1566303): Removing directories from multiple clients throws ESTALE errors
-- [#1566386](https://bugzilla.redhat.com/1566386): Disable choose-local in groups virt and gluster-block
-- [#1566732](https://bugzilla.redhat.com/1566732): EIO errors on some operations when volume has mixed brick versions on a disperse volume
-- [#1567209](https://bugzilla.redhat.com/1567209): Geo-rep: faulty session due to OSError: [Errno 95] Operation not supported
-- [#1567880](https://bugzilla.redhat.com/1567880): Grant Deepshikha access to all CI-related infrastructure
-- [#1567881](https://bugzilla.redhat.com/1567881): Halo replication I/O path is not working
-- [#1568348](https://bugzilla.redhat.com/1568348): Rebalance on few nodes doesn't seem to complete - stuck at FUTEX_WAIT
-- [#1568521](https://bugzilla.redhat.com/1568521): shard files present even after deleting vm from ovirt UI
-- [#1568820](https://bugzilla.redhat.com/1568820): Add generated HMAC token in header for webhook calls
-- [#1568844](https://bugzilla.redhat.com/1568844): [snapshot-scheduler]Prevent access of shared storage volume from the outside client
-- [#1569198](https://bugzilla.redhat.com/1569198): bitrot scrub status does not show the brick where the object (file) is corrupted
-- [#1569489](https://bugzilla.redhat.com/1569489): Need heal-timeout to be configured as low as 5 seconds
-- [#1570011](https://bugzilla.redhat.com/1570011): test case is failing ./tests/bugs/glusterd/add-brick-and-validate-replicated-volume-options.t while brick mux is enabled
-- [#1570538](https://bugzilla.redhat.com/1570538): linux untar errors out at completion during disperse volume inservice upgrade
-- [#1570962](https://bugzilla.redhat.com/1570962): print the path of the corrupted object in scrub status
-- [#1571069](https://bugzilla.redhat.com/1571069): [geo-rep]: Lot of changelogs retries and "dict is null" errors in geo-rep logs
-- [#1572076](https://bugzilla.redhat.com/1572076): Dictionary response is not captured in syncop_(f)xattrop
-- [#1572581](https://bugzilla.redhat.com/1572581): Remove-brick failed on Distributed volume while rm -rf is in-progress
-- [#1572586](https://bugzilla.redhat.com/1572586): dht: do not allow migration if file is open
-- [#1573066](https://bugzilla.redhat.com/1573066): growing glusterd memory usage with connected RHGSWA
-- [#1573119](https://bugzilla.redhat.com/1573119): Amends in volume profile option 'gluster-block'
-- [#1573220](https://bugzilla.redhat.com/1573220): Memory leak in volume tier status command
-- [#1574259](https://bugzilla.redhat.com/1574259): Errors unintentionally reported for snapshot status
-- [#1574305](https://bugzilla.redhat.com/1574305): rm command hangs in fuse_request_send
-- [#1574606](https://bugzilla.redhat.com/1574606): the regression test "tests/bugs/posix/bug-990028.t" fails
-- [#1575294](https://bugzilla.redhat.com/1575294): lease recall callback should be avoided on closed
-- [#1575386](https://bugzilla.redhat.com/1575386): GlusterFS 4.1.0 tracker
-- [#1575707](https://bugzilla.redhat.com/1575707): Gluster volume smb share options are getting overwritten after restating the gluster volume
-- [#1576814](https://bugzilla.redhat.com/1576814): GlusterFS can be improved
-- [#1577162](https://bugzilla.redhat.com/1577162): gfapi: broken symbol versions
-- [#1579674](https://bugzilla.redhat.com/1579674): Remove EIO from the dht_inode_missing macro
-- [#1579736](https://bugzilla.redhat.com/1579736): Additional log messages in dht_readdir(p)_cbk
-- [#1579757](https://bugzilla.redhat.com/1579757): DHT Log flooding in mount log "key=trusted.glusterfs.dht.mds [Invalid argument]"
-- [#1580215](https://bugzilla.redhat.com/1580215): [geo-rep]: Lot of changelogs retries and "dict is null" errors in geo-rep logs
-- [#1580540](https://bugzilla.redhat.com/1580540): make getfattr return proper response for "glusterfs.gfidtopath" xattr for files created when gfid2path was off
-- [#1581548](https://bugzilla.redhat.com/1581548): writes succeed when only good brick is down in 1x3 volume
-- [#1581745](https://bugzilla.redhat.com/1581745): bug-1309462.t is failing reliably due to changes in security.capability changes in the kernel
-- [#1582056](https://bugzilla.redhat.com/1582056): Input/Output errors on a disperse volume with concurrent reads and writes
-- [#1582063](https://bugzilla.redhat.com/1582063): rpc: The gluster auth version is always AUTH_GLUSTERFS_v2
-- [#1582068](https://bugzilla.redhat.com/1582068): ctime: Rename and unlink does not update ctime
-- [#1582072](https://bugzilla.redhat.com/1582072): posix/ctime: Access time is not updated for file with a hardlink
-- [#1582080](https://bugzilla.redhat.com/1582080): posix/ctime: The first lookup on file is not healing the gfid
-- [#1582199](https://bugzilla.redhat.com/1582199): posix unwinds readdirp calls with readdir signature
-- [#1582286](https://bugzilla.redhat.com/1582286): Brick-mux regressions failing on 4.1 branch
-- [#1582531](https://bugzilla.redhat.com/1582531): posix/ctime: Mtime is not updated on setting it to older date
-- [#1582549](https://bugzilla.redhat.com/1582549): api: missing __THROW on pub function decls
-- [#1583016](https://bugzilla.redhat.com/1583016): libgfapi: glfs init fails on afr volume with ctime feature enabled
-- [#1583734](https://bugzilla.redhat.com/1583734): rpc_transport_unref() called for an unregistered socket fd
-- [#1583769](https://bugzilla.redhat.com/1583769): Fix incorrect rebalance log message
-- [#1584633](https://bugzilla.redhat.com/1584633): Brick process crashed after upgrade from RHGS-3.3.1 async(7.4) to RHGS-3.4(7.5)
-- [#1585894](https://bugzilla.redhat.com/1585894): posix/ctime: EC self heal of directory is blocked with ctime feature enabled
-- [#1587908](https://bugzilla.redhat.com/1587908): Fix deadlock in failure codepath of shard fsync
-- [#1590128](https://bugzilla.redhat.com/1590128): xdata is leaking in server3_3_seek