diff options
| author | Jeff Darcy <jdarcy@redhat.com> | 2017-01-31 14:49:45 -0500 |
|---|---|---|
| committer | Shyamsundar Ranganathan <srangana@redhat.com> | 2017-02-01 19:54:58 -0500 |
| commit | 83803b4b2d70e9e6e16bb050d7ac8e49ba420893 (patch) | |
| tree | 9a6c1f3f9a723bf578f78c624d3ce9f44baac6db /tests/basic/afr/client-side-heal.t | |
| parent | 80b04666ec7019e132f76f734a88559457702f1b (diff) | |
core: run many bricks within one glusterfsd process
This patch adds support for multiple brick translator stacks running in
a single brick server process. This reduces our per-brick memory usage
by approximately 3x, and our appetite for TCP ports even more. It also
creates potential to avoid process/thread thrashing, and to improve QoS
by scheduling more carefully across the bricks, but realizing that
potential will require further work.
Multiplexing is controlled by the "cluster.brick-multiplex" global
option. By default it's off, and bricks are started in separate
processes as before. If multiplexing is enabled, then *compatible*
bricks (mostly those with the same transport options) will be started in
the same process.
Backport of:
> Change-Id: I45059454e51d6f4cbb29a4953359c09a408695cb
> BUG: 1385758
> Reviewed-on: https://review.gluster.org/14763
Change-Id: I4bce9080f6c93d50171823298fdf920258317ee8
BUG: 1418091
Signed-off-by: Jeff Darcy <jdarcy@redhat.com>
Reviewed-on: https://review.gluster.org/16496
Smoke: Gluster Build System <jenkins@build.gluster.org>
NetBSD-regression: NetBSD Build System <jenkins@build.gluster.org>
CentOS-regression: Gluster Build System <jenkins@build.gluster.org>
Reviewed-by: Shyamsundar Ranganathan <srangana@redhat.com>
Diffstat (limited to 'tests/basic/afr/client-side-heal.t')
| -rwxr-xr-x | tests/basic/afr/client-side-heal.t | 10 |
1 files changed, 5 insertions, 5 deletions
diff --git a/tests/basic/afr/client-side-heal.t b/tests/basic/afr/client-side-heal.t index d87f4b14063..eba7dc2b3c4 100755 --- a/tests/basic/afr/client-side-heal.t +++ b/tests/basic/afr/client-side-heal.t @@ -13,7 +13,7 @@ TEST $CLI volume set $V0 cluster.data-self-heal off TEST $CLI volume set $V0 cluster.metadata-self-heal off TEST $CLI volume start $V0 -TEST glusterfs --volfile-id=$V0 --volfile-server=$H0 --entry-timeout=0 $M0; +TEST $GFS --volfile-id=$V0 --volfile-server=$H0 $M0; echo "some data" > $M0/datafile EXPECT 0 echo $? TEST touch $M0/mdatafile @@ -46,11 +46,11 @@ TEST ls $M0/mdatafile #To trigger inode refresh for sure, the volume is unmounted and mounted each time. #Check that data heal does not happen. EXPECT_WITHIN $UMOUNT_TIMEOUT "Y" force_umount $M0 -TEST glusterfs --volfile-id=$V0 --volfile-server=$H0 --entry-timeout=0 $M0; +TEST $GFS --volfile-id=$V0 --volfile-server=$H0 $M0; TEST cat $M0/datafile #Check that entry heal does not happen. EXPECT_WITHIN $UMOUNT_TIMEOUT "Y" force_umount $M0 -TEST glusterfs --volfile-id=$V0 --volfile-server=$H0 --entry-timeout=0 $M0; +TEST $GFS --volfile-id=$V0 --volfile-server=$H0 $M0; TEST ls $M0/dir #No heal must have happened @@ -68,12 +68,12 @@ EXPECT 7 get_pending_heal_count $V0 #Inode refresh must trigger data and entry heals. #To trigger inode refresh for sure, the volume is unmounted and mounted each time. EXPECT_WITHIN $UMOUNT_TIMEOUT "Y" force_umount $M0 -TEST glusterfs --volfile-id=$V0 --volfile-server=$H0 --entry-timeout=0 $M0; +TEST $GFS --volfile-id=$V0 --volfile-server=$H0 $M0; TEST cat $M0/datafile EXPECT_WITHIN $HEAL_TIMEOUT 6 get_pending_heal_count $V0 EXPECT_WITHIN $UMOUNT_TIMEOUT "Y" force_umount $M0 -TEST glusterfs --volfile-id=$V0 --volfile-server=$H0 --entry-timeout=0 $M0; +TEST $GFS --volfile-id=$V0 --volfile-server=$H0 $M0; TEST ls $M0/dir EXPECT 5 get_pending_heal_count $V0 |
