diff options
author | Aravinda VK <avishwan@redhat.com> | 2014-06-23 13:43:20 +0530 |
---|---|---|
committer | Venky Shankar <vshankar@redhat.com> | 2014-07-23 20:36:02 -0700 |
commit | f4110bed9230025ef7cac8e550057e316f9de1cb (patch) | |
tree | 75bd99f7a32bc6a7cf2311fbb17366dcdd54a8ca /geo-replication/syncdaemon/resource.py | |
parent | 8a24b1487d440c23b00b56aba5b9bfb77af6ef03 (diff) |
geo-rep: minimize xsync crawl usage and set upper limit to xsync crawl
For effective handling of deletes and renames use history crawl
as much as possible. History crawl will run in loop till it syncs
all data before live changelog time.
When it uses xsync crawl(fallback when changelog not available, or
very first crawl) it sets upper limit to crawl.
After completing History crawl, it checks actual end time returned
by history api to compare with register time, if actual end is less
than register time then run history crawl one more time.
If first turn history processing time is less than the CHANGELOG
ROLLOVER TIME then sleep for the difference, After sleep if it is
guaranteed that rollover will happen and switches to live changelog
consumption without switching to xsync. This sleep is only when
history processing completed < CHANGELOG_ROLLOVER_TIME and sleep
only after the first turn, So will not affect the performance.
BUG: 1112238
Change-Id: I644ef24b07e42e81cec96a025ebd21244a555ec0
Signed-off-by: Aravinda VK <avishwan@redhat.com>
Reviewed-on: http://review.gluster.org/8151
Tested-by: Gluster Build System <jenkins@build.gluster.com>
Reviewed-by: Venky Shankar <vshankar@redhat.com>
Tested-by: Venky Shankar <vshankar@redhat.com>
Diffstat (limited to 'geo-replication/syncdaemon/resource.py')
-rw-r--r-- | geo-replication/syncdaemon/resource.py | 55 |
1 files changed, 36 insertions, 19 deletions
diff --git a/geo-replication/syncdaemon/resource.py b/geo-replication/syncdaemon/resource.py index b537ff65003..c84265739c5 100644 --- a/geo-replication/syncdaemon/resource.py +++ b/geo-replication/syncdaemon/resource.py @@ -1267,6 +1267,7 @@ class GLUSTER(AbstractUrl, SlaveLocal, SlaveRemote): # register the crawlers and start crawling # g1 ==> Xsync, g2 ==> config.change_detector(changelog by default) # g3 ==> changelog History + changelog_register_failed = False (inf, ouf, ra, wa) = gconf.rpc_fd.split(',') os.close(int(ra)) os.close(int(wa)) @@ -1278,7 +1279,6 @@ class GLUSTER(AbstractUrl, SlaveLocal, SlaveRemote): "local %s, remote %s" % (CHANGELOG_AGENT_CLIENT_VERSION, rv)) - g1.register() try: workdir = g2.setup_working_dir() # register with the changelog library @@ -1288,38 +1288,55 @@ class GLUSTER(AbstractUrl, SlaveLocal, SlaveRemote): workdir, gconf.changelog_log_file, g2.CHANGELOG_LOG_LEVEL, g2.CHANGELOG_CONN_RETRIES) - g2.register(changelog_agent) - g3.register(changelog_agent) - except ChangelogException as e: - logging.debug("Changelog register failed: %s - %s" % - (e.errno, e.strerror)) - + register_time = int(time.time()) + g2.register(register_time, changelog_agent) + g3.register(register_time, changelog_agent) + except ChangelogException: + changelog_register_failed = True + register_time = int(time.time()) + logging.info("Changelog register failed, fallback to xsync") + + g1.register(register_time) + logging.info("Register time: %s" % register_time) # oneshot: Try to use changelog history api, if not # available switch to FS crawl # Note: if config.change_detector is xsync then # it will not use changelog history api try: - g3.crawlwrap(oneshot=True) + if not changelog_register_failed: + g3.crawlwrap(oneshot=True) + else: + g1.crawlwrap(oneshot=True) except (ChangelogException, NoPurgeTimeAvailable, PartialHistoryAvailable) as e: if isinstance(e, ChangelogException): - logging.debug('Changelog history crawl failed, failback ' - 'to xsync: %s - %s' % (e.errno, e.strerror)) - elif isinstance(e, NoPurgeTimeAvailable): - logging.debug('Using xsync crawl since no purge time ' - 'available') + logging.info('Changelog history crawl failed, fallback ' + 'to xsync: %s - %s' % (e.errno, e.strerror)) elif isinstance(e, PartialHistoryAvailable): - logging.debug('Using xsync crawl after consuming history ' - 'till %s' % str(e)) - g1.crawlwrap(oneshot=True) + logging.info('Partial history available, using xsync crawl' + ' after consuming history ' + 'till %s' % str(e)) + g1.crawlwrap(oneshot=True, no_stime_update=True) + + # Reset xsync upper limit. g2, g3 are changelog and history + # instances, but if change_detector is set to xsync then + # g1, g2, g3 will be xsync instances. + g1.xsync_upper_limit = None + if getattr(g2, "xsync_upper_limit", None) is not None: + g2.xsync_upper_limit = None + + if getattr(g3, "xsync_upper_limit", None) is not None: + g3.xsync_upper_limit = None # crawl loop: Try changelog crawl, if failed # switch to FS crawl try: - g2.crawlwrap() + if not changelog_register_failed: + g2.crawlwrap() + else: + g1.crawlwrap() except ChangelogException as e: - logging.debug('Changelog crawl failed, failback to xsync: ' - '%s - %s' % (e.errno, e.strerror)) + logging.info('Changelog crawl failed, fallback to xsync') g1.crawlwrap() else: sup(self, *args) |