From 39b258131e6067f36445bd53f6c644709f57b5c2 Mon Sep 17 00:00:00 2001 From: Manisha Saini Date: Wed, 28 Oct 2020 01:30:33 +0530 Subject: [Test] Test Memory leak for arbiter volume on deleting file Change-Id: I1a9cefd16ddb376d8c496089d114c92efa1fd1ea Signed-off-by: Manisha Saini --- .../test_memory_leaks_with_files_delete.py | 113 +++++++++++++++++++++ 1 file changed, 113 insertions(+) create mode 100644 tests/functional/resource_leak/test_memory_leaks_with_files_delete.py diff --git a/tests/functional/resource_leak/test_memory_leaks_with_files_delete.py b/tests/functional/resource_leak/test_memory_leaks_with_files_delete.py new file mode 100644 index 000000000..ab29fdbe7 --- /dev/null +++ b/tests/functional/resource_leak/test_memory_leaks_with_files_delete.py @@ -0,0 +1,113 @@ +# Copyright (C) 2020 Red Hat, Inc. +# +# This program is free software; you can redistribute it and/or modify +# it under the terms of the GNU General Public License as published by +# the Free Software Foundation; either version 2 of the License, or +# any later version. +# +# This program is distributed in the hope that it will be useful, +# but WITHOUT ANY WARRANTY; without even the implied warranty of +# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the +# GNU General Public License for more details. +# +# You should have received a copy of the GNU General Public License along` +# with this program; if not, write to the Free Software Foundation, Inc., +# 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA. + +from glusto.core import Glusto as g +from glustolibs.gluster.gluster_base_class import GlusterBaseClass, runs_on +from glustolibs.gluster.exceptions import ExecutionError +from glustolibs.gluster.glusterdir import get_dir_contents +from glustolibs.io.memory_and_cpu_utils import ( + wait_for_logging_processes_to_stop) +from glustolibs.gluster.brick_libs import get_all_bricks + + +@runs_on([['arbiter', 'distributed-arbiter'], ['glusterfs']]) +class TestMemoryLeakWithRm(GlusterBaseClass): + + def setUp(self): + + self.get_super_method(self, 'setUp')() + + # Set test_id for get gathering + self.test_id = self.id() + + # Set I/O flag to false + self.is_io_running = False + + # Creating Volume and mounting the volume + ret = self.setup_volume_and_mount_volume(self.mounts) + if not ret: + raise ExecutionError("Volume creation or mount failed: %s" + % self.volname) + + def tearDown(self): + + # Unmounting and cleaning volume + ret = self.unmount_volume_and_cleanup_volume(self.mounts) + if not ret: + raise ExecutionError("Unable to delete volume %s" % self.volname) + + self.get_super_method(self, 'tearDown')() + + def test_memory_leak_with_rm(self): + """ + Test case: + 1. Create a volume, start it and mount it. + 2. Create 10,000 files each of size 200K + 3. Delete the files created at step 2 + 4. Check if the files are deleted from backend + 5. Check if there are any memory leaks and OOM killers. + """ + # Start monitoring resource usage on servers and clients + monitor_proc_dict = self.start_memory_and_cpu_usage_logging( + self.test_id, count=30) + self.assertIsNotNone(monitor_proc_dict, + "Failed to start monitoring on servers and " + "clients") + # Create files on mount point + cmd = ('cd %s;for i in {1..10000};' + 'do dd if=/dev/urandom bs=200K count=1 of=file$i;done;' + 'rm -rf %s/file*' + % (self.mounts[0].mountpoint, self.mounts[0].mountpoint)) + ret, _, _ = g.run(self.mounts[0].client_system, cmd) + self.assertEqual(ret, 0, "Failed to create and delete files on" + " mountpoint") + g.log.info("Successfully created and removed files on mountpoint") + + # Delete files from mount point and check if all files + # are deleted or not from mount point as well as backend bricks. + ret, _, _ = g.run(self.clients[0], + "rm -rf {}/*".format(self.mounts[0].mountpoint)) + self.assertFalse(ret, "rm -rf * failed on mount point") + + ret = get_dir_contents(self.clients[0], + "{}/".format(self.mounts[0].mountpoint)) + self.assertEqual(ret, [], "Unexpected: Files and directories still " + "seen from mount point") + + for brick in get_all_bricks(self.mnode, self.volname): + node, brick_path = brick.split(":") + ret = get_dir_contents(node, "{}/".format(brick_path)) + self.assertEqual(ret, [], "Unexpected: Files and dirs still seen " + "on brick %s on node %s" % (brick_path, node)) + g.log.info("rm -rf * on mount point successful") + + # Wait for monitoring processes to complete + ret = wait_for_logging_processes_to_stop(monitor_proc_dict, + cluster=True) + self.assertTrue(ret, + "ERROR: Failed to stop monitoring processes") + + # Check if there are any memory leaks and OOM killers + ret = self.check_for_memory_leaks_and_oom_kills_on_servers( + self.test_id) + self.assertFalse(ret, + "Memory leak and OOM kills check failed on servers") + + ret = self.check_for_memory_leaks_and_oom_kills_on_clients( + self.test_id) + self.assertFalse(ret, + "Memory leak and OOM kills check failed on clients") + g.log.info("No memory leaks or OOM kills found on serves and clients") -- cgit