From d0604ca7d194a8f158096850ac931d5f5a0b8529 Mon Sep 17 00:00:00 2001 From: vamahaja Date: Thu, 17 Sep 2020 09:32:15 +0530 Subject: [TestFix][Tier2] Update pytest marker for 'tier2' tests Change-Id: I5db480da820c82b7b15f90ed70a4f60643079cb0 Signed-off-by: vamahaja --- tests/functional/arbiter/test_arbiter.py | 14 +++++++------- .../gluster_stability/test_brickmux_stability.py | 2 +- .../gluster_stability/test_gluster_block_stability.py | 4 ++-- tests/functional/heketi/test_block_volumes_heketi.py | 4 ++-- tests/functional/heketi/test_heketi_cluster_operations.py | 12 ++++++------ tests/functional/heketi/test_heketi_create_volume.py | 8 ++++---- tests/functional/heketi/test_heketi_device_operations.py | 8 ++++---- tests/functional/heketi/test_heketi_metrics.py | 10 +++++----- tests/functional/heketi/test_heketi_node_operations.py | 6 +++--- tests/functional/heketi/test_heketi_volume_operations.py | 12 ++++++------ .../functional/heketi/test_server_state_examine_gluster.py | 4 ++-- tests/functional/heketi/test_volume_creation.py | 6 +++--- tests/functional/heketi/test_volume_multi_req.py | 2 +- .../provisioning/test_dynamic_provisioning_block.py | 8 ++++---- .../provisioning/test_dynamic_provisioning_file.py | 4 ++-- tests/functional/provisioning/test_pv_resize.py | 8 ++++---- tests/functional/provisioning/test_storage_class_cases.py | 12 ++++++------ 17 files changed, 62 insertions(+), 62 deletions(-) diff --git a/tests/functional/arbiter/test_arbiter.py b/tests/functional/arbiter/test_arbiter.py index 44901b45..884e7c4b 100755 --- a/tests/functional/arbiter/test_arbiter.py +++ b/tests/functional/arbiter/test_arbiter.py @@ -325,7 +325,7 @@ class TestArbiterVolumeCreateExpandDelete(baseclass.BaseClass): self.verify_amount_and_proportion_of_arbiter_and_data_bricks( vol_info, arbiter_bricks=2, data_bricks=4) - @pytest.mark.tier1 + @pytest.mark.tier2 # NOTE(vponomar): do not create big volumes setting value less than 64 # for 'avg_file_size'. It will cause creation of very huge amount of files # making one test run very loooooooong. @@ -594,7 +594,7 @@ class TestArbiterVolumeCreateExpandDelete(baseclass.BaseClass): self.verify_amount_and_proportion_of_arbiter_and_data_bricks( vol_info, arbiter_bricks=2, data_bricks=4) - @pytest.mark.tier1 + @pytest.mark.tier2 @ddt.data(True, False) def test_expand_arbiter_volume_setting_tags_on_nodes_or_devices( self, node_tags): @@ -682,7 +682,7 @@ class TestArbiterVolumeCreateExpandDelete(baseclass.BaseClass): for brick in bricks['data_list']: self.assertIn(brick['name'].split(':')[0], data_hosts) - @pytest.mark.tier1 + @pytest.mark.tier2 @ddt.data( (4, '250M', True), (8, '122M', True), @@ -845,7 +845,7 @@ class TestArbiterVolumeCreateExpandDelete(baseclass.BaseClass): gluster_ops.match_heketi_and_gluster_volumes_by_prefix( heketi_volume_names, "{}_".format(prefix)) - @pytest.mark.tier1 + @pytest.mark.tier2 @podcmd.GlustoPod() def test_arbiter_volume_node_tag_removal(self): """Test remove tags from nodes and check if arbiter volume is @@ -1219,7 +1219,7 @@ class TestArbiterVolumeCreateExpandDelete(baseclass.BaseClass): "is created {} and and number of nodes having the arbiter" " tag {}".format(arbiter_brick_ip, arbiter_tag_node_ip)) - @pytest.mark.tier1 + @pytest.mark.tier2 @ddt.data( ((None, None, None, True), ('required', 'disabled', 'disabled', True)), @@ -1241,7 +1241,7 @@ class TestArbiterVolumeCreateExpandDelete(baseclass.BaseClass): else: self._arbiter_volume_node_tag_operations(node_tags) - @pytest.mark.tier1 + @pytest.mark.tier2 @podcmd.GlustoPod() def test_create_arbiter_volume_with_avg_file_size_and_expand(self): """ @@ -1838,7 +1838,7 @@ class TestArbiterVolumeCreateExpandDelete(baseclass.BaseClass): "Failed to match vg {} from gluster side with vg {} " "from heketi side".format(device, h_devices_with_tag)) - @pytest.mark.tier1 + @pytest.mark.tier2 @ddt.data( (1, False), (2, True), diff --git a/tests/functional/gluster_stability/test_brickmux_stability.py b/tests/functional/gluster_stability/test_brickmux_stability.py index a2134fc2..21e3daf9 100644 --- a/tests/functional/gluster_stability/test_brickmux_stability.py +++ b/tests/functional/gluster_stability/test_brickmux_stability.py @@ -23,7 +23,7 @@ class TestBrickMux(BaseClass): super(TestBrickMux, self).setUp() self.node = self.ocp_master_node[0] - @pytest.mark.tier1 + @pytest.mark.tier2 def test_brick_multiplex_pids_with_diff_vol_option_values(self): """Test Brick Pid's should be same when values of vol options are diff """ diff --git a/tests/functional/gluster_stability/test_gluster_block_stability.py b/tests/functional/gluster_stability/test_gluster_block_stability.py index 896d2ce0..be2f2428 100644 --- a/tests/functional/gluster_stability/test_gluster_block_stability.py +++ b/tests/functional/gluster_stability/test_gluster_block_stability.py @@ -713,7 +713,7 @@ class TestGlusterBlockStability(GlusterBlockBaseClass): self._perform_initiator_node_reboot_and_block_validations( ini_node, is_ini_taget_same=True) - @pytest.mark.tier1 + @pytest.mark.tier2 def test_validate_gluster_ip_utilized_by_blockvolumes(self): """ Validate if all gluster nodes IP are utilized by blockvolume when using HA=2 @@ -1865,7 +1865,7 @@ class TestGlusterBlockStability(GlusterBlockBaseClass): self.node, delete_log_level.format(TCMU_CONF), gluster_node=g_node) @skip("Blocked by BZ-1755903") - @pytest.mark.tier1 + @pytest.mark.tier2 def test_tcmu_log_levels(self): """Check tcmu log levels and verify log levels""" g_node, get_system_time = self.gluster_servers[0], "date '+%F %T'" diff --git a/tests/functional/heketi/test_block_volumes_heketi.py b/tests/functional/heketi/test_block_volumes_heketi.py index a9092b06..cbea98ae 100644 --- a/tests/functional/heketi/test_block_volumes_heketi.py +++ b/tests/functional/heketi/test_block_volumes_heketi.py @@ -224,7 +224,7 @@ class TestBlockVolumeOps(GlusterBlockBaseClass): self.assertEqual(v, vol_info[bhv_name] ["options"][k]) - @pytest.mark.tier1 + @pytest.mark.tier2 @ddt.data(True, False) def test_create_blockvolume_with_different_auth_values(self, auth_value): """To validate block volume creation with different auth values""" @@ -272,7 +272,7 @@ class TestBlockVolumeOps(GlusterBlockBaseClass): ("Block volume Names are not same %s as %s", (block_vol_info["name"], vol_name))) - @pytest.mark.tier1 + @pytest.mark.tier2 @podcmd.GlustoPod() def test_create_max_num_blockhostingvolumes(self): num_of_bv = 10 diff --git a/tests/functional/heketi/test_heketi_cluster_operations.py b/tests/functional/heketi/test_heketi_cluster_operations.py index 68239965..c9289ff5 100644 --- a/tests/functional/heketi/test_heketi_cluster_operations.py +++ b/tests/functional/heketi/test_heketi_cluster_operations.py @@ -12,7 +12,7 @@ from openshiftstoragelibs.waiter import Waiter class TestClusterOperationsTestCases(baseclass.BaseClass): """Class for heketi cluster creation related test cases""" - @pytest.mark.tier1 + @pytest.mark.tier2 @ddt.data("", "block", "file") def test_heketi_cluster_create(self, disable_volume_type): """Test heketi cluster creation""" @@ -38,7 +38,7 @@ class TestClusterOperationsTestCases(baseclass.BaseClass): cluster_info["block"], err_msg % ("block", "False")) self.assertTrue(cluster_info["file"], err_msg % ("file", "False")) - @pytest.mark.tier1 + @pytest.mark.tier2 def test_heketi_cluster_list(self): """Test and validateheketi cluster list operation""" # Create heketi cluster @@ -57,7 +57,7 @@ class TestClusterOperationsTestCases(baseclass.BaseClass): % (cluster_info["id"], cluster_list["clusters"])) self.assertIn(cluster_info["id"], cluster_list["clusters"], err_msg) - @pytest.mark.tier1 + @pytest.mark.tier2 def test_heketi_cluster_info(self): """Test and validateheketi cluster info operation""" # Create heketi cluster @@ -83,7 +83,7 @@ class TestClusterOperationsTestCases(baseclass.BaseClass): for param, value in params: self.assertEqual(get_cluster_info[param], value) - @pytest.mark.tier1 + @pytest.mark.tier2 def test_heketi_cluster_delete(self): """Test and validateheketi cluster delete operation""" # Create heketi cluster @@ -103,7 +103,7 @@ class TestClusterOperationsTestCases(baseclass.BaseClass): % (cluster_info["id"], cluster_list["clusters"])) self.assertNotIn(cluster_info["id"], cluster_list["clusters"], err_msg) - @pytest.mark.tier1 + @pytest.mark.tier2 def test_create_heketi_cluster_and_add_node(self): """Test heketi node add to a newly created cluster""" storage_host_info = g.config.get("additional_gluster_servers") @@ -212,7 +212,7 @@ class TestClusterOperationsTestCases(baseclass.BaseClass): self.assertEqual( zone, storage_zone, err_msg % ("zone", zone, storage_zone)) - @pytest.mark.tier1 + @pytest.mark.tier2 def test_heketi_server_operations_cleanup_on_idle_setup(self): """Run heketi db clean up on an idle setup""" h_node, h_url = self.heketi_client_node, self.heketi_server_url diff --git a/tests/functional/heketi/test_heketi_create_volume.py b/tests/functional/heketi/test_heketi_create_volume.py index a15d4447..f7fb6126 100644 --- a/tests/functional/heketi/test_heketi_create_volume.py +++ b/tests/functional/heketi/test_heketi_create_volume.py @@ -110,7 +110,7 @@ class TestHeketiVolume(BaseClass): "of Heketi volumes before and after volume creation: %s\n%s" % ( existing_h_vol_list, h_vol_list)) - @pytest.mark.tier1 + @pytest.mark.tier2 @podcmd.GlustoPod() def test_create_vol_and_retrieve_vol_info(self): """Validate heketi and gluster volume info""" @@ -194,7 +194,7 @@ class TestHeketiVolume(BaseClass): "\n%s" % (volume_ids[2], existing_volumes)) g.log.info("Sucessfully verified the topology info") - @pytest.mark.tier1 + @pytest.mark.tier2 def test_to_check_deletion_of_cluster(self): """Validate deletion of cluster with volumes""" # List heketi volumes @@ -435,7 +435,7 @@ class TestHeketiVolume(BaseClass): heketi_blockvolume_delete(h_node, h_url, block_vol) heketi_volume_delete(h_node, h_url, volume, raise_on_error=False) - @pytest.mark.tier1 + @pytest.mark.tier2 @ddt.data("", "block") def test_verify_delete_heketi_volumes_pending_entries_in_db( self, vol_type): @@ -592,7 +592,7 @@ class TestHeketiVolume(BaseClass): act_brick_count, exp_brick_count, err_msg.format(act_brick_count, exp_brick_count, "brick")) - @pytest.mark.tier1 + @pytest.mark.tier2 @ddt.data('', 'block') def test_verify_create_heketi_volumes_pending_entries_in_db( self, vol_type): diff --git a/tests/functional/heketi/test_heketi_device_operations.py b/tests/functional/heketi/test_heketi_device_operations.py index 71430a14..c8a9b991 100755 --- a/tests/functional/heketi/test_heketi_device_operations.py +++ b/tests/functional/heketi/test_heketi_device_operations.py @@ -298,7 +298,7 @@ class TestHeketiDeviceOperations(BaseClass): "Some of the '%s' volume bricks is present of the removed " "'%s' device." % (vol_info['id'], lowest_device_id)) - @pytest.mark.tier1 + @pytest.mark.tier2 @ddt.data((True, False), (False, False), (True, True)) @ddt.unpack def test_heketi_device_remove_delete( @@ -428,7 +428,7 @@ class TestHeketiDeviceOperations(BaseClass): heketi_device_disable, heketi_node, heketi_url, device_id) raise - @pytest.mark.tier1 + @pytest.mark.tier2 def test_heketi_device_delete(self): """Test Heketi device delete operation""" @@ -488,7 +488,7 @@ class TestHeketiDeviceOperations(BaseClass): "after the device deletion" % (device_id, node_id)) self.assertNotIn(device_id, node_info_after_deletion, msg) - @pytest.mark.tier1 + @pytest.mark.tier2 def test_heketi_device_info(self): """Validate whether device related information is displayed""" @@ -554,7 +554,7 @@ class TestHeketiDeviceOperations(BaseClass): self.assertEqual(brick[key], brick_from_t[key]) self.assertEqual(brick_match_count, len(device_info['bricks'])) - @pytest.mark.tier1 + @pytest.mark.tier2 def test_device_delete_with_bricks(self): """Validate device deletion with existing bricks on the device""" h_node, h_url = self.heketi_client_node, self.heketi_server_url diff --git a/tests/functional/heketi/test_heketi_metrics.py b/tests/functional/heketi/test_heketi_metrics.py index 9f161607..ec0fc57b 100644 --- a/tests/functional/heketi/test_heketi_metrics.py +++ b/tests/functional/heketi/test_heketi_metrics.py @@ -172,7 +172,7 @@ class TestHeketiMetrics(BaseClass): vol_count['cluster'], json=True) self.assertEqual(vol_count['value'], len(cluster_info['volumes'])) - @pytest.mark.tier1 + @pytest.mark.tier2 def test_heketi_metrics_with_topology_info(self): """Validate heketi metrics generation""" self.verify_heketi_metrics_with_topology_info() @@ -222,7 +222,7 @@ class TestHeketiMetrics(BaseClass): self.verify_heketi_metrics_with_topology_info() - @pytest.mark.tier1 + @pytest.mark.tier2 def test_heketi_metrics_validating_vol_count_on_vol_creation(self): """Validate heketi metrics VolumeCount after volume creation""" @@ -247,7 +247,7 @@ class TestHeketiMetrics(BaseClass): self.verify_volume_count() - @pytest.mark.tier1 + @pytest.mark.tier2 def test_heketi_metrics_validating_vol_count_on_vol_deletion(self): """Validate heketi metrics VolumeCount after volume deletion""" @@ -287,7 +287,7 @@ class TestHeketiMetrics(BaseClass): self.assertNotIn(vol['id'], volume_list) self.verify_volume_count() - @pytest.mark.tier1 + @pytest.mark.tier2 def test_heketi_metrics_validating_cluster_count(self): """Validate 'cluster count' in heketi metrics""" cluster_list = heketi_cluster_list( @@ -305,7 +305,7 @@ class TestHeketiMetrics(BaseClass): self.assertEqual( len(cluster_list['clusters']), metrics['heketi_cluster_count']) - @pytest.mark.tier1 + @pytest.mark.tier2 def test_heketi_metrics_validating_existing_node_count(self): """Validate existing 'node count' in heketi metrics""" metrics = get_heketi_metrics( diff --git a/tests/functional/heketi/test_heketi_node_operations.py b/tests/functional/heketi/test_heketi_node_operations.py index adcbf6ef..65e3554d 100644 --- a/tests/functional/heketi/test_heketi_node_operations.py +++ b/tests/functional/heketi/test_heketi_node_operations.py @@ -54,7 +54,7 @@ class TestHeketiNodeOperations(baseclass.BaseClass): "Heketi volume list %s is not equal to gluster volume list %s" % (node_ips, hostnames)) - @pytest.mark.tier1 + @pytest.mark.tier2 def test_heketi_node_info(self): """Test heketi node info operation """ @@ -254,7 +254,7 @@ class TestHeketiNodeOperations(baseclass.BaseClass): # Add node to valid cluster id self.heketi_node_add_with_valid_cluster() - @pytest.mark.tier1 + @pytest.mark.tier2 def test_validate_heketi_node_add_with_db_check(self): """Test heketi db check after node add operation""" if not self.is_containerized_gluster(): @@ -285,7 +285,7 @@ class TestHeketiNodeOperations(baseclass.BaseClass): " not as expected".format(initial_node_count, final_node_count)) self.assertEqual(initial_node_count + 1, final_node_count, msg) - @pytest.mark.tier1 + @pytest.mark.tier2 def test_heketi_node_add_with_invalid_cluster(self): """Test heketi node add operation with invalid cluster id""" storage_hostname, cluster_id = None, utils.get_random_str(size=33) diff --git a/tests/functional/heketi/test_heketi_volume_operations.py b/tests/functional/heketi/test_heketi_volume_operations.py index ef9ae213..b3192d02 100644 --- a/tests/functional/heketi/test_heketi_volume_operations.py +++ b/tests/functional/heketi/test_heketi_volume_operations.py @@ -32,7 +32,7 @@ class TestHeketiVolumeOperations(BaseClass): super(TestHeketiVolumeOperations, cls).setUpClass() cls.volume_size = 1 - @pytest.mark.tier1 + @pytest.mark.tier2 def test_heketi_with_default_options(self): """ Test to create volume with default options. @@ -52,7 +52,7 @@ class TestHeketiVolumeOperations(BaseClass): "Expected Size: %s, Actual Size: %s" % (self.volume_size, vol_info['size']))) - @pytest.mark.tier1 + @pytest.mark.tier2 def test_heketi_with_expand_volume(self): """ Test volume expand and size if updated correctly in heketi-cli info @@ -86,7 +86,7 @@ class TestHeketiVolumeOperations(BaseClass): "Size: %s" % (str(expected_size), str(volume_info['size'])))) - @pytest.mark.tier1 + @pytest.mark.tier2 def test_heketi_volume_mount(self): self.node = self.ocp_master_node[0] try: @@ -179,7 +179,7 @@ class TestHeketiVolumeOperations(BaseClass): pids.append([g_node, pid]) return pids - @pytest.mark.tier1 + @pytest.mark.tier2 @podcmd.GlustoPod() def test_heketi_volume_snapshot_create_with_one_brick_down(self): """ @@ -242,7 +242,7 @@ class TestHeketiVolumeOperations(BaseClass): "Expecting Snapshot count before {} and after creation {} to be " "same".format(snap_list_before, snap_list_after)) - @pytest.mark.tier1 + @pytest.mark.tier2 def test_heketi_volume_create_mutiple_sizes(self): """Validate creation of heketi volume with differnt sizes""" sizes, required_space = [15, 50, 100], 495 @@ -259,7 +259,7 @@ class TestHeketiVolumeOperations(BaseClass): vol_id = heketi_volume_create(h_node, h_url, size, json=True)['id'] self.addCleanup(heketi_volume_delete, h_node, h_url, vol_id) - @pytest.mark.tier1 + @pytest.mark.tier2 @podcmd.GlustoPod() def test_heketi_volume_snapshot_delete(self): """Test heketi volume snapshot delete operation""" diff --git a/tests/functional/heketi/test_server_state_examine_gluster.py b/tests/functional/heketi/test_server_state_examine_gluster.py index 6b66f13a..0a93c9eb 100644 --- a/tests/functional/heketi/test_server_state_examine_gluster.py +++ b/tests/functional/heketi/test_server_state_examine_gluster.py @@ -27,7 +27,7 @@ class TestHeketiServerStateExamineGluster(BaseClass): self.skipTest("heketi-client package %s does not support server " "state examine gluster" % version.v_str) - @pytest.mark.tier1 + @pytest.mark.tier2 def test_volume_inconsistencies(self): # Examine Gluster cluster and Heketi that there is no inconsistencies out = heketi_ops.heketi_examine_gluster( @@ -130,7 +130,7 @@ class TestHeketiServerStateExamineGluster(BaseClass): db_bricks_count, topology_bricks_count)) self.assertEqual(topology_bricks_count, db_bricks_count, msg) - @pytest.mark.tier1 + @pytest.mark.tier2 @ddt.data('', 'block') def test_compare_heketi_volumes(self, vol_type): """Validate file/block volume count using heketi gluster examine""" diff --git a/tests/functional/heketi/test_volume_creation.py b/tests/functional/heketi/test_volume_creation.py index 54c808dd..d96fa9e0 100644 --- a/tests/functional/heketi/test_volume_creation.py +++ b/tests/functional/heketi/test_volume_creation.py @@ -158,7 +158,7 @@ class TestVolumeCreationTestCases(BaseClass): vol_fail, "Volume should have not been created. Out: %s" % vol_fail) - @pytest.mark.tier1 + @pytest.mark.tier2 @podcmd.GlustoPod() def test_volume_create_replica_2(self): """Validate creation of a replica 2 volume""" @@ -189,7 +189,7 @@ class TestVolumeCreationTestCases(BaseClass): "Brick amount is expected to be 2. " "Actual amount is '%s'" % brick_amount) - @pytest.mark.tier1 + @pytest.mark.tier2 @podcmd.GlustoPod() def test_volume_create_snapshot_enabled(self): """Validate volume creation with snapshot enabled""" @@ -386,7 +386,7 @@ class TestVolumeCreationTestCases(BaseClass): % (vol_name, gluster_v_info['brickCount'])) self.assertFalse(int(gluster_v_info['brickCount']) % 3, msg) - @pytest.mark.tier1 + @pytest.mark.tier2 def test_create_volume_with_same_name(self): """Test create two volumes with the same name and verify that 2nd one is failing with the appropriate error. diff --git a/tests/functional/heketi/test_volume_multi_req.py b/tests/functional/heketi/test_volume_multi_req.py index 0780be1e..44ea803b 100644 --- a/tests/functional/heketi/test_volume_multi_req.py +++ b/tests/functional/heketi/test_volume_multi_req.py @@ -336,7 +336,7 @@ class TestVolumeMultiReq(BaseClass): # verify this volume in heketi self.assertIn(c2.heketiVolumeName, now_vols) - @pytest.mark.tier1 + @pytest.mark.tier2 # NOTE(jjm): I've noticed that on the system I'm using (RHEL7). # with count=8 things start to back up a bit. # I needed to increase some timeouts to get this to pass. diff --git a/tests/functional/provisioning/test_dynamic_provisioning_block.py b/tests/functional/provisioning/test_dynamic_provisioning_block.py index cce553b0..215354da 100755 --- a/tests/functional/provisioning/test_dynamic_provisioning_block.py +++ b/tests/functional/provisioning/test_dynamic_provisioning_block.py @@ -118,7 +118,7 @@ class TestDynamicProvisioningBlockP0(GlusterBlockBaseClass): """ self.dynamic_provisioning_glusterblock(set_hacount=True) - @pytest.mark.tier1 + @pytest.mark.tier2 def test_dynamic_provisioning_glusterblock_hacount_false(self): """Validate storage-class mandatory parameters for block """ @@ -255,7 +255,7 @@ class TestDynamicProvisioningBlockP0(GlusterBlockBaseClass): ret, out, err = async_io.async_communicate() self.assertEqual(ret, 0, "IO %s failed on %s" % (io_cmd, self.node)) - @pytest.mark.tier1 + @pytest.mark.tier2 def test_glusterblock_logs_presence_verification(self): """Validate presence of glusterblock provisioner POD and it's status""" @@ -549,7 +549,7 @@ class TestDynamicProvisioningBlockP0(GlusterBlockBaseClass): # create a block pvc greater than default BHV size self.create_and_wait_for_pvc(pvc_size=(default_bhv_size + 1)) - @pytest.mark.tier1 + @pytest.mark.tier2 def test_expansion_of_block_hosting_volume_using_heketi(self): """Verify that after expanding block hosting volume we are able to consume the expanded space""" @@ -839,7 +839,7 @@ class TestDynamicProvisioningBlockP0(GlusterBlockBaseClass): " No. of BHV after the test : {}".format(len(bhv_list), bhv_post)) self.assertEqual(bhv_post, (len(bhv_list) + 2), err_msg) - @pytest.mark.tier1 + @pytest.mark.tier2 def test_100gb_block_pvc_create_and_delete_twice(self): """Validate creation and deletion of blockvoume of size 100GB""" # Define required space, bhv size required for on 100GB block PVC diff --git a/tests/functional/provisioning/test_dynamic_provisioning_file.py b/tests/functional/provisioning/test_dynamic_provisioning_file.py index c711186c..c1f9d078 100644 --- a/tests/functional/provisioning/test_dynamic_provisioning_file.py +++ b/tests/functional/provisioning/test_dynamic_provisioning_file.py @@ -114,7 +114,7 @@ class TestDynamicProvisioningP0(BaseClass): g.log.info("test_dynamic_provisioning_glusterfile") self.dynamic_provisioning_glusterfile(False) - @pytest.mark.tier1 + @pytest.mark.tier2 def test_dynamic_provisioning_glusterfile_volname_prefix(self): """Validate dynamic provisioning for gluster file with vol name prefix """ @@ -481,7 +481,7 @@ class TestDynamicProvisioningP0(BaseClass): oc_delete(self.node, 'pv', pv_name) wait_for_resource_absence(self.node, 'pv', pv_name) - @pytest.mark.tier1 + @pytest.mark.tier2 def test_usage_of_default_storage_class(self): """Validate PVs creation for SC with default custom volname prefix""" diff --git a/tests/functional/provisioning/test_pv_resize.py b/tests/functional/provisioning/test_pv_resize.py index abadc4ee..a7baca00 100644 --- a/tests/functional/provisioning/test_pv_resize.py +++ b/tests/functional/provisioning/test_pv_resize.py @@ -50,7 +50,7 @@ class TestPvResizeClass(BaseClass): g.log.error(msg) raise self.skipTest(msg) - @pytest.mark.tier1 + @pytest.mark.tier2 @ddt.data( (True, True), (False, True), @@ -258,7 +258,7 @@ class TestPvResizeClass(BaseClass): """Validate PVC resize when resized by exact available free space""" self._pv_resize(exceed_free_space=False) - @pytest.mark.tier1 + @pytest.mark.tier2 def test_pv_resize_try_shrink_pv_size(self): """Validate whether reducing the PV size is allowed""" dir_path = "/mnt/" @@ -295,7 +295,7 @@ class TestPvResizeClass(BaseClass): self.assertEqual( ret, 0, "Failed to execute command %s on %s" % (cmd, node)) - @pytest.mark.tier1 + @pytest.mark.tier2 def test_pv_resize_when_heketi_down(self): """Create a PVC and try to expand it when heketi is down, It should fail. After heketi is up, expand PVC should work. @@ -418,7 +418,7 @@ class TestPvResizeClass(BaseClass): resize_pvc(self.ocp_master_node[0], pvc_name, available_size_gb) verify_pvc_size(self.ocp_master_node[0], pvc_name, available_size_gb) - @pytest.mark.tier1 + @pytest.mark.tier2 def test_pv_resize_device_disabled(self): """Validate resize after disabling all devices except one""" h_node, h_url = self.heketi_client_node, self.heketi_server_url diff --git a/tests/functional/provisioning/test_storage_class_cases.py b/tests/functional/provisioning/test_storage_class_cases.py index 05ee49cc..3a2b9dfa 100644 --- a/tests/functional/provisioning/test_storage_class_cases.py +++ b/tests/functional/provisioning/test_storage_class_cases.py @@ -181,7 +181,7 @@ class TestStorageClassCases(BaseClass): validate_multipath_pod( self.ocp_master_node[0], pod_name, hacount, list(mpaths)[0]) - @pytest.mark.tier1 + @pytest.mark.tier2 @ddt.data( {"volumetype": "dist-rep:3"}, {"resturl": "http://10.0.0.1:8080"}, @@ -195,7 +195,7 @@ class TestStorageClassCases(BaseClass): """Validate glusterfile storage with different incorrect parameters""" self.create_sc_with_parameter("glusterfile", parameter=parameter) - @pytest.mark.tier1 + @pytest.mark.tier2 @ddt.data( {"resturl": "http://10.0.0.1:8080"}, {"restsecretname": "fakerestsecretname", @@ -331,7 +331,7 @@ class TestStorageClassCases(BaseClass): "Invalid chapauthenabled value '%s'" % chapauthenabled ) - @pytest.mark.tier1 + @pytest.mark.tier2 def test_create_and_verify_pvc_with_volume_name_prefix(self): """create and verify pvc with volname prefix on an app pod""" if get_openshift_version() < "3.9": @@ -358,7 +358,7 @@ class TestStorageClassCases(BaseClass): "Failed to read Endpoints of %s on %s " % ( pv_name, self.ocp_master_node[0])) - @pytest.mark.tier1 + @pytest.mark.tier2 def test_try_to_create_sc_with_duplicated_name(self): """Verify SC creation fails with duplicate name""" sc_name = "test-sc-duplicated-name-" + utils.get_random_str() @@ -367,7 +367,7 @@ class TestStorageClassCases(BaseClass): with self.assertRaises(AssertionError): self.create_storage_class(sc_name=sc_name) - @pytest.mark.tier1 + @pytest.mark.tier2 @ddt.data('secretName', 'secretNamespace', None) def test_sc_glusterfile_missing_parameter(self, parameter): """Validate glusterfile storage with missing parameters""" @@ -397,7 +397,7 @@ class TestStorageClassCases(BaseClass): with self.assertRaises(ExecutionError): verify_pvc_status_is_bound(node, pvc_name, timeout=1) - @pytest.mark.tier1 + @pytest.mark.tier2 def test_sc_create_with_clusterid(self): """Create storage class with 'cluster id'""" h_cluster_list = heketi_cluster_list( -- cgit