summaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
-rwxr-xr-xtests/functional/arbiter/test_arbiter.py14
-rw-r--r--tests/functional/gluster_stability/test_brickmux_stability.py2
-rw-r--r--tests/functional/gluster_stability/test_gluster_block_stability.py4
-rw-r--r--tests/functional/heketi/test_block_volumes_heketi.py4
-rw-r--r--tests/functional/heketi/test_heketi_cluster_operations.py12
-rw-r--r--tests/functional/heketi/test_heketi_create_volume.py8
-rwxr-xr-xtests/functional/heketi/test_heketi_device_operations.py8
-rw-r--r--tests/functional/heketi/test_heketi_metrics.py10
-rw-r--r--tests/functional/heketi/test_heketi_node_operations.py6
-rw-r--r--tests/functional/heketi/test_heketi_volume_operations.py12
-rw-r--r--tests/functional/heketi/test_server_state_examine_gluster.py4
-rw-r--r--tests/functional/heketi/test_volume_creation.py6
-rw-r--r--tests/functional/heketi/test_volume_multi_req.py2
-rwxr-xr-xtests/functional/provisioning/test_dynamic_provisioning_block.py8
-rw-r--r--tests/functional/provisioning/test_dynamic_provisioning_file.py4
-rw-r--r--tests/functional/provisioning/test_pv_resize.py8
-rw-r--r--tests/functional/provisioning/test_storage_class_cases.py12
17 files changed, 62 insertions, 62 deletions
diff --git a/tests/functional/arbiter/test_arbiter.py b/tests/functional/arbiter/test_arbiter.py
index 44901b45..884e7c4b 100755
--- a/tests/functional/arbiter/test_arbiter.py
+++ b/tests/functional/arbiter/test_arbiter.py
@@ -325,7 +325,7 @@ class TestArbiterVolumeCreateExpandDelete(baseclass.BaseClass):
self.verify_amount_and_proportion_of_arbiter_and_data_bricks(
vol_info, arbiter_bricks=2, data_bricks=4)
- @pytest.mark.tier1
+ @pytest.mark.tier2
# NOTE(vponomar): do not create big volumes setting value less than 64
# for 'avg_file_size'. It will cause creation of very huge amount of files
# making one test run very loooooooong.
@@ -594,7 +594,7 @@ class TestArbiterVolumeCreateExpandDelete(baseclass.BaseClass):
self.verify_amount_and_proportion_of_arbiter_and_data_bricks(
vol_info, arbiter_bricks=2, data_bricks=4)
- @pytest.mark.tier1
+ @pytest.mark.tier2
@ddt.data(True, False)
def test_expand_arbiter_volume_setting_tags_on_nodes_or_devices(
self, node_tags):
@@ -682,7 +682,7 @@ class TestArbiterVolumeCreateExpandDelete(baseclass.BaseClass):
for brick in bricks['data_list']:
self.assertIn(brick['name'].split(':')[0], data_hosts)
- @pytest.mark.tier1
+ @pytest.mark.tier2
@ddt.data(
(4, '250M', True),
(8, '122M', True),
@@ -845,7 +845,7 @@ class TestArbiterVolumeCreateExpandDelete(baseclass.BaseClass):
gluster_ops.match_heketi_and_gluster_volumes_by_prefix(
heketi_volume_names, "{}_".format(prefix))
- @pytest.mark.tier1
+ @pytest.mark.tier2
@podcmd.GlustoPod()
def test_arbiter_volume_node_tag_removal(self):
"""Test remove tags from nodes and check if arbiter volume is
@@ -1219,7 +1219,7 @@ class TestArbiterVolumeCreateExpandDelete(baseclass.BaseClass):
"is created {} and and number of nodes having the arbiter"
" tag {}".format(arbiter_brick_ip, arbiter_tag_node_ip))
- @pytest.mark.tier1
+ @pytest.mark.tier2
@ddt.data(
((None, None, None, True),
('required', 'disabled', 'disabled', True)),
@@ -1241,7 +1241,7 @@ class TestArbiterVolumeCreateExpandDelete(baseclass.BaseClass):
else:
self._arbiter_volume_node_tag_operations(node_tags)
- @pytest.mark.tier1
+ @pytest.mark.tier2
@podcmd.GlustoPod()
def test_create_arbiter_volume_with_avg_file_size_and_expand(self):
"""
@@ -1838,7 +1838,7 @@ class TestArbiterVolumeCreateExpandDelete(baseclass.BaseClass):
"Failed to match vg {} from gluster side with vg {} "
"from heketi side".format(device, h_devices_with_tag))
- @pytest.mark.tier1
+ @pytest.mark.tier2
@ddt.data(
(1, False),
(2, True),
diff --git a/tests/functional/gluster_stability/test_brickmux_stability.py b/tests/functional/gluster_stability/test_brickmux_stability.py
index a2134fc2..21e3daf9 100644
--- a/tests/functional/gluster_stability/test_brickmux_stability.py
+++ b/tests/functional/gluster_stability/test_brickmux_stability.py
@@ -23,7 +23,7 @@ class TestBrickMux(BaseClass):
super(TestBrickMux, self).setUp()
self.node = self.ocp_master_node[0]
- @pytest.mark.tier1
+ @pytest.mark.tier2
def test_brick_multiplex_pids_with_diff_vol_option_values(self):
"""Test Brick Pid's should be same when values of vol options are diff
"""
diff --git a/tests/functional/gluster_stability/test_gluster_block_stability.py b/tests/functional/gluster_stability/test_gluster_block_stability.py
index 896d2ce0..be2f2428 100644
--- a/tests/functional/gluster_stability/test_gluster_block_stability.py
+++ b/tests/functional/gluster_stability/test_gluster_block_stability.py
@@ -713,7 +713,7 @@ class TestGlusterBlockStability(GlusterBlockBaseClass):
self._perform_initiator_node_reboot_and_block_validations(
ini_node, is_ini_taget_same=True)
- @pytest.mark.tier1
+ @pytest.mark.tier2
def test_validate_gluster_ip_utilized_by_blockvolumes(self):
""" Validate if all gluster nodes IP are
utilized by blockvolume when using HA=2
@@ -1865,7 +1865,7 @@ class TestGlusterBlockStability(GlusterBlockBaseClass):
self.node, delete_log_level.format(TCMU_CONF), gluster_node=g_node)
@skip("Blocked by BZ-1755903")
- @pytest.mark.tier1
+ @pytest.mark.tier2
def test_tcmu_log_levels(self):
"""Check tcmu log levels and verify log levels"""
g_node, get_system_time = self.gluster_servers[0], "date '+%F %T'"
diff --git a/tests/functional/heketi/test_block_volumes_heketi.py b/tests/functional/heketi/test_block_volumes_heketi.py
index a9092b06..cbea98ae 100644
--- a/tests/functional/heketi/test_block_volumes_heketi.py
+++ b/tests/functional/heketi/test_block_volumes_heketi.py
@@ -224,7 +224,7 @@ class TestBlockVolumeOps(GlusterBlockBaseClass):
self.assertEqual(v, vol_info[bhv_name]
["options"][k])
- @pytest.mark.tier1
+ @pytest.mark.tier2
@ddt.data(True, False)
def test_create_blockvolume_with_different_auth_values(self, auth_value):
"""To validate block volume creation with different auth values"""
@@ -272,7 +272,7 @@ class TestBlockVolumeOps(GlusterBlockBaseClass):
("Block volume Names are not same %s as %s",
(block_vol_info["name"], vol_name)))
- @pytest.mark.tier1
+ @pytest.mark.tier2
@podcmd.GlustoPod()
def test_create_max_num_blockhostingvolumes(self):
num_of_bv = 10
diff --git a/tests/functional/heketi/test_heketi_cluster_operations.py b/tests/functional/heketi/test_heketi_cluster_operations.py
index 68239965..c9289ff5 100644
--- a/tests/functional/heketi/test_heketi_cluster_operations.py
+++ b/tests/functional/heketi/test_heketi_cluster_operations.py
@@ -12,7 +12,7 @@ from openshiftstoragelibs.waiter import Waiter
class TestClusterOperationsTestCases(baseclass.BaseClass):
"""Class for heketi cluster creation related test cases"""
- @pytest.mark.tier1
+ @pytest.mark.tier2
@ddt.data("", "block", "file")
def test_heketi_cluster_create(self, disable_volume_type):
"""Test heketi cluster creation"""
@@ -38,7 +38,7 @@ class TestClusterOperationsTestCases(baseclass.BaseClass):
cluster_info["block"], err_msg % ("block", "False"))
self.assertTrue(cluster_info["file"], err_msg % ("file", "False"))
- @pytest.mark.tier1
+ @pytest.mark.tier2
def test_heketi_cluster_list(self):
"""Test and validateheketi cluster list operation"""
# Create heketi cluster
@@ -57,7 +57,7 @@ class TestClusterOperationsTestCases(baseclass.BaseClass):
% (cluster_info["id"], cluster_list["clusters"]))
self.assertIn(cluster_info["id"], cluster_list["clusters"], err_msg)
- @pytest.mark.tier1
+ @pytest.mark.tier2
def test_heketi_cluster_info(self):
"""Test and validateheketi cluster info operation"""
# Create heketi cluster
@@ -83,7 +83,7 @@ class TestClusterOperationsTestCases(baseclass.BaseClass):
for param, value in params:
self.assertEqual(get_cluster_info[param], value)
- @pytest.mark.tier1
+ @pytest.mark.tier2
def test_heketi_cluster_delete(self):
"""Test and validateheketi cluster delete operation"""
# Create heketi cluster
@@ -103,7 +103,7 @@ class TestClusterOperationsTestCases(baseclass.BaseClass):
% (cluster_info["id"], cluster_list["clusters"]))
self.assertNotIn(cluster_info["id"], cluster_list["clusters"], err_msg)
- @pytest.mark.tier1
+ @pytest.mark.tier2
def test_create_heketi_cluster_and_add_node(self):
"""Test heketi node add to a newly created cluster"""
storage_host_info = g.config.get("additional_gluster_servers")
@@ -212,7 +212,7 @@ class TestClusterOperationsTestCases(baseclass.BaseClass):
self.assertEqual(
zone, storage_zone, err_msg % ("zone", zone, storage_zone))
- @pytest.mark.tier1
+ @pytest.mark.tier2
def test_heketi_server_operations_cleanup_on_idle_setup(self):
"""Run heketi db clean up on an idle setup"""
h_node, h_url = self.heketi_client_node, self.heketi_server_url
diff --git a/tests/functional/heketi/test_heketi_create_volume.py b/tests/functional/heketi/test_heketi_create_volume.py
index a15d4447..f7fb6126 100644
--- a/tests/functional/heketi/test_heketi_create_volume.py
+++ b/tests/functional/heketi/test_heketi_create_volume.py
@@ -110,7 +110,7 @@ class TestHeketiVolume(BaseClass):
"of Heketi volumes before and after volume creation: %s\n%s" % (
existing_h_vol_list, h_vol_list))
- @pytest.mark.tier1
+ @pytest.mark.tier2
@podcmd.GlustoPod()
def test_create_vol_and_retrieve_vol_info(self):
"""Validate heketi and gluster volume info"""
@@ -194,7 +194,7 @@ class TestHeketiVolume(BaseClass):
"\n%s" % (volume_ids[2], existing_volumes))
g.log.info("Sucessfully verified the topology info")
- @pytest.mark.tier1
+ @pytest.mark.tier2
def test_to_check_deletion_of_cluster(self):
"""Validate deletion of cluster with volumes"""
# List heketi volumes
@@ -435,7 +435,7 @@ class TestHeketiVolume(BaseClass):
heketi_blockvolume_delete(h_node, h_url, block_vol)
heketi_volume_delete(h_node, h_url, volume, raise_on_error=False)
- @pytest.mark.tier1
+ @pytest.mark.tier2
@ddt.data("", "block")
def test_verify_delete_heketi_volumes_pending_entries_in_db(
self, vol_type):
@@ -592,7 +592,7 @@ class TestHeketiVolume(BaseClass):
act_brick_count, exp_brick_count,
err_msg.format(act_brick_count, exp_brick_count, "brick"))
- @pytest.mark.tier1
+ @pytest.mark.tier2
@ddt.data('', 'block')
def test_verify_create_heketi_volumes_pending_entries_in_db(
self, vol_type):
diff --git a/tests/functional/heketi/test_heketi_device_operations.py b/tests/functional/heketi/test_heketi_device_operations.py
index 71430a14..c8a9b991 100755
--- a/tests/functional/heketi/test_heketi_device_operations.py
+++ b/tests/functional/heketi/test_heketi_device_operations.py
@@ -298,7 +298,7 @@ class TestHeketiDeviceOperations(BaseClass):
"Some of the '%s' volume bricks is present of the removed "
"'%s' device." % (vol_info['id'], lowest_device_id))
- @pytest.mark.tier1
+ @pytest.mark.tier2
@ddt.data((True, False), (False, False), (True, True))
@ddt.unpack
def test_heketi_device_remove_delete(
@@ -428,7 +428,7 @@ class TestHeketiDeviceOperations(BaseClass):
heketi_device_disable, heketi_node, heketi_url, device_id)
raise
- @pytest.mark.tier1
+ @pytest.mark.tier2
def test_heketi_device_delete(self):
"""Test Heketi device delete operation"""
@@ -488,7 +488,7 @@ class TestHeketiDeviceOperations(BaseClass):
"after the device deletion" % (device_id, node_id))
self.assertNotIn(device_id, node_info_after_deletion, msg)
- @pytest.mark.tier1
+ @pytest.mark.tier2
def test_heketi_device_info(self):
"""Validate whether device related information is displayed"""
@@ -554,7 +554,7 @@ class TestHeketiDeviceOperations(BaseClass):
self.assertEqual(brick[key], brick_from_t[key])
self.assertEqual(brick_match_count, len(device_info['bricks']))
- @pytest.mark.tier1
+ @pytest.mark.tier2
def test_device_delete_with_bricks(self):
"""Validate device deletion with existing bricks on the device"""
h_node, h_url = self.heketi_client_node, self.heketi_server_url
diff --git a/tests/functional/heketi/test_heketi_metrics.py b/tests/functional/heketi/test_heketi_metrics.py
index 9f161607..ec0fc57b 100644
--- a/tests/functional/heketi/test_heketi_metrics.py
+++ b/tests/functional/heketi/test_heketi_metrics.py
@@ -172,7 +172,7 @@ class TestHeketiMetrics(BaseClass):
vol_count['cluster'], json=True)
self.assertEqual(vol_count['value'], len(cluster_info['volumes']))
- @pytest.mark.tier1
+ @pytest.mark.tier2
def test_heketi_metrics_with_topology_info(self):
"""Validate heketi metrics generation"""
self.verify_heketi_metrics_with_topology_info()
@@ -222,7 +222,7 @@ class TestHeketiMetrics(BaseClass):
self.verify_heketi_metrics_with_topology_info()
- @pytest.mark.tier1
+ @pytest.mark.tier2
def test_heketi_metrics_validating_vol_count_on_vol_creation(self):
"""Validate heketi metrics VolumeCount after volume creation"""
@@ -247,7 +247,7 @@ class TestHeketiMetrics(BaseClass):
self.verify_volume_count()
- @pytest.mark.tier1
+ @pytest.mark.tier2
def test_heketi_metrics_validating_vol_count_on_vol_deletion(self):
"""Validate heketi metrics VolumeCount after volume deletion"""
@@ -287,7 +287,7 @@ class TestHeketiMetrics(BaseClass):
self.assertNotIn(vol['id'], volume_list)
self.verify_volume_count()
- @pytest.mark.tier1
+ @pytest.mark.tier2
def test_heketi_metrics_validating_cluster_count(self):
"""Validate 'cluster count' in heketi metrics"""
cluster_list = heketi_cluster_list(
@@ -305,7 +305,7 @@ class TestHeketiMetrics(BaseClass):
self.assertEqual(
len(cluster_list['clusters']), metrics['heketi_cluster_count'])
- @pytest.mark.tier1
+ @pytest.mark.tier2
def test_heketi_metrics_validating_existing_node_count(self):
"""Validate existing 'node count' in heketi metrics"""
metrics = get_heketi_metrics(
diff --git a/tests/functional/heketi/test_heketi_node_operations.py b/tests/functional/heketi/test_heketi_node_operations.py
index adcbf6ef..65e3554d 100644
--- a/tests/functional/heketi/test_heketi_node_operations.py
+++ b/tests/functional/heketi/test_heketi_node_operations.py
@@ -54,7 +54,7 @@ class TestHeketiNodeOperations(baseclass.BaseClass):
"Heketi volume list %s is not equal to gluster volume list %s"
% (node_ips, hostnames))
- @pytest.mark.tier1
+ @pytest.mark.tier2
def test_heketi_node_info(self):
"""Test heketi node info operation
"""
@@ -254,7 +254,7 @@ class TestHeketiNodeOperations(baseclass.BaseClass):
# Add node to valid cluster id
self.heketi_node_add_with_valid_cluster()
- @pytest.mark.tier1
+ @pytest.mark.tier2
def test_validate_heketi_node_add_with_db_check(self):
"""Test heketi db check after node add operation"""
if not self.is_containerized_gluster():
@@ -285,7 +285,7 @@ class TestHeketiNodeOperations(baseclass.BaseClass):
" not as expected".format(initial_node_count, final_node_count))
self.assertEqual(initial_node_count + 1, final_node_count, msg)
- @pytest.mark.tier1
+ @pytest.mark.tier2
def test_heketi_node_add_with_invalid_cluster(self):
"""Test heketi node add operation with invalid cluster id"""
storage_hostname, cluster_id = None, utils.get_random_str(size=33)
diff --git a/tests/functional/heketi/test_heketi_volume_operations.py b/tests/functional/heketi/test_heketi_volume_operations.py
index ef9ae213..b3192d02 100644
--- a/tests/functional/heketi/test_heketi_volume_operations.py
+++ b/tests/functional/heketi/test_heketi_volume_operations.py
@@ -32,7 +32,7 @@ class TestHeketiVolumeOperations(BaseClass):
super(TestHeketiVolumeOperations, cls).setUpClass()
cls.volume_size = 1
- @pytest.mark.tier1
+ @pytest.mark.tier2
def test_heketi_with_default_options(self):
"""
Test to create volume with default options.
@@ -52,7 +52,7 @@ class TestHeketiVolumeOperations(BaseClass):
"Expected Size: %s, Actual Size: %s"
% (self.volume_size, vol_info['size'])))
- @pytest.mark.tier1
+ @pytest.mark.tier2
def test_heketi_with_expand_volume(self):
"""
Test volume expand and size if updated correctly in heketi-cli info
@@ -86,7 +86,7 @@ class TestHeketiVolumeOperations(BaseClass):
"Size: %s" % (str(expected_size),
str(volume_info['size']))))
- @pytest.mark.tier1
+ @pytest.mark.tier2
def test_heketi_volume_mount(self):
self.node = self.ocp_master_node[0]
try:
@@ -179,7 +179,7 @@ class TestHeketiVolumeOperations(BaseClass):
pids.append([g_node, pid])
return pids
- @pytest.mark.tier1
+ @pytest.mark.tier2
@podcmd.GlustoPod()
def test_heketi_volume_snapshot_create_with_one_brick_down(self):
"""
@@ -242,7 +242,7 @@ class TestHeketiVolumeOperations(BaseClass):
"Expecting Snapshot count before {} and after creation {} to be "
"same".format(snap_list_before, snap_list_after))
- @pytest.mark.tier1
+ @pytest.mark.tier2
def test_heketi_volume_create_mutiple_sizes(self):
"""Validate creation of heketi volume with differnt sizes"""
sizes, required_space = [15, 50, 100], 495
@@ -259,7 +259,7 @@ class TestHeketiVolumeOperations(BaseClass):
vol_id = heketi_volume_create(h_node, h_url, size, json=True)['id']
self.addCleanup(heketi_volume_delete, h_node, h_url, vol_id)
- @pytest.mark.tier1
+ @pytest.mark.tier2
@podcmd.GlustoPod()
def test_heketi_volume_snapshot_delete(self):
"""Test heketi volume snapshot delete operation"""
diff --git a/tests/functional/heketi/test_server_state_examine_gluster.py b/tests/functional/heketi/test_server_state_examine_gluster.py
index 6b66f13a..0a93c9eb 100644
--- a/tests/functional/heketi/test_server_state_examine_gluster.py
+++ b/tests/functional/heketi/test_server_state_examine_gluster.py
@@ -27,7 +27,7 @@ class TestHeketiServerStateExamineGluster(BaseClass):
self.skipTest("heketi-client package %s does not support server "
"state examine gluster" % version.v_str)
- @pytest.mark.tier1
+ @pytest.mark.tier2
def test_volume_inconsistencies(self):
# Examine Gluster cluster and Heketi that there is no inconsistencies
out = heketi_ops.heketi_examine_gluster(
@@ -130,7 +130,7 @@ class TestHeketiServerStateExamineGluster(BaseClass):
db_bricks_count, topology_bricks_count))
self.assertEqual(topology_bricks_count, db_bricks_count, msg)
- @pytest.mark.tier1
+ @pytest.mark.tier2
@ddt.data('', 'block')
def test_compare_heketi_volumes(self, vol_type):
"""Validate file/block volume count using heketi gluster examine"""
diff --git a/tests/functional/heketi/test_volume_creation.py b/tests/functional/heketi/test_volume_creation.py
index 54c808dd..d96fa9e0 100644
--- a/tests/functional/heketi/test_volume_creation.py
+++ b/tests/functional/heketi/test_volume_creation.py
@@ -158,7 +158,7 @@ class TestVolumeCreationTestCases(BaseClass):
vol_fail,
"Volume should have not been created. Out: %s" % vol_fail)
- @pytest.mark.tier1
+ @pytest.mark.tier2
@podcmd.GlustoPod()
def test_volume_create_replica_2(self):
"""Validate creation of a replica 2 volume"""
@@ -189,7 +189,7 @@ class TestVolumeCreationTestCases(BaseClass):
"Brick amount is expected to be 2. "
"Actual amount is '%s'" % brick_amount)
- @pytest.mark.tier1
+ @pytest.mark.tier2
@podcmd.GlustoPod()
def test_volume_create_snapshot_enabled(self):
"""Validate volume creation with snapshot enabled"""
@@ -386,7 +386,7 @@ class TestVolumeCreationTestCases(BaseClass):
% (vol_name, gluster_v_info['brickCount']))
self.assertFalse(int(gluster_v_info['brickCount']) % 3, msg)
- @pytest.mark.tier1
+ @pytest.mark.tier2
def test_create_volume_with_same_name(self):
"""Test create two volumes with the same name and verify that 2nd one
is failing with the appropriate error.
diff --git a/tests/functional/heketi/test_volume_multi_req.py b/tests/functional/heketi/test_volume_multi_req.py
index 0780be1e..44ea803b 100644
--- a/tests/functional/heketi/test_volume_multi_req.py
+++ b/tests/functional/heketi/test_volume_multi_req.py
@@ -336,7 +336,7 @@ class TestVolumeMultiReq(BaseClass):
# verify this volume in heketi
self.assertIn(c2.heketiVolumeName, now_vols)
- @pytest.mark.tier1
+ @pytest.mark.tier2
# NOTE(jjm): I've noticed that on the system I'm using (RHEL7).
# with count=8 things start to back up a bit.
# I needed to increase some timeouts to get this to pass.
diff --git a/tests/functional/provisioning/test_dynamic_provisioning_block.py b/tests/functional/provisioning/test_dynamic_provisioning_block.py
index cce553b0..215354da 100755
--- a/tests/functional/provisioning/test_dynamic_provisioning_block.py
+++ b/tests/functional/provisioning/test_dynamic_provisioning_block.py
@@ -118,7 +118,7 @@ class TestDynamicProvisioningBlockP0(GlusterBlockBaseClass):
"""
self.dynamic_provisioning_glusterblock(set_hacount=True)
- @pytest.mark.tier1
+ @pytest.mark.tier2
def test_dynamic_provisioning_glusterblock_hacount_false(self):
"""Validate storage-class mandatory parameters for block
"""
@@ -255,7 +255,7 @@ class TestDynamicProvisioningBlockP0(GlusterBlockBaseClass):
ret, out, err = async_io.async_communicate()
self.assertEqual(ret, 0, "IO %s failed on %s" % (io_cmd, self.node))
- @pytest.mark.tier1
+ @pytest.mark.tier2
def test_glusterblock_logs_presence_verification(self):
"""Validate presence of glusterblock provisioner POD and it's status"""
@@ -549,7 +549,7 @@ class TestDynamicProvisioningBlockP0(GlusterBlockBaseClass):
# create a block pvc greater than default BHV size
self.create_and_wait_for_pvc(pvc_size=(default_bhv_size + 1))
- @pytest.mark.tier1
+ @pytest.mark.tier2
def test_expansion_of_block_hosting_volume_using_heketi(self):
"""Verify that after expanding block hosting volume we are able to
consume the expanded space"""
@@ -839,7 +839,7 @@ class TestDynamicProvisioningBlockP0(GlusterBlockBaseClass):
" No. of BHV after the test : {}".format(len(bhv_list), bhv_post))
self.assertEqual(bhv_post, (len(bhv_list) + 2), err_msg)
- @pytest.mark.tier1
+ @pytest.mark.tier2
def test_100gb_block_pvc_create_and_delete_twice(self):
"""Validate creation and deletion of blockvoume of size 100GB"""
# Define required space, bhv size required for on 100GB block PVC
diff --git a/tests/functional/provisioning/test_dynamic_provisioning_file.py b/tests/functional/provisioning/test_dynamic_provisioning_file.py
index c711186c..c1f9d078 100644
--- a/tests/functional/provisioning/test_dynamic_provisioning_file.py
+++ b/tests/functional/provisioning/test_dynamic_provisioning_file.py
@@ -114,7 +114,7 @@ class TestDynamicProvisioningP0(BaseClass):
g.log.info("test_dynamic_provisioning_glusterfile")
self.dynamic_provisioning_glusterfile(False)
- @pytest.mark.tier1
+ @pytest.mark.tier2
def test_dynamic_provisioning_glusterfile_volname_prefix(self):
"""Validate dynamic provisioning for gluster file with vol name prefix
"""
@@ -481,7 +481,7 @@ class TestDynamicProvisioningP0(BaseClass):
oc_delete(self.node, 'pv', pv_name)
wait_for_resource_absence(self.node, 'pv', pv_name)
- @pytest.mark.tier1
+ @pytest.mark.tier2
def test_usage_of_default_storage_class(self):
"""Validate PVs creation for SC with default custom volname prefix"""
diff --git a/tests/functional/provisioning/test_pv_resize.py b/tests/functional/provisioning/test_pv_resize.py
index abadc4ee..a7baca00 100644
--- a/tests/functional/provisioning/test_pv_resize.py
+++ b/tests/functional/provisioning/test_pv_resize.py
@@ -50,7 +50,7 @@ class TestPvResizeClass(BaseClass):
g.log.error(msg)
raise self.skipTest(msg)
- @pytest.mark.tier1
+ @pytest.mark.tier2
@ddt.data(
(True, True),
(False, True),
@@ -258,7 +258,7 @@ class TestPvResizeClass(BaseClass):
"""Validate PVC resize when resized by exact available free space"""
self._pv_resize(exceed_free_space=False)
- @pytest.mark.tier1
+ @pytest.mark.tier2
def test_pv_resize_try_shrink_pv_size(self):
"""Validate whether reducing the PV size is allowed"""
dir_path = "/mnt/"
@@ -295,7 +295,7 @@ class TestPvResizeClass(BaseClass):
self.assertEqual(
ret, 0, "Failed to execute command %s on %s" % (cmd, node))
- @pytest.mark.tier1
+ @pytest.mark.tier2
def test_pv_resize_when_heketi_down(self):
"""Create a PVC and try to expand it when heketi is down, It should
fail. After heketi is up, expand PVC should work.
@@ -418,7 +418,7 @@ class TestPvResizeClass(BaseClass):
resize_pvc(self.ocp_master_node[0], pvc_name, available_size_gb)
verify_pvc_size(self.ocp_master_node[0], pvc_name, available_size_gb)
- @pytest.mark.tier1
+ @pytest.mark.tier2
def test_pv_resize_device_disabled(self):
"""Validate resize after disabling all devices except one"""
h_node, h_url = self.heketi_client_node, self.heketi_server_url
diff --git a/tests/functional/provisioning/test_storage_class_cases.py b/tests/functional/provisioning/test_storage_class_cases.py
index 05ee49cc..3a2b9dfa 100644
--- a/tests/functional/provisioning/test_storage_class_cases.py
+++ b/tests/functional/provisioning/test_storage_class_cases.py
@@ -181,7 +181,7 @@ class TestStorageClassCases(BaseClass):
validate_multipath_pod(
self.ocp_master_node[0], pod_name, hacount, list(mpaths)[0])
- @pytest.mark.tier1
+ @pytest.mark.tier2
@ddt.data(
{"volumetype": "dist-rep:3"},
{"resturl": "http://10.0.0.1:8080"},
@@ -195,7 +195,7 @@ class TestStorageClassCases(BaseClass):
"""Validate glusterfile storage with different incorrect parameters"""
self.create_sc_with_parameter("glusterfile", parameter=parameter)
- @pytest.mark.tier1
+ @pytest.mark.tier2
@ddt.data(
{"resturl": "http://10.0.0.1:8080"},
{"restsecretname": "fakerestsecretname",
@@ -331,7 +331,7 @@ class TestStorageClassCases(BaseClass):
"Invalid chapauthenabled value '%s'" % chapauthenabled
)
- @pytest.mark.tier1
+ @pytest.mark.tier2
def test_create_and_verify_pvc_with_volume_name_prefix(self):
"""create and verify pvc with volname prefix on an app pod"""
if get_openshift_version() < "3.9":
@@ -358,7 +358,7 @@ class TestStorageClassCases(BaseClass):
"Failed to read Endpoints of %s on %s " % (
pv_name, self.ocp_master_node[0]))
- @pytest.mark.tier1
+ @pytest.mark.tier2
def test_try_to_create_sc_with_duplicated_name(self):
"""Verify SC creation fails with duplicate name"""
sc_name = "test-sc-duplicated-name-" + utils.get_random_str()
@@ -367,7 +367,7 @@ class TestStorageClassCases(BaseClass):
with self.assertRaises(AssertionError):
self.create_storage_class(sc_name=sc_name)
- @pytest.mark.tier1
+ @pytest.mark.tier2
@ddt.data('secretName', 'secretNamespace', None)
def test_sc_glusterfile_missing_parameter(self, parameter):
"""Validate glusterfile storage with missing parameters"""
@@ -397,7 +397,7 @@ class TestStorageClassCases(BaseClass):
with self.assertRaises(ExecutionError):
verify_pvc_status_is_bound(node, pvc_name, timeout=1)
- @pytest.mark.tier1
+ @pytest.mark.tier2
def test_sc_create_with_clusterid(self):
"""Create storage class with 'cluster id'"""
h_cluster_list = heketi_cluster_list(