Merge "Adding description for testcases - volume part4"
diff --git a/releasenotes/notes/25/subunt-describe-call-verbose-arg-fix.yaml b/releasenotes/notes/25/subunt-describe-call-verbose-arg-fix.yaml
new file mode 100644
index 0000000..d2a644e
--- /dev/null
+++ b/releasenotes/notes/25/subunt-describe-call-verbose-arg-fix.yaml
@@ -0,0 +1,10 @@
+---
+fixes:
+ - |
+ Fixed bug #1890060. tempest subunit_describe_calls --verbose not working with Cliff CLI.
+ The subunit_describe_calls --verbose argument was a boolean and worked in the non Cliff CLI
+ which is now deprecated, but does not work with cliff since --verbase is a standard cliff
+ argument which is an int. Since the tool is in lib directory we cannot change the interface,
+ so we add a new argument -a --all-stdout that will allow cliff CLI to support the
+ feature in subunnit_describe_calls to print request and response headers and bodies
+ to stdout.
\ No newline at end of file
diff --git a/releasenotes/notes/add-can-migrate-between-any-hosts-config-option-x8ah4f9737a28e9b.yaml b/releasenotes/notes/add-can-migrate-between-any-hosts-config-option-x8ah4f9737a28e9b.yaml
new file mode 100644
index 0000000..26fe01a
--- /dev/null
+++ b/releasenotes/notes/add-can-migrate-between-any-hosts-config-option-x8ah4f9737a28e9b.yaml
@@ -0,0 +1,9 @@
+---
+features:
+ - Add a new config option can_migrate_between_any_hosts in the
+ compute-feature-enabled section, which can be set to False for environment
+ with non homogeneous compute nodes, so that it can select a destination
+ host for migrating automatically, otherwise the testcase may fail
+ unexpectedly. e.g., if source host is with CPU "E5-2699 v4" and the
+ selected target host is with CPU "E5-2670 v3", the live-migration will
+ fail because of the downgrade issue.
diff --git a/tempest/api/compute/admin/test_live_migration.py b/tempest/api/compute/admin/test_live_migration.py
index a845c72..941315e 100644
--- a/tempest/api/compute/admin/test_live_migration.py
+++ b/tempest/api/compute/admin/test_live_migration.py
@@ -71,6 +71,10 @@
def _live_migrate(self, server_id, target_host, state,
volume_backed=False):
+ # If target_host is None, check whether source host is different with
+ # the new host after migration.
+ if target_host is None:
+ source_host = self.get_host_for_server(server_id)
self._migrate_server_to(server_id, target_host, volume_backed)
waiters.wait_for_server_status(self.servers_client, server_id, state)
migration_list = (self.admin_migration_client.list_migrations()
@@ -82,8 +86,12 @@
if (live_migration['instance_uuid'] == server_id):
msg += "\n%s" % live_migration
msg += "]"
- self.assertEqual(target_host, self.get_host_for_server(server_id),
- msg)
+ if target_host is None:
+ self.assertNotEqual(source_host,
+ self.get_host_for_server(server_id), msg)
+ else:
+ self.assertEqual(target_host, self.get_host_for_server(server_id),
+ msg)
class LiveMigrationTest(LiveMigrationTestBase):
@@ -105,7 +113,11 @@
server_id = self.create_test_server(wait_until="ACTIVE",
volume_backed=volume_backed)['id']
source_host = self.get_host_for_server(server_id)
- destination_host = self.get_host_other_than(server_id)
+ if not CONF.compute_feature_enabled.can_migrate_between_any_hosts:
+ # not to specify a host so that the scheduler will pick one
+ destination_host = None
+ else:
+ destination_host = self.get_host_other_than(server_id)
if state == 'PAUSED':
self.admin_servers_client.pause_server(server_id)
@@ -123,11 +135,17 @@
self._live_migrate(server_id, source_host, state, volume_backed)
@decorators.idempotent_id('1dce86b8-eb04-4c03-a9d8-9c1dc3ee0c7b')
+ @testtools.skipUnless(CONF.compute_feature_enabled.
+ block_migration_for_live_migration,
+ 'Block Live migration not available')
def test_live_block_migration(self):
"""Test live migrating an active server"""
self._test_live_migration()
@decorators.idempotent_id('1e107f21-61b2-4988-8f22-b196e938ab88')
+ @testtools.skipUnless(CONF.compute_feature_enabled.
+ block_migration_for_live_migration,
+ 'Block Live migration not available')
@testtools.skipUnless(CONF.compute_feature_enabled.pause,
'Pause is not available.')
def test_live_block_migration_paused(self):
@@ -155,7 +173,11 @@
"""Test live migrating a server with volume attached"""
server = self.create_test_server(wait_until="ACTIVE")
server_id = server['id']
- target_host = self.get_host_other_than(server_id)
+ if not CONF.compute_feature_enabled.can_migrate_between_any_hosts:
+ # not to specify a host so that the scheduler will pick one
+ target_host = None
+ else:
+ target_host = self.get_host_other_than(server_id)
volume = self.create_volume()
diff --git a/tempest/api/compute/servers/test_list_server_filters.py b/tempest/api/compute/servers/test_list_server_filters.py
index 7f62c64..990dd52 100644
--- a/tempest/api/compute/servers/test_list_server_filters.py
+++ b/tempest/api/compute/servers/test_list_server_filters.py
@@ -296,21 +296,32 @@
for ip in ip_list:
self.assertNotIn(ip_list[ip], map(lambda x: x['id'], servers))
- @decorators.skip_because(bug="1540645")
@decorators.idempotent_id('a905e287-c35e-42f2-b132-d02b09f3654a')
def test_list_servers_filtered_by_ip_regex(self):
"""Filter the list of servers by part of server ip address"""
- # Here should be listed all servers
if not self.fixed_network_name:
msg = 'fixed_network_name needs to be configured to run this test'
raise self.skipException(msg)
- self.s1 = self.client.show_server(self.s1['id'])['server']
- addr_spec = self.s1['addresses'][self.fixed_network_name][0]
- ip = addr_spec['addr'][0:-3]
+ # query addresses of the 3 servers
+ addrs = []
+ for s in [self.s1, self.s2, self.s3]:
+ s_show = self.client.show_server(s['id'])['server']
+ addr_spec = s_show['addresses'][self.fixed_network_name][0]
+ addrs.append(addr_spec['addr'])
+ # find common part of the 3 ip addresses
+ prefix = ''
+ addrs_len = [len(a) for a in addrs]
+ addrs_len.sort()
+ # iterate over the smallest length of an ip
+ for i in range(addrs_len[0]):
+ if not addrs[0][i] == addrs[1][i] == addrs[2][i]:
+ break
+ prefix += addrs[0][i]
+
if addr_spec['version'] == 4:
- params = {'ip': ip}
+ params = {'ip': prefix}
else:
- params = {'ip6': ip}
+ params = {'ip6': prefix}
# capture all servers in case something goes wrong
all_servers = self.client.list_servers(detail=True)
body = self.client.list_servers(**params)
diff --git a/tempest/api/identity/v2/test_api_discovery.py b/tempest/api/identity/v2/test_api_discovery.py
index 5b9d38c..afda104 100644
--- a/tempest/api/identity/v2/test_api_discovery.py
+++ b/tempest/api/identity/v2/test_api_discovery.py
@@ -18,11 +18,12 @@
class TestApiDiscovery(base.BaseIdentityV2Test):
- """Tests for API discovery features."""
+ """Tests for identity v2 API discovery features."""
@decorators.attr(type='smoke')
@decorators.idempotent_id('ea889a68-a15f-4166-bfb1-c12456eae853')
def test_api_version_resources(self):
+ """Test showing identity v2 api version resources"""
descr = self.non_admin_client.show_api_description()['version']
expected_resources = ('id', 'links', 'media-types', 'status',
'updated')
@@ -34,6 +35,7 @@
@decorators.attr(type='smoke')
@decorators.idempotent_id('007a0be0-78fe-4fdb-bbee-e9216cc17bb2')
def test_api_media_types(self):
+ """Test showing identity v2 api version media type"""
descr = self.non_admin_client.show_api_description()['version']
# Get MIME type bases and descriptions
media_types = [(media_type['base'], media_type['type']) for
@@ -49,6 +51,7 @@
@decorators.attr(type='smoke')
@decorators.idempotent_id('77fd6be0-8801-48e6-b9bf-38cdd2f253ec')
def test_api_version_statuses(self):
+ """Test showing identity v2 api version status"""
descr = self.non_admin_client.show_api_description()['version']
status = descr['status'].lower()
supported_statuses = ['current', 'stable', 'experimental',
diff --git a/tempest/api/identity/v2/test_extension.py b/tempest/api/identity/v2/test_extension.py
index c538c14..13555bd 100644
--- a/tempest/api/identity/v2/test_extension.py
+++ b/tempest/api/identity/v2/test_extension.py
@@ -18,10 +18,11 @@
class ExtensionTestJSON(base.BaseIdentityV2Test):
+ """Test extensions in identity v2 API"""
@decorators.idempotent_id('85f3f661-f54c-4d48-b563-72ae952b9383')
def test_list_extensions(self):
- # List all the extensions
+ """List all the identity extensions via v2 API"""
body = self.non_admin_client.list_extensions()['extensions']['values']
self.assertNotEmpty(body)
keys = ['name', 'updated', 'alias', 'links',
diff --git a/tempest/api/identity/v2/test_tenants.py b/tempest/api/identity/v2/test_tenants.py
index b2a6d13..1752b65 100644
--- a/tempest/api/identity/v2/test_tenants.py
+++ b/tempest/api/identity/v2/test_tenants.py
@@ -19,11 +19,13 @@
class IdentityTenantsTest(base.BaseIdentityV2Test):
+ """Test listing tenants in identity v2 API"""
credentials = ['primary', 'alt']
@decorators.idempotent_id('ecae2459-243d-4ba1-ad02-65f15dc82b78')
def test_list_tenants_returns_only_authorized_tenants(self):
+ """Test listing tenants only returns authorized tenants via v2 API"""
alt_tenant_name = self.os_alt.credentials.tenant_name
resp = self.non_admin_tenants_client.list_tenants()
diff --git a/tempest/api/identity/v2/test_tokens.py b/tempest/api/identity/v2/test_tokens.py
index 64b81c2..a928ad9 100644
--- a/tempest/api/identity/v2/test_tokens.py
+++ b/tempest/api/identity/v2/test_tokens.py
@@ -20,10 +20,11 @@
class TokensTest(base.BaseIdentityV2Test):
+ """Test tokens in identity v2 API"""
@decorators.idempotent_id('65ae3b78-91ff-467b-a705-f6678863b8ec')
def test_create_token(self):
-
+ """Test creating token for user via v2 API"""
token_client = self.non_admin_token_client
# get a token for the user
diff --git a/tempest/api/identity/v2/test_users.py b/tempest/api/identity/v2/test_users.py
index 2eea860..a63b45c 100644
--- a/tempest/api/identity/v2/test_users.py
+++ b/tempest/api/identity/v2/test_users.py
@@ -28,6 +28,7 @@
class IdentityUsersTest(base.BaseIdentityV2Test):
+ """Test user password in identity v2 API"""
@classmethod
def resource_setup(cls):
@@ -85,6 +86,7 @@
'immutable user source and solely '
'provides read-only access to users.')
def test_user_update_own_password(self):
+ """test updating user's own password via v2 API"""
old_pass = self.creds.password
old_token = self.non_admin_users_client.token
new_pass = data_utils.rand_password()
diff --git a/tempest/api/identity/v3/test_api_discovery.py b/tempest/api/identity/v3/test_api_discovery.py
index e87d1cd..ebb96fd 100644
--- a/tempest/api/identity/v3/test_api_discovery.py
+++ b/tempest/api/identity/v3/test_api_discovery.py
@@ -22,10 +22,11 @@
class TestApiDiscovery(base.BaseIdentityV3Test):
- """Tests for API discovery features."""
+ """Tests for identity API discovery features."""
@decorators.idempotent_id('79aec9ae-710f-4c54-a4fc-3aa25b4feac3')
def test_identity_v3_existence(self):
+ """Test that identity v3 version should exist"""
versions = self.non_admin_versions_client.list_versions()
found = any(
"v3" in version.get('id')
@@ -35,9 +36,12 @@
@decorators.idempotent_id('721f480f-35b6-46c7-846e-047e6acea0dc')
@decorators.attr(type='smoke')
def test_list_api_versions(self):
- # NOTE: Actually this API doesn't depend on v3 API at all, because
- # the API operation is "GET /" without v3's endpoint. The reason of
- # this test path is just v3 API is CURRENT on Keystone side.
+ """Test listing identity api versions
+
+ NOTE: Actually this API doesn't depend on v3 API at all, because
+ the API operation is "GET /" without v3's endpoint. The reason of
+ this test path is just v3 API is CURRENT on Keystone side.
+ """
versions = self.non_admin_versions_client.list_versions()
expected_resources = ('id', 'links', 'media-types', 'status',
'updated')
@@ -49,6 +53,7 @@
@decorators.attr(type='smoke')
@decorators.idempotent_id('b9232f5e-d9e5-4d97-b96c-28d3db4de1bd')
def test_api_version_resources(self):
+ """Test showing identity v3 api version resources"""
descr = self.non_admin_client.show_api_description()['version']
expected_resources = ('id', 'links', 'media-types', 'status',
'updated')
@@ -60,6 +65,7 @@
@decorators.attr(type='smoke')
@decorators.idempotent_id('657c1970-4722-4189-8831-7325f3bc4265')
def test_api_media_types(self):
+ """Test showing identity v3 api version media type"""
descr = self.non_admin_client.show_api_description()['version']
# Get MIME type bases and descriptions
media_types = [(media_type['base'], media_type['type']) for
@@ -75,6 +81,7 @@
@decorators.attr(type='smoke')
@decorators.idempotent_id('8879a470-abfb-47bb-bb8d-5a7fd279ad1e')
def test_api_version_statuses(self):
+ """Test showing identity v3 api version status"""
descr = self.non_admin_client.show_api_description()['version']
status = descr['status'].lower()
supported_statuses = ['current', 'stable', 'experimental',
diff --git a/tempest/api/identity/v3/test_domains.py b/tempest/api/identity/v3/test_domains.py
index 9f132dd..bb62ea6 100644
--- a/tempest/api/identity/v3/test_domains.py
+++ b/tempest/api/identity/v3/test_domains.py
@@ -21,6 +21,7 @@
class DefaultDomainTestJSON(base.BaseIdentityV3Test):
+ """Test identity default domains"""
@classmethod
def setup_clients(cls):
@@ -35,5 +36,6 @@
@decorators.attr(type='smoke')
@decorators.idempotent_id('17a5de24-e6a0-4e4a-a9ee-d85b6e5612b5')
def test_default_domain_exists(self):
+ """Test showing default domain"""
domain = self.domains_client.show_domain(self.domain_id)['domain']
self.assertTrue(domain['enabled'])
diff --git a/tempest/api/identity/v3/test_projects.py b/tempest/api/identity/v3/test_projects.py
index bbb4013..338b57b 100644
--- a/tempest/api/identity/v3/test_projects.py
+++ b/tempest/api/identity/v3/test_projects.py
@@ -19,11 +19,13 @@
class IdentityV3ProjectsTest(base.BaseIdentityV3Test):
+ """Test identity projects"""
credentials = ['primary', 'alt']
@decorators.idempotent_id('86128d46-e170-4644-866a-cc487f699e1d')
def test_list_projects_returns_only_authorized_projects(self):
+ """Test listing projects only returns authorized projects"""
alt_project_name = self.os_alt.credentials.project_name
resp = self.non_admin_users_client.list_user_projects(
self.os_primary.credentials.user_id)
diff --git a/tempest/api/identity/v3/test_tokens.py b/tempest/api/identity/v3/test_tokens.py
index cb05f39..b201285 100644
--- a/tempest/api/identity/v3/test_tokens.py
+++ b/tempest/api/identity/v3/test_tokens.py
@@ -24,9 +24,11 @@
class TokensV3Test(base.BaseIdentityV3Test):
+ """Test identity tokens"""
@decorators.idempotent_id('a9512ac3-3909-48a4-b395-11f438e16260')
def test_validate_token(self):
+ """Test validating token for user"""
creds = self.os_primary.credentials
user_id = creds.user_id
username = creds.username
@@ -69,7 +71,7 @@
@decorators.idempotent_id('6f8e4436-fc96-4282-8122-e41df57197a9')
def test_create_token(self):
-
+ """Test creating token for user"""
creds = self.os_primary.credentials
user_id = creds.user_id
username = creds.username
@@ -120,9 +122,12 @@
@decorators.idempotent_id('0f9f5a5f-d5cd-4a86-8a5b-c5ded151f212')
def test_token_auth_creation_existence_deletion(self):
- # Tests basic token auth functionality in a way that is compatible with
- # pre-provisioned credentials. The default user is used for token
- # authentication.
+ """Test auth/check existence/delete token for user
+
+ Tests basic token auth functionality in a way that is compatible with
+ pre-provisioned credentials. The default user is used for token
+ authentication.
+ """
# Valid user's token is authenticated
user = self.os_primary.credentials
diff --git a/tempest/api/identity/v3/test_users.py b/tempest/api/identity/v3/test_users.py
index d4e7612..6425ea9 100644
--- a/tempest/api/identity/v3/test_users.py
+++ b/tempest/api/identity/v3/test_users.py
@@ -28,6 +28,7 @@
class IdentityV3UsersTest(base.BaseIdentityV3Test):
+ """Test identity user password"""
@classmethod
def resource_setup(cls):
@@ -82,6 +83,7 @@
'immutable user source and solely '
'provides read-only access to users.')
def test_user_update_own_password(self):
+ """Test updating user's own password"""
old_pass = self.creds.password
old_token = self.non_admin_client.token
new_pass = data_utils.rand_password()
@@ -111,6 +113,7 @@
'immutable user source and solely '
'provides read-only access to users.')
def test_password_history_check_self_service_api(self):
+ """Test checking password changing history"""
old_pass = self.creds.password
new_pass1 = data_utils.rand_password()
new_pass2 = data_utils.rand_password()
@@ -141,6 +144,7 @@
'Security compliance not available.')
@decorators.idempotent_id('a7ad8bbf-2cff-4520-8c1d-96332e151658')
def test_user_account_lockout(self):
+ """Test locking out user account after failure attempts"""
if (CONF.identity.user_lockout_failure_attempts <= 0 or
CONF.identity.user_lockout_duration <= 0):
raise self.skipException(
diff --git a/tempest/api/network/admin/test_ports.py b/tempest/api/network/admin/test_ports.py
index 51f2857..5f9f29f 100644
--- a/tempest/api/network/admin/test_ports.py
+++ b/tempest/api/network/admin/test_ports.py
@@ -48,6 +48,8 @@
"name": data_utils.rand_name(self.__class__.__name__)}
body = self.admin_ports_client.create_port(**post_body)
port = body['port']
+ self.addCleanup(self.admin_ports_client.wait_for_resource_deletion,
+ port['id'])
self.addCleanup(
test_utils.call_and_ignore_notfound_exc,
self.admin_ports_client.delete_port, port['id'])
@@ -63,6 +65,8 @@
"name": data_utils.rand_name(self.__class__.__name__)}
body = self.admin_ports_client.create_port(**post_body)
port = body['port']
+ self.addCleanup(self.admin_ports_client.wait_for_resource_deletion,
+ port['id'])
self.addCleanup(
test_utils.call_and_ignore_notfound_exc,
self.admin_ports_client.delete_port, port['id'])
@@ -82,6 +86,8 @@
"name": data_utils.rand_name(self.__class__.__name__)}
body = self.admin_ports_client.create_port(**post_body)
port = body['port']
+ self.addCleanup(self.admin_ports_client.wait_for_resource_deletion,
+ port['id'])
self.addCleanup(
test_utils.call_and_ignore_notfound_exc,
self.admin_ports_client.delete_port, port['id'])
@@ -110,6 +116,8 @@
name=data_utils.rand_name(self.__class__.__name__),
network_id=self.network['id'])
port = body['port']
+ self.addCleanup(self.admin_ports_client.wait_for_resource_deletion,
+ port['id'])
self.addCleanup(test_utils.call_and_ignore_notfound_exc,
self.admin_ports_client.delete_port, port['id'])
body = self.admin_ports_client.show_port(port['id'])
diff --git a/tempest/api/network/test_ports.py b/tempest/api/network/test_ports.py
index c6d049a..479578d 100644
--- a/tempest/api/network/test_ports.py
+++ b/tempest/api/network/test_ports.py
@@ -77,6 +77,8 @@
name=data_utils.rand_name(self.__class__.__name__))
port = body['port']
# Schedule port deletion with verification upon test completion
+ self.addCleanup(self.ports_client.wait_for_resource_deletion,
+ port['id'])
self.addCleanup(self._delete_port, port['id'])
self.assertTrue(port['admin_state_up'])
# Verify port update
@@ -99,6 +101,10 @@
created_ports = body['ports']
port1 = created_ports[0]
port2 = created_ports[1]
+ self.addCleanup(self.ports_client.wait_for_resource_deletion,
+ port1['id'])
+ self.addCleanup(self.ports_client.wait_for_resource_deletion,
+ port2['id'])
self.addCleanup(self._delete_port, port1['id'])
self.addCleanup(self._delete_port, port2['id'])
self.assertEqual(port1['network_id'], network1['id'])
@@ -126,6 +132,8 @@
body = self.ports_client.create_port(
network_id=net_id,
name=data_utils.rand_name(self.__class__.__name__))
+ self.addCleanup(self.ports_client.wait_for_resource_deletion,
+ body['port']['id'])
self.addCleanup(test_utils.call_and_ignore_notfound_exc,
self.ports_client.delete_port, body['port']['id'])
port = body['port']
@@ -183,11 +191,15 @@
port_1 = self.ports_client.create_port(
network_id=network['id'],
name=data_utils.rand_name(self.__class__.__name__))
+ self.addCleanup(self.ports_client.wait_for_resource_deletion,
+ port_1['port']['id'])
self.addCleanup(test_utils.call_and_ignore_notfound_exc,
self.ports_client.delete_port, port_1['port']['id'])
port_2 = self.ports_client.create_port(
network_id=network['id'],
name=data_utils.rand_name(self.__class__.__name__))
+ self.addCleanup(self.ports_client.wait_for_resource_deletion,
+ port_2['port']['id'])
self.addCleanup(test_utils.call_and_ignore_notfound_exc,
self.ports_client.delete_port, port_2['port']['id'])
# List ports filtered by fixed_ips
@@ -241,6 +253,8 @@
network_id=network['id'],
name=data_utils.rand_name(self.__class__.__name__),
fixed_ips=fixed_ips)
+ self.addCleanup(self.ports_client.wait_for_resource_deletion,
+ port_1['port']['id'])
self.addCleanup(test_utils.call_and_ignore_notfound_exc,
self.ports_client.delete_port, port_1['port']['id'])
fixed_ips = [{'subnet_id': subnet['id'], 'ip_address': ip_address_2}]
@@ -248,6 +262,8 @@
network_id=network['id'],
name=data_utils.rand_name(self.__class__.__name__),
fixed_ips=fixed_ips)
+ self.addCleanup(self.ports_client.wait_for_resource_deletion,
+ port_2['port']['id'])
self.addCleanup(test_utils.call_and_ignore_notfound_exc,
self.ports_client.delete_port, port_2['port']['id'])
@@ -307,6 +323,8 @@
# Add router interface to port created above
self.routers_client.add_router_interface(router['id'],
port_id=port['port']['id'])
+ self.addCleanup(self.ports_client.wait_for_resource_deletion,
+ port['port']['id'])
self.addCleanup(test_utils.call_and_ignore_notfound_exc,
self.routers_client.remove_router_interface,
router['id'], port_id=port['port']['id'])
@@ -343,6 +361,8 @@
# Create a port with multiple IP addresses
port = self.create_port(network,
fixed_ips=fixed_ips)
+ self.addCleanup(self.ports_client.wait_for_resource_deletion,
+ port['id'])
self.addCleanup(test_utils.call_and_ignore_notfound_exc,
self.ports_client.delete_port, port['id'])
self.assertEqual(2, len(port['fixed_ips']))
@@ -386,6 +406,8 @@
"admin_state_up": True,
"fixed_ips": fixed_ip_1}
body = self.ports_client.create_port(**post_body)
+ self.addCleanup(self.ports_client.wait_for_resource_deletion,
+ body['port']['id'])
self.addCleanup(test_utils.call_and_ignore_notfound_exc,
self.ports_client.delete_port, body['port']['id'])
port = body['port']
@@ -456,6 +478,8 @@
network_id=self.network['id'],
mac_address=free_mac_address,
name=data_utils.rand_name(self.__class__.__name__))
+ self.addCleanup(self.ports_client.wait_for_resource_deletion,
+ body['port']['id'])
self.addCleanup(test_utils.call_and_ignore_notfound_exc,
self.ports_client.delete_port, body['port']['id'])
port = body['port']
@@ -474,6 +498,8 @@
network = self._create_network()
self._create_subnet(network)
port = self.create_port(network, security_groups=[])
+ self.addCleanup(self.ports_client.wait_for_resource_deletion,
+ port['id'])
self.addCleanup(test_utils.call_and_ignore_notfound_exc,
self.ports_client.delete_port, port['id'])
self.assertIsNotNone(port['security_groups'])
diff --git a/tempest/api/volume/admin/test_volumes_actions.py b/tempest/api/volume/admin/test_volumes_actions.py
index 33e503f..ecddfba 100644
--- a/tempest/api/volume/admin/test_volumes_actions.py
+++ b/tempest/api/volume/admin/test_volumes_actions.py
@@ -108,5 +108,4 @@
waiters.wait_for_volume_resource_status(self.volumes_client,
volume_id, 'available')
vol_info = self.volumes_client.show_volume(volume_id)['volume']
- self.assertIn('attachments', vol_info)
self.assertEmpty(vol_info['attachments'])
diff --git a/tempest/api/volume/test_volumes_actions.py b/tempest/api/volume/test_volumes_actions.py
index dd0d0db..5b50bfa 100644
--- a/tempest/api/volume/test_volumes_actions.py
+++ b/tempest/api/volume/test_volumes_actions.py
@@ -92,7 +92,6 @@
self.volume['id'], 'available')
self.addCleanup(self.volumes_client.detach_volume, self.volume['id'])
volume = self.volumes_client.show_volume(self.volume['id'])['volume']
- self.assertIn('attachments', volume)
attachment = volume['attachments'][0]
self.assertEqual('/dev/%s' %
diff --git a/tempest/api/volume/test_volumes_get.py b/tempest/api/volume/test_volumes_get.py
index ade2deb..91728ab 100644
--- a/tempest/api/volume/test_volumes_get.py
+++ b/tempest/api/volume/test_volumes_get.py
@@ -37,11 +37,9 @@
kwargs['name'] = v_name
kwargs['metadata'] = metadata
volume = self.volumes_client.create_volume(**kwargs)['volume']
- self.assertIn('id', volume)
self.addCleanup(self.delete_volume, self.volumes_client, volume['id'])
waiters.wait_for_volume_resource_status(self.volumes_client,
volume['id'], 'available')
- self.assertIn('name', volume)
self.assertEqual(volume['name'], v_name,
"The created volume name is not equal "
"to the requested name")
@@ -101,7 +99,6 @@
'availability_zone': volume['availability_zone'],
'size': CONF.volume.volume_size}
new_volume = self.volumes_client.create_volume(**params)['volume']
- self.assertIn('id', new_volume)
self.addCleanup(self.delete_volume, self.volumes_client,
new_volume['id'])
waiters.wait_for_volume_resource_status(self.volumes_client,
@@ -153,7 +150,5 @@
@decorators.idempotent_id('c4f2431e-4920-4736-9e00-4040386b6feb')
def test_show_volume_summary(self):
"""Test showing volume summary"""
- volume_summary = \
- self.volumes_client.show_volume_summary()['volume-summary']
- for key in ['total_size', 'total_count']:
- self.assertIn(key, volume_summary)
+ # check response schema
+ self.volumes_client.show_volume_summary()
diff --git a/tempest/cmd/subunit_describe_calls.py b/tempest/cmd/subunit_describe_calls.py
index e029538..172fbaa 100644
--- a/tempest/cmd/subunit_describe_calls.py
+++ b/tempest/cmd/subunit_describe_calls.py
@@ -30,6 +30,8 @@
* ``--ports, -p``: (Optional) The path to a JSON file describing the ports
being used by different services
* ``--verbose, -v``: (Optional) Print Request and Response Headers and Body
+ data to stdout in the non cliff deprecated CLI
+* ``--all-stdout, -a``: (Optional) Print Request and Response Headers and Body
data to stdout
@@ -278,7 +280,7 @@
return url_parser
-def output(url_parser, output_file, verbose):
+def output(url_parser, output_file, all_stdout):
if output_file is not None:
with open(output_file, "w") as outfile:
outfile.write(json.dumps(url_parser.test_logs))
@@ -294,7 +296,7 @@
sys.stdout.write('\t- {0} {1} request for {2} to {3}\n'.format(
item.get('status_code'), item.get('verb'),
item.get('service'), item.get('url')))
- if verbose:
+ if all_stdout:
sys.stdout.write('\t\t- request headers: {0}\n'.format(
item.get('request_headers')))
sys.stdout.write('\t\t- request body: {0}\n'.format(
@@ -313,7 +315,7 @@
"please use: 'tempest subunit-describe-calls'")
cl_args = ArgumentParser().parse_args()
parser = parse(cl_args.subunit, cl_args.non_subunit_name, cl_args.ports)
- output(parser, cl_args.output_file, cl_args.verbose)
+ output(parser, cl_args.output_file, cl_args.all_stdout)
def _parser_add_args(parser):
@@ -339,9 +341,23 @@
help="A JSON file describing the ports for each service."
)
- parser.add_argument(
- "-v", "--verbose", action='store_true', default=False,
- help="Add Request and Response header and body data to stdout."
+ group = parser.add_mutually_exclusive_group()
+ # the -v and --verbose command are for the old subunit-describe-calls
+ # main() CLI interface. It does not work with the new
+ # tempest subunit-describe-callss CLI. So when the main CLI approach is
+ # deleted this argument is not needed.
+ group.add_argument(
+ "-v", "--verbose", action='store_true', dest='all_stdout',
+ help='Add Request and Response header and body data to stdout print.'
+ ' NOTE: This argument deprecated and does not work with'
+ ' tempest subunit-describe-calls CLI.'
+ ' Use new option: "-a", "--all-stdout"'
+ )
+ group.add_argument(
+ "-a", "--all-stdout", action='store_true',
+ help="Add Request and Response header and body data to stdout print."
+ " Note: this argument work with the subunit-describe-calls and"
+ " tempest subunit-describe-calls CLI commands."
)
diff --git a/tempest/config.py b/tempest/config.py
index eca2023..989b8ef 100644
--- a/tempest/config.py
+++ b/tempest/config.py
@@ -481,6 +481,12 @@
'MIN_LIBVIRT_VERSION is >= 1.2.17 on all '
'branches from stable/rocky and will be '
'removed in a future release.'),
+ cfg.BoolOpt('can_migrate_between_any_hosts',
+ default=True,
+ help="Does the test environment support migrating between "
+ "any hosts? In environments with non-homogeneous compute "
+ "nodes you can set this to False so that it will select "
+ "destination host for migrating automatically"),
cfg.BoolOpt('vnc_console',
default=False,
help='Enable VNC console. This configuration value should '
diff --git a/tempest/lib/api_schema/response/compute/v2_1/parameter_types.py b/tempest/lib/api_schema/response/compute/v2_1/parameter_types.py
index 28ed816..8aed37d 100644
--- a/tempest/lib/api_schema/response/compute/v2_1/parameter_types.py
+++ b/tempest/lib/api_schema/response/compute/v2_1/parameter_types.py
@@ -120,3 +120,10 @@
# 7: SUSPENDED
'enum': [0, 1, 3, 4, 6, 7]
}
+
+uuid_or_null = {
+ 'anyOf': [
+ {'type': 'string', 'format': 'uuid'},
+ {'type': 'null'}
+ ]
+}
diff --git a/tempest/lib/api_schema/response/volume/volumes.py b/tempest/lib/api_schema/response/volume/volumes.py
new file mode 100644
index 0000000..ffcf488
--- /dev/null
+++ b/tempest/lib/api_schema/response/volume/volumes.py
@@ -0,0 +1,368 @@
+# Copyright 2018 ZTE Corporation. All rights reserved.
+#
+# Licensed under the Apache License, Version 2.0 (the "License"); you may
+# not use this file except in compliance with the License. You may obtain
+# a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
+# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
+# License for the specific language governing permissions and limitations
+# under the License.
+
+import copy
+
+from tempest.lib.api_schema.response.compute.v2_1 import parameter_types
+
+attachments = {
+ 'type': 'array',
+ 'items': {
+ 'type': 'object',
+ 'properties': {
+ 'server_id': {'type': 'string', 'format': 'uuid'},
+ 'attachment_id': {'type': 'string', 'format': 'uuid'},
+ 'attached_at': parameter_types.date_time_or_null,
+ 'host_name': {'type': ['string', 'null']},
+ 'volume_id': {'type': 'string', 'format': 'uuid'},
+ 'device': {'type': ['string', 'null']},
+ 'id': {'type': 'string', 'format': 'uuid'}
+ },
+ 'additionalProperties': False,
+ 'required': ['server_id', 'attachment_id', 'host_name',
+ 'volume_id', 'device', 'id']
+ }
+}
+
+common_show_volume = {
+ 'type': 'object',
+ 'properties': {
+ 'migration_status': {'type': ['string', 'null']},
+ 'attachments': attachments,
+ 'links': parameter_types.links,
+ 'availability_zone': {'type': ['string', 'null']},
+ 'os-vol-host-attr:host': {
+ 'type': ['string', 'null'], 'pattern': '.+@.+#.+'},
+ 'encrypted': {'type': 'boolean'},
+ 'updated_at': parameter_types.date_time_or_null,
+ 'replication_status': {'type': ['string', 'null']},
+ 'snapshot_id': parameter_types.uuid_or_null,
+ 'id': {'type': 'string', 'format': 'uuid'},
+ 'size': {'type': 'integer'},
+ 'user_id': {'type': 'string', 'format': 'uuid'},
+ 'os-vol-tenant-attr:tenant_id': {'type': 'string',
+ 'format': 'uuid'},
+ 'os-vol-mig-status-attr:migstat': {'type': ['string', 'null']},
+ 'metadata': {'type': 'object'},
+ 'status': {'type': 'string'},
+ 'volume_image_metadata': {'type': ['object', 'null']},
+ 'description': {'type': ['string', 'null']},
+ 'multiattach': {'type': 'boolean'},
+ 'source_volid': parameter_types.uuid_or_null,
+ 'consistencygroup_id': parameter_types.uuid_or_null,
+ 'os-vol-mig-status-attr:name_id': parameter_types.uuid_or_null,
+ 'name': {'type': ['string', 'null']},
+ 'bootable': {'type': 'string'},
+ 'created_at': parameter_types.date_time,
+ 'volume_type': {'type': ['string', 'null']},
+ # TODO(zhufl): group_id is added in 3.13, we should move it to the
+ # 3.13 schema file when microversion is supported in volume interfaces
+ 'group_id': parameter_types.uuid_or_null,
+ # TODO(zhufl): provider_id is added in 3.21, we should move it to the
+ # 3.21 schema file when microversion is supported in volume interfaces
+ 'provider_id': parameter_types.uuid_or_null,
+ # TODO(zhufl): service_uuid and shared_targets are added in 3.48,
+ # we should move them to the 3.48 schema file when microversion
+ # is supported in volume interfaces.
+ 'service_uuid': parameter_types.uuid_or_null,
+ 'shared_targets': {'type': 'boolean'}
+ },
+ 'additionalProperties': False,
+ 'required': ['attachments', 'links', 'encrypted',
+ 'updated_at', 'replication_status', 'id',
+ 'size', 'user_id', 'availability_zone',
+ 'metadata', 'status', 'description',
+ 'multiattach', 'consistencygroup_id',
+ 'name', 'bootable', 'created_at',
+ 'volume_type', 'snapshot_id', 'source_volid']
+}
+
+list_volumes_no_detail = {
+ 'status_code': [200],
+ 'response_body': {
+ 'type': 'object',
+ 'properties': {
+ 'volumes': {
+ 'type': 'array',
+ 'items': {
+ 'type': 'object',
+ 'properties': {
+ 'links': parameter_types.links,
+ 'id': {'type': 'string', 'format': 'uuid'},
+ 'name': {'type': ['string', 'null']},
+ # TODO(zhufl): count is added in 3.45, we should move
+ # it to the 3.45 schema file when microversion is
+ # supported in volume interfaces
+ # 'count': {'type': 'integer'}
+ },
+ 'additionalProperties': False,
+ 'required': ['links', 'id', 'name']
+ }
+ },
+ 'volumes_links': parameter_types.links
+ },
+ 'additionalProperties': False,
+ 'required': ['volumes']
+ }
+}
+
+show_volume = {
+ 'status_code': [200],
+ 'response_body': {
+ 'type': 'object',
+ 'properties': {
+ 'volume': common_show_volume
+ },
+ 'additionalProperties': False,
+ 'required': ['volume']
+ }
+}
+
+list_volumes_detail = copy.deepcopy(common_show_volume)
+# TODO(zhufl): count is added in 3.45, we should move it to the 3.45 schema
+# file when microversion is supported in volume interfaces
+# list_volumes_detail['properties'].update({'count': {'type': 'integer'}})
+list_volumes_with_detail = {
+ 'status_code': [200],
+ 'response_body': {
+ 'type': 'object',
+ 'properties': {
+ 'volumes': {
+ 'type': 'array',
+ 'items': list_volumes_detail
+ },
+ 'volumes_links': parameter_types.links
+ },
+ 'additionalProperties': False,
+ 'required': ['volumes']
+ }
+}
+
+create_volume = {
+ 'status_code': [202],
+ 'response_body': {
+ 'type': 'object',
+ 'properties': {
+ 'volume': {
+ 'type': 'object',
+ 'properties': {
+ 'migration_status': {'type': ['string', 'null']},
+ 'attachments': attachments,
+ 'links': parameter_types.links,
+ 'availability_zone': {'type': ['string', 'null']},
+ 'encrypted': {'type': 'boolean'},
+ 'updated_at': parameter_types.date_time_or_null,
+ 'replication_status': {'type': ['string', 'null']},
+ 'snapshot_id': parameter_types.uuid_or_null,
+ 'id': {'type': 'string', 'format': 'uuid'},
+ 'size': {'type': 'integer'},
+ 'user_id': {'type': 'string', 'format': 'uuid'},
+ 'metadata': {'type': 'object'},
+ 'status': {'type': 'string'},
+ 'description': {'type': ['string', 'null']},
+ 'multiattach': {'type': 'boolean'},
+ 'source_volid': parameter_types.uuid_or_null,
+ 'consistencygroup_id': parameter_types.uuid_or_null,
+ 'name': {'type': ['string', 'null']},
+ 'bootable': {'type': 'string'},
+ 'created_at': parameter_types.date_time,
+ 'volume_type': {'type': ['string', 'null']},
+ # TODO(zhufl): group_id is added in 3.13, we should move
+ # it to the 3.13 schema file when microversion is
+ # supported in volume interfaces.
+ 'group_id': parameter_types.uuid_or_null,
+ # TODO(zhufl): provider_id is added in 3.21, we should
+ # move it to the 3.21 schema file when microversion is
+ # supported in volume interfaces
+ 'provider_id': parameter_types.uuid_or_null,
+ # TODO(zhufl): service_uuid and shared_targets are added
+ # in 3.48, we should move them to the 3.48 schema file
+ # when microversion is supported in volume interfaces.
+ 'service_uuid': parameter_types.uuid_or_null,
+ 'shared_targets': {'type': 'boolean'}
+ },
+ 'additionalProperties': False,
+ 'required': ['attachments', 'links', 'encrypted',
+ 'updated_at', 'replication_status', 'id',
+ 'size', 'user_id', 'availability_zone',
+ 'metadata', 'status', 'description',
+ 'multiattach', 'consistencygroup_id',
+ 'name', 'bootable', 'created_at',
+ 'volume_type', 'snapshot_id', 'source_volid']
+ }
+ },
+ 'additionalProperties': False,
+ 'required': ['volume']
+ }
+}
+
+update_volume = copy.deepcopy(create_volume)
+update_volume.update({'status_code': [200]})
+
+delete_volume = {'status_code': [202]}
+
+show_volume_summary = {
+ 'status_code': [200],
+ 'response_body': {
+ 'type': 'object',
+ 'properties': {
+ 'volume-summary': {
+ 'type': 'object',
+ 'properties': {
+ 'total_size': {'type': 'integer'},
+ 'total_count': {'type': 'integer'},
+ # TODO(zhufl): metadata is added in 3.36, we should move
+ # it to the 3.36 schema file when microversion is
+ # supported in volume interfaces
+ 'metadata': {'type': 'object'},
+ },
+ 'additionalProperties': False,
+ 'required': ['total_size', 'total_count']
+ }
+ },
+ 'additionalProperties': False,
+ 'required': ['volume-summary']
+ }
+}
+
+# TODO(zhufl): This is under discussion, so will be merged in a seperate patch.
+# https://bugs.launchpad.net/cinder/+bug/1880566
+# upload_volume = {
+# 'status_code': [202],
+# 'response_body': {
+# 'type': 'object',
+# 'properties': {
+# 'os-volume_upload_image': {
+# 'type': 'object',
+# 'properties': {
+# 'status': {'type': 'string'},
+# 'image_name': {'type': 'string'},
+# 'disk_format': {'type': 'string'},
+# 'container_format': {'type': 'string'},
+# 'is_public': {'type': 'boolean'},
+# 'visibility': {'type': 'string'},
+# 'protected': {'type': 'boolean'},
+# 'updated_at': parameter_types.date_time_or_null,
+# 'image_id': {'type': 'string', 'format': 'uuid'},
+# 'display_description': {'type': ['string', 'null']},
+# 'id': {'type': 'string', 'format': 'uuid'},
+# 'size': {'type': 'integer'},
+# 'volume_type': {
+# 'type': ['object', 'null'],
+# 'properties': {
+# 'created_at': parameter_types.date_time,
+# 'deleted': {'type': 'boolean'},
+# 'deleted_at': parameter_types.date_time_or_null,
+# 'description': {'type': ['string', 'null']},
+# 'extra_specs': {
+# 'type': 'object',
+# 'patternProperties': {
+# '^.+$': {'type': 'string'}
+# }
+# },
+# 'id': {'type': 'string', 'format': 'uuid'},
+# 'is_public': {'type': 'boolean'},
+# 'name': {'type': ['string', 'null']},
+# 'qos_specs_id': parameter_types.uuid_or_null,
+# 'updated_at': parameter_types.date_time_or_null
+# },
+# }
+# },
+# 'additionalProperties': False,
+# 'required': ['status', 'image_name', 'updated_at',
+# 'image_id',
+# 'display_description', 'id', 'size',
+# 'volume_type', 'disk_format',
+# 'container_format']
+# }
+# },
+# 'additionalProperties': False,
+# 'required': ['os-volume_upload_image']
+# }
+# }
+
+attach_volume = {'status_code': [202]}
+set_bootable_volume = {'status_code': [200]}
+detach_volume = {'status_code': [202]}
+reserve_volume = {'status_code': [202]}
+unreserve_volume = {'status_code': [202]}
+extend_volume = {'status_code': [202]}
+reset_volume_status = {'status_code': [202]}
+update_volume_readonly = {'status_code': [202]}
+force_delete_volume = {'status_code': [202]}
+retype_volume = {'status_code': [202]}
+force_detach_volume = {'status_code': [202]}
+
+create_volume_metadata = {
+ 'status_code': [200],
+ 'response_body': {
+ 'type': 'object',
+ 'properties': {
+ 'metadata': {'type': 'object'},
+ },
+ 'additionalProperties': False,
+ 'required': ['metadata']
+ }
+}
+
+show_volume_metadata = {
+ 'status_code': [200],
+ 'response_body': {
+ 'type': 'object',
+ 'properties': {
+ 'metadata': {'type': 'object'},
+ },
+ 'additionalProperties': False,
+ 'required': ['metadata']
+ }
+}
+update_volume_metadata = copy.deepcopy(show_volume_metadata)
+
+show_volume_metadata_item = {
+ 'status_code': [200],
+ 'response_body': {
+ 'type': 'object',
+ 'properties': {
+ 'meta': {'type': 'object'},
+ },
+ 'additionalProperties': False,
+ 'required': ['meta']
+ }
+}
+update_volume_metadata_item = copy.deepcopy(show_volume_metadata_item)
+delete_volume_metadata_item = {'status_code': [200]}
+
+update_volume_image_metadata = {
+ 'status_code': [200],
+ 'response_body': {
+ 'type': 'object',
+ 'properties': {'metadata': {'type': 'object'}},
+ 'additionalProperties': False,
+ 'required': ['metadata']
+ }
+}
+delete_volume_image_metadata = {'status_code': [200]}
+show_volume_image_metadata = {
+ 'status_code': [200],
+ 'response_body': {
+ 'type': 'object',
+ 'properties': {
+ 'metadata': {'type': 'object'},
+ },
+ 'additionalProperties': False,
+ 'required': ['metadata']
+ }
+}
+
+unmanage_volume = {'status_code': [202]}
diff --git a/tempest/lib/common/utils/linux/remote_client.py b/tempest/lib/common/utils/linux/remote_client.py
index 8ac1d38..71fed02 100644
--- a/tempest/lib/common/utils/linux/remote_client.py
+++ b/tempest/lib/common/utils/linux/remote_client.py
@@ -11,7 +11,6 @@
# under the License.
import functools
-import re
import sys
import netaddr
@@ -134,9 +133,8 @@
This method will not unmount the config drive, so unmount_config_drive
must be used for cleanup.
"""
- cmd_blkid = 'blkid | grep -i config-2'
- result = self.exec_command(cmd_blkid)
- dev_name = re.match('([^:]+)', result).group()
+ cmd_blkid = 'blkid -L config-2 -o device'
+ dev_name = self.exec_command(cmd_blkid).strip()
try:
self.exec_command('sudo mount %s /mnt' % dev_name)
diff --git a/tempest/lib/services/volume/v3/volumes_client.py b/tempest/lib/services/volume/v3/volumes_client.py
index 9c6729b..b8535d8 100644
--- a/tempest/lib/services/volume/v3/volumes_client.py
+++ b/tempest/lib/services/volume/v3/volumes_client.py
@@ -17,6 +17,7 @@
import six
from six.moves.urllib import parse as urllib
+from tempest.lib.api_schema.response.volume import volumes as schema
from tempest.lib.common import rest_client
from tempest.lib import exceptions as lib_exc
from tempest.lib.services.volume import base_client
@@ -55,14 +56,16 @@
https://docs.openstack.org/api-ref/block-storage/v3/index.html#list-accessible-volumes
"""
url = 'volumes'
+ list_schema = schema.list_volumes_no_detail
if detail:
+ list_schema = schema.list_volumes_with_detail
url += '/detail'
if params:
url += '?%s' % self._prepare_params(params)
resp, body = self.get(url)
body = json.loads(body)
- self.expected_success(200, resp.status)
+ self.validate_response(list_schema, resp, body)
return rest_client.ResponseBody(resp, body)
def migrate_volume(self, volume_id, **kwargs):
@@ -83,7 +86,7 @@
url = "volumes/%s" % volume_id
resp, body = self.get(url)
body = json.loads(body)
- self.expected_success(200, resp.status)
+ self.validate_response(schema.show_volume, resp, body)
return rest_client.ResponseBody(resp, body)
def create_volume(self, **kwargs):
@@ -96,7 +99,7 @@
post_body = json.dumps({'volume': kwargs})
resp, body = self.post('volumes', post_body)
body = json.loads(body)
- self.expected_success(202, resp.status)
+ self.validate_response(schema.create_volume, resp, body)
return rest_client.ResponseBody(resp, body)
def update_volume(self, volume_id, **kwargs):
@@ -109,7 +112,7 @@
put_body = json.dumps({'volume': kwargs})
resp, body = self.put('volumes/%s' % volume_id, put_body)
body = json.loads(body)
- self.expected_success(200, resp.status)
+ self.validate_response(schema.update_volume, resp, body)
return rest_client.ResponseBody(resp, body)
def delete_volume(self, volume_id, **params):
@@ -123,7 +126,7 @@
if params:
url += '?%s' % urllib.urlencode(params)
resp, body = self.delete(url)
- self.expected_success(202, resp.status)
+ self.validate_response(schema.delete_volume, resp, body)
return rest_client.ResponseBody(resp, body)
def show_volume_summary(self, **params):
@@ -138,7 +141,7 @@
url += '?%s' % urllib.urlencode(params)
resp, body = self.get(url)
body = json.loads(body)
- self.expected_success(200, resp.status)
+ self.validate_response(schema.show_volume_summary, resp, body)
return rest_client.ResponseBody(resp, body)
def upload_volume(self, volume_id, **kwargs):
@@ -152,6 +155,10 @@
url = 'volumes/%s/action' % (volume_id)
resp, body = self.post(url, post_body)
body = json.loads(body)
+ # TODO(zhufl): This is under discussion, so will be merged
+ # in a seperate patch.
+ # https://bugs.launchpad.net/cinder/+bug/1880566
+ # self.validate_response(schema.upload_volume, resp, body)
self.expected_success(202, resp.status)
return rest_client.ResponseBody(resp, body)
@@ -165,7 +172,7 @@
post_body = json.dumps({'os-attach': kwargs})
url = 'volumes/%s/action' % (volume_id)
resp, body = self.post(url, post_body)
- self.expected_success(202, resp.status)
+ self.validate_response(schema.attach_volume, resp, body)
return rest_client.ResponseBody(resp, body)
def set_bootable_volume(self, volume_id, **kwargs):
@@ -178,7 +185,7 @@
post_body = json.dumps({'os-set_bootable': kwargs})
url = 'volumes/%s/action' % (volume_id)
resp, body = self.post(url, post_body)
- self.expected_success(200, resp.status)
+ self.validate_response(schema.set_bootable_volume, resp, body)
return rest_client.ResponseBody(resp, body)
def detach_volume(self, volume_id):
@@ -186,7 +193,7 @@
post_body = json.dumps({'os-detach': {}})
url = 'volumes/%s/action' % (volume_id)
resp, body = self.post(url, post_body)
- self.expected_success(202, resp.status)
+ self.validate_response(schema.detach_volume, resp, body)
return rest_client.ResponseBody(resp, body)
def reserve_volume(self, volume_id):
@@ -194,7 +201,7 @@
post_body = json.dumps({'os-reserve': {}})
url = 'volumes/%s/action' % (volume_id)
resp, body = self.post(url, post_body)
- self.expected_success(202, resp.status)
+ self.validate_response(schema.reserve_volume, resp, body)
return rest_client.ResponseBody(resp, body)
def unreserve_volume(self, volume_id):
@@ -202,7 +209,7 @@
post_body = json.dumps({'os-unreserve': {}})
url = 'volumes/%s/action' % (volume_id)
resp, body = self.post(url, post_body)
- self.expected_success(202, resp.status)
+ self.validate_response(schema.unreserve_volume, resp, body)
return rest_client.ResponseBody(resp, body)
def is_resource_deleted(self, id):
@@ -237,7 +244,7 @@
post_body = json.dumps({'os-extend': kwargs})
url = 'volumes/%s/action' % (volume_id)
resp, body = self.post(url, post_body)
- self.expected_success(202, resp.status)
+ self.validate_response(schema.extend_volume, resp, body)
return rest_client.ResponseBody(resp, body)
def reset_volume_status(self, volume_id, **kwargs):
@@ -249,7 +256,7 @@
"""
post_body = json.dumps({'os-reset_status': kwargs})
resp, body = self.post('volumes/%s/action' % volume_id, post_body)
- self.expected_success(202, resp.status)
+ self.validate_response(schema.reset_volume_status, resp, body)
return rest_client.ResponseBody(resp, body)
def update_volume_readonly(self, volume_id, **kwargs):
@@ -262,14 +269,14 @@
post_body = json.dumps({'os-update_readonly_flag': kwargs})
url = 'volumes/%s/action' % (volume_id)
resp, body = self.post(url, post_body)
- self.expected_success(202, resp.status)
+ self.validate_response(schema.update_volume_readonly, resp, body)
return rest_client.ResponseBody(resp, body)
def force_delete_volume(self, volume_id):
"""Force Delete Volume."""
post_body = json.dumps({'os-force_delete': {}})
resp, body = self.post('volumes/%s/action' % volume_id, post_body)
- self.expected_success(202, resp.status)
+ self.validate_response(schema.force_delete_volume, resp, body)
return rest_client.ResponseBody(resp, body)
def create_volume_metadata(self, volume_id, metadata):
@@ -283,7 +290,7 @@
url = "volumes/%s/metadata" % volume_id
resp, body = self.post(url, put_body)
body = json.loads(body)
- self.expected_success(200, resp.status)
+ self.validate_response(schema.create_volume_metadata, resp, body)
return rest_client.ResponseBody(resp, body)
def show_volume_metadata(self, volume_id):
@@ -291,7 +298,7 @@
url = "volumes/%s/metadata" % volume_id
resp, body = self.get(url)
body = json.loads(body)
- self.expected_success(200, resp.status)
+ self.validate_response(schema.show_volume_metadata, resp, body)
return rest_client.ResponseBody(resp, body)
def update_volume_metadata(self, volume_id, metadata):
@@ -305,7 +312,7 @@
url = "volumes/%s/metadata" % volume_id
resp, body = self.put(url, put_body)
body = json.loads(body)
- self.expected_success(200, resp.status)
+ self.validate_response(schema.update_volume_metadata, resp, body)
return rest_client.ResponseBody(resp, body)
def show_volume_metadata_item(self, volume_id, id):
@@ -313,7 +320,7 @@
url = "volumes/%s/metadata/%s" % (volume_id, id)
resp, body = self.get(url)
body = json.loads(body)
- self.expected_success(200, resp.status)
+ self.validate_response(schema.show_volume_metadata_item, resp, body)
return rest_client.ResponseBody(resp, body)
def update_volume_metadata_item(self, volume_id, id, meta_item):
@@ -322,14 +329,14 @@
url = "volumes/%s/metadata/%s" % (volume_id, id)
resp, body = self.put(url, put_body)
body = json.loads(body)
- self.expected_success(200, resp.status)
+ self.validate_response(schema.update_volume_metadata_item, resp, body)
return rest_client.ResponseBody(resp, body)
def delete_volume_metadata_item(self, volume_id, id):
"""Delete metadata item for the volume."""
url = "volumes/%s/metadata/%s" % (volume_id, id)
resp, body = self.delete(url)
- self.expected_success(200, resp.status)
+ self.validate_response(schema.delete_volume_metadata_item, resp, body)
return rest_client.ResponseBody(resp, body)
def retype_volume(self, volume_id, **kwargs):
@@ -341,7 +348,7 @@
"""
post_body = json.dumps({'os-retype': kwargs})
resp, body = self.post('volumes/%s/action' % volume_id, post_body)
- self.expected_success(202, resp.status)
+ self.validate_response(schema.retype_volume, resp, body)
return rest_client.ResponseBody(resp, body)
def force_detach_volume(self, volume_id, **kwargs):
@@ -354,7 +361,7 @@
post_body = json.dumps({'os-force_detach': kwargs})
url = 'volumes/%s/action' % volume_id
resp, body = self.post(url, post_body)
- self.expected_success(202, resp.status)
+ self.validate_response(schema.force_detach_volume, resp, body)
return rest_client.ResponseBody(resp, body)
def update_volume_image_metadata(self, volume_id, **kwargs):
@@ -368,7 +375,7 @@
url = "volumes/%s/action" % (volume_id)
resp, body = self.post(url, post_body)
body = json.loads(body)
- self.expected_success(200, resp.status)
+ self.validate_response(schema.update_volume_image_metadata, resp, body)
return rest_client.ResponseBody(resp, body)
def delete_volume_image_metadata(self, volume_id, key_name):
@@ -376,7 +383,7 @@
post_body = json.dumps({'os-unset_image_metadata': {'key': key_name}})
url = "volumes/%s/action" % (volume_id)
resp, body = self.post(url, post_body)
- self.expected_success(200, resp.status)
+ self.validate_response(schema.delete_volume_image_metadata, resp, body)
return rest_client.ResponseBody(resp, body)
def show_volume_image_metadata(self, volume_id):
@@ -385,7 +392,7 @@
url = "volumes/%s/action" % volume_id
resp, body = self.post(url, post_body)
body = json.loads(body)
- self.expected_success(200, resp.status)
+ self.validate_response(schema.show_volume_image_metadata, resp, body)
return rest_client.ResponseBody(resp, body)
def unmanage_volume(self, volume_id):
@@ -397,5 +404,5 @@
"""
post_body = json.dumps({'os-unmanage': {}})
resp, body = self.post('volumes/%s/action' % volume_id, post_body)
- self.expected_success(202, resp.status)
+ self.validate_response(schema.unmanage_volume, resp, body)
return rest_client.ResponseBody(resp, body)
diff --git a/tempest/manager.py b/tempest/manager.py
index e3174d4..b485ef2 100644
--- a/tempest/manager.py
+++ b/tempest/manager.py
@@ -33,7 +33,7 @@
def __init__(self, credentials, scope='project'):
msg = ("tempest.manager.Manager is not a stable interface and as such "
- "it should not imported directly. It will be removed as "
+ "it should not be imported directly. It will be removed as "
"soon as the client manager becomes available in tempest.lib.")
LOG.warning(msg)
dscv = CONF.identity.disable_ssl_certificate_validation
diff --git a/tempest/scenario/manager.py b/tempest/scenario/manager.py
index a80b77d..add4c37 100644
--- a/tempest/scenario/manager.py
+++ b/tempest/scenario/manager.py
@@ -91,10 +91,32 @@
volume_microversion=self.volume_request_microversion,
placement_microversion=self.placement_request_microversion))
+ def setup_compute_client(cls):
+ """Compute and Compute security groups client"""
+ cls.compute_images_client = cls.os_primary.compute_images_client
+ cls.keypairs_client = cls.os_primary.keypairs_client
+ cls.compute_security_groups_client = (
+ cls.os_primary.compute_security_groups_client)
+ cls.compute_security_group_rules_client = (
+ cls.os_primary.compute_security_group_rules_client)
+ cls.servers_client = cls.os_primary.servers_client
+ cls.interface_client = cls.os_primary.interfaces_client
+
+ def setup_network_client(cls):
+ """Neutron network client"""
+ cls.networks_client = cls.os_primary.networks_client
+ cls.ports_client = cls.os_primary.ports_client
+ cls.routers_client = cls.os_primary.routers_client
+ cls.subnets_client = cls.os_primary.subnets_client
+ cls.floating_ips_client = cls.os_primary.floating_ips_client
+ cls.security_groups_client = cls.os_primary.security_groups_client
+ cls.security_group_rules_client = (
+ cls.os_primary.security_group_rules_client)
+
@classmethod
def setup_clients(cls):
+ """This setup the service clients for the tests"""
super(ScenarioTest, cls).setup_clients()
- # Clients (in alphabetical order)
cls.flavors_client = cls.os_primary.flavors_client
cls.compute_floating_ips_client = (
cls.os_primary.compute_floating_ips_client)
@@ -108,37 +130,20 @@
raise lib_exc.InvalidConfiguration(
'Either api_v1 or api_v2 must be True in '
'[image-feature-enabled].')
- # Compute image client
- cls.compute_images_client = cls.os_primary.compute_images_client
- cls.keypairs_client = cls.os_primary.keypairs_client
- # Nova security groups client
- cls.compute_security_groups_client = (
- cls.os_primary.compute_security_groups_client)
- cls.compute_security_group_rules_client = (
- cls.os_primary.compute_security_group_rules_client)
- cls.servers_client = cls.os_primary.servers_client
- cls.interface_client = cls.os_primary.interfaces_client
- # Neutron network client
- cls.networks_client = cls.os_primary.networks_client
- cls.ports_client = cls.os_primary.ports_client
- cls.routers_client = cls.os_primary.routers_client
- cls.subnets_client = cls.os_primary.subnets_client
- cls.floating_ips_client = cls.os_primary.floating_ips_client
- cls.security_groups_client = cls.os_primary.security_groups_client
- cls.security_group_rules_client = (
- cls.os_primary.security_group_rules_client)
- # Use the latest available volume clients
+
+ cls.setup_compute_client(cls)
+ cls.setup_network_client(cls)
if CONF.service_available.cinder:
cls.volumes_client = cls.os_primary.volumes_client_latest
cls.snapshots_client = cls.os_primary.snapshots_client_latest
cls.backups_client = cls.os_primary.backups_client_latest
# ## Test functions library
- #
# The create_[resource] functions only return body and discard the
# resp part which is not used in scenario tests
def create_port(self, network_id, client=None, **kwargs):
+ """Creates port"""
if not client:
client = self.ports_client
name = data_utils.rand_name(self.__class__.__name__)
@@ -156,6 +161,13 @@
return port
def create_keypair(self, client=None):
+ """Creates keypair
+
+ Keypair is a public key of OpenSSH key pair used for accessing
+ and create servers
+ Keypair can also be created by a private key for the same purpose
+ Here, the keys are randomly generated[public/private]
+ """
if not client:
client = self.keypairs_client
name = data_utils.rand_name(self.__class__.__name__)
@@ -295,6 +307,13 @@
def create_volume(self, size=None, name=None, snapshot_id=None,
imageRef=None, volume_type=None):
+ """Creates volume
+
+ This wrapper utility creates volume and waits for volume to be
+ in 'available' state.
+ This method returns the volume's full representation by GET request.
+ """
+
if size is None:
size = CONF.volume.volume_size
if imageRef:
@@ -334,6 +353,11 @@
def create_backup(self, volume_id, name=None, description=None,
force=False, snapshot_id=None, incremental=False,
container=None):
+ """Creates backup
+
+ This wrapper utility creates backup and waits for backup to be
+ in 'available' state.
+ """
name = name or data_utils.rand_name(
self.__class__.__name__ + "-backup")
@@ -351,6 +375,12 @@
return backup
def restore_backup(self, backup_id):
+ """Restore backup
+
+ This wrapper utility restores backup and waits for backup to be
+ in 'available' state.
+ """
+
restore = self.backups_client.restore_backup(backup_id)['restore']
self.addCleanup(self.volumes_client.delete_volume,
restore['volume_id'])
@@ -362,8 +392,31 @@
self.assertEqual(backup_id, restore['backup_id'])
return restore
+ def rebuild_server(self, server_id, image=None,
+ preserve_ephemeral=False, wait=True,
+ rebuild_kwargs=None):
+ if image is None:
+ image = CONF.compute.image_ref
+ rebuild_kwargs = rebuild_kwargs or {}
+ LOG.debug("Rebuilding server (id: %s, image: %s, preserve eph: %s)",
+ server_id, image, preserve_ephemeral)
+ self.servers_client.rebuild_server(
+ server_id=server_id,
+ image_ref=image,
+ preserve_ephemeral=preserve_ephemeral,
+ **rebuild_kwargs)
+ if wait:
+ waiters.wait_for_server_status(self.servers_client,
+ server_id, 'ACTIVE')
+
def create_volume_snapshot(self, volume_id, name=None, description=None,
metadata=None, force=False):
+ """Creates volume
+
+ This wrapper utility creates volume snapshot and waits for backup
+ to be in 'available' state.
+ """
+
name = name or data_utils.rand_name(
self.__class__.__name__ + '-snapshot')
snapshot = self.snapshots_client.create_snapshot(
@@ -372,6 +425,7 @@
display_name=name,
description=description,
metadata=metadata)['snapshot']
+
self.addCleanup(self.snapshots_client.wait_for_resource_deletion,
snapshot['id'])
self.addCleanup(self.snapshots_client.delete_snapshot, snapshot['id'])
@@ -401,6 +455,23 @@
admin_volume_type_client.delete_volume_type(volume_type['id'])
def create_volume_type(self, client=None, name=None, backend_name=None):
+ """Creates volume type
+
+ In a multiple-storage back-end configuration,
+ each back end has a name (volume_backend_name).
+ The name of the back end is declared as an extra-specification
+ of a volume type (such as, volume_backend_name=LVM).
+ When a volume is created, the scheduler chooses an
+ appropriate back end to handle the request, according
+ to the volume type specified by the user.
+ The scheduler uses volume types to explicitly create volumes on
+ specific back ends.
+
+ Before using volume type, a volume type has to be declared
+ to Block Storage. In addition to that, an extra-specification
+ has to be created to link the volume type to a back end name.
+ """
+
if not client:
client = self.os_admin.volume_types_client_latest
if not name:
@@ -416,6 +487,7 @@
volume_type = client.create_volume_type(
name=randomized_name, extra_specs=extra_specs)['volume_type']
+ self.assertIn('id', volume_type)
self.addCleanup(self._cleanup_volume_type, volume_type)
return volume_type
@@ -456,7 +528,7 @@
return rules
def _create_security_group(self):
- # Create security group
+ """Create security group and add rules to security group"""
sg_name = data_utils.rand_name(self.__class__.__name__)
sg_desc = sg_name + " description"
secgroup = self.compute_security_groups_client.create_security_group(
@@ -470,7 +542,6 @@
# Add rules to the security group
self._create_loginable_secgroup_rule(secgroup['id'])
-
return secgroup
def get_remote_client(self, ip_address, username=None, private_key=None,
@@ -502,36 +573,7 @@
linux_client.validate_authentication()
return linux_client
- def _image_create(self, name, fmt, path,
- disk_format=None, properties=None):
- if properties is None:
- properties = {}
- name = data_utils.rand_name('%s-' % name)
- params = {
- 'name': name,
- 'container_format': fmt,
- 'disk_format': disk_format or fmt,
- }
- if CONF.image_feature_enabled.api_v1:
- params['is_public'] = 'False'
- params['properties'] = properties
- params = {'headers': common_image.image_meta_to_headers(**params)}
- else:
- params['visibility'] = 'private'
- # Additional properties are flattened out in the v2 API.
- params.update(properties)
- body = self.image_client.create_image(**params)
- image = body['image'] if 'image' in body else body
- self.addCleanup(self.image_client.delete_image, image['id'])
- self.assertEqual("queued", image['status'])
- with open(path, 'rb') as image_file:
- if CONF.image_feature_enabled.api_v1:
- self.image_client.update_image(image['id'], data=image_file)
- else:
- self.image_client.store_image_file(image['id'], image_file)
- return image['id']
-
- def glance_image_create(self):
+ def image_create(self, name='scenario-img'):
img_path = CONF.scenario.img_file
if not os.path.exists(img_path):
# TODO(kopecmartin): replace LOG.warning for rasing
@@ -553,16 +595,38 @@
"properties: %s",
img_path, img_container_format, img_disk_format,
img_properties)
- image = self._image_create('scenario-img',
- img_container_format,
- img_path,
- disk_format=img_disk_format,
- properties=img_properties)
- LOG.debug("image:%s", image)
-
- return image
+ if img_properties is None:
+ img_properties = {}
+ name = data_utils.rand_name('%s-' % name)
+ params = {
+ 'name': name,
+ 'container_format': img_container_format,
+ 'disk_format': img_disk_format or img_container_format,
+ }
+ if CONF.image_feature_enabled.api_v1:
+ params['is_public'] = 'False'
+ if img_properties:
+ params['properties'] = img_properties
+ params = {'headers': common_image.image_meta_to_headers(**params)}
+ else:
+ params['visibility'] = 'private'
+ # Additional properties are flattened out in the v2 API.
+ if img_properties:
+ params.update(img_properties)
+ body = self.image_client.create_image(**params)
+ image = body['image'] if 'image' in body else body
+ self.addCleanup(self.image_client.delete_image, image['id'])
+ self.assertEqual("queued", image['status'])
+ with open(img_path, 'rb') as image_file:
+ if CONF.image_feature_enabled.api_v1:
+ self.image_client.update_image(image['id'], data=image_file)
+ else:
+ self.image_client.store_image_file(image['id'], image_file)
+ LOG.debug("image:%s", image['id'])
+ return image['id']
def _log_console_output(self, servers=None, client=None):
+ """Console log output"""
if not CONF.compute_feature_enabled.console_output:
LOG.debug('Console output not supported, cannot log')
return
@@ -581,11 +645,12 @@
"for the console log", server['id'])
def _log_net_info(self, exc):
- # network debug is called as part of ssh init
+ """network debug is called as part of ssh init"""
if not isinstance(exc, lib_exc.SSHTimeout):
LOG.debug('Network information on a devstack host')
def create_server_snapshot(self, server, name=None):
+ """Creates server snapshot"""
# Glance client
_image_client = self.image_client
# Compute client
@@ -603,7 +668,7 @@
_image_client.delete_image, image_id)
if CONF.image_feature_enabled.api_v1:
- # In glance v1 the additional properties are stored in the headers.
+ # In glance v1 the additional properties are stored in the headers
resp = _image_client.check_image(image_id)
snapshot_image = common_image.get_image_meta_from_headers(resp)
image_props = snapshot_image.get('properties', {})
@@ -633,6 +698,7 @@
return snapshot_image
def nova_volume_attach(self, server, volume_to_attach):
+ """Attach nova volume"""
volume = self.servers_client.attach_volume(
server['id'], volumeId=volume_to_attach['id'])['volumeAttachment']
self.assertEqual(volume_to_attach['id'], volume['id'])
@@ -643,12 +709,14 @@
return self.volumes_client.show_volume(volume['id'])['volume']
def nova_volume_detach(self, server, volume):
+ """Detach nova volume"""
self.servers_client.detach_volume(server['id'], volume['id'])
waiters.wait_for_volume_resource_status(self.volumes_client,
volume['id'], 'available')
def ping_ip_address(self, ip_address, should_succeed=True,
ping_timeout=None, mtu=None, server=None):
+ """ping ip address"""
timeout = ping_timeout or CONF.validation.ping_timeout
cmd = ['ping', '-c1', '-w1']
@@ -709,6 +777,7 @@
:raises: AssertError if the result of the connectivity check does
not match the value of the should_connect param
"""
+
LOG.debug('checking network connections to IP %s with user: %s',
ip_address, username)
if should_connect:
@@ -732,7 +801,7 @@
LOG.exception(extra_msg)
raise
- def create_floating_ip(self, thing, pool_name=None):
+ def create_floating_ip(self, server, pool_name=None):
"""Create a floating IP and associates to a server on Nova"""
if not pool_name:
@@ -743,11 +812,17 @@
self.compute_floating_ips_client.delete_floating_ip,
floating_ip['id'])
self.compute_floating_ips_client.associate_floating_ip_to_server(
- floating_ip['ip'], thing['id'])
+ floating_ip['ip'], server['id'])
return floating_ip
def create_timestamp(self, ip_address, dev_name=None, mount_path='/mnt',
private_key=None, server=None):
+ """Creates timestamp
+
+ This wrapper utility does ssh, creates timestamp and returns the
+ created timestamp.
+ """
+
ssh_client = self.get_remote_client(ip_address,
private_key=private_key,
server=server)
@@ -765,6 +840,11 @@
def get_timestamp(self, ip_address, dev_name=None, mount_path='/mnt',
private_key=None, server=None):
+ """Returns timestamp
+
+ This wrapper utility does ssh and returns the timestamp.
+ """
+
ssh_client = self.get_remote_client(ip_address,
private_key=private_key,
server=server)
@@ -782,6 +862,7 @@
Based on the configuration we're in, return a correct ip
address for validating that a guest is up.
"""
+
if CONF.validation.connect_method == 'floating':
# The tests calling this method don't have a floating IP
# and can't make use of the validation resources. So the
@@ -807,6 +888,8 @@
@classmethod
def get_host_for_server(cls, server_id):
+ """Gets host of server"""
+
server_details = cls.os_admin.servers_client.show_server(server_id)
return server_details['server']['OS-EXT-SRV-ATTR:host']
@@ -825,6 +908,12 @@
security_group=None,
delete_on_termination=False,
name=None):
+ """Boot instance from resource
+
+ This wrapper utility boots instance from resource with block device
+ mapping with source info passed in arguments
+ """
+
create_kwargs = dict()
if keypair:
create_kwargs['key_name'] = keypair['name']
@@ -841,6 +930,7 @@
return self.create_server(image_id='', **create_kwargs)
def create_volume_from_image(self):
+ """Create volume from image"""
img_uuid = CONF.compute.image_ref
vol_name = data_utils.rand_name(
self.__class__.__name__ + '-volume-origin')
@@ -896,8 +986,17 @@
namestart='subnet-smoke', **kwargs):
"""Create a subnet for the given network
+ This utility creates subnet for the given network
within the cidr block configured for tenant networks.
+
+ :param **kwargs:
+ See extra parameters below
+
+ :Keyword Arguments:
+
+ * *ip_version = ip version of the given network,
"""
+
if not subnets_client:
subnets_client = self.subnets_client
@@ -1000,22 +1099,23 @@
"Unable to get network by name: %s" % network_name)
return net[0]
- def create_floating_ip(self, thing, external_network_id=None,
+ def create_floating_ip(self, server, external_network_id=None,
port_id=None, client=None):
"""Create a floating IP and associates to a resource/port on Neutron"""
+
if not external_network_id:
external_network_id = CONF.network.public_network_id
if not client:
client = self.floating_ips_client
if not port_id:
- port_id, ip4 = self._get_server_port_id_and_ip4(thing)
+ port_id, ip4 = self._get_server_port_id_and_ip4(server)
else:
ip4 = None
kwargs = {
'floating_network_id': external_network_id,
'port_id': port_id,
- 'tenant_id': thing.get('project_id') or thing['tenant_id'],
+ 'tenant_id': server.get('project_id') or server['tenant_id'],
'fixed_ip_address': ip4,
}
if CONF.network.subnet_id:
@@ -1035,6 +1135,7 @@
:param status: target status
:raises: AssertionError if status doesn't match
"""
+
floatingip_id = floating_ip['id']
def refresh():
@@ -1061,6 +1162,7 @@
private_key,
should_connect=True,
servers_for_debug=None):
+ """Checks tenant network connectivity"""
if not CONF.network.project_networks_reachable:
msg = 'Tenant networks not configured to be reachable.'
LOG.info(msg)
@@ -1093,6 +1195,7 @@
:returns: True, if the connection succeeded and it was expected to
succeed. False otherwise.
"""
+
method_name = '%s_check' % protocol
connectivity_checker = getattr(source, method_name)
@@ -1156,6 +1259,7 @@
:param project_id: secgroup will be created in this project
:returns: the created security group
"""
+
if client is None:
client = self.security_groups_client
if not project_id:
@@ -1197,6 +1301,7 @@
port_range_max: 22
}
"""
+
if sec_group_rules_client is None:
sec_group_rules_client = self.security_group_rules_client
if security_groups_client is None:
@@ -1287,6 +1392,7 @@
network has, a tenant router will be created and returned that
routes traffic to the public network.
"""
+
if not client:
client = self.routers_client
if not project_id:
@@ -1327,6 +1433,7 @@
'provider:segmentation_id': '42'}
:returns: network, subnet, router
"""
+
if CONF.network.shared_physical_network:
# NOTE(Shrews): This exception is for environments where tenant
# credential isolation is available, but network separation is
@@ -1383,6 +1490,7 @@
def create_encryption_type(self, client=None, type_id=None, provider=None,
key_size=None, cipher=None,
control_location=None):
+ """Creates an encryption type for volume"""
if not client:
client = self.admin_encryption_types_client
if not type_id:
@@ -1396,6 +1504,7 @@
def create_encrypted_volume(self, encryption_provider, volume_type,
key_size=256, cipher='aes-xts-plain64',
control_location='front-end'):
+ """Creates an encrypted volume"""
volume_type = self.create_volume_type(name=volume_type)
self.create_encryption_type(type_id=volume_type['id'],
provider=encryption_provider,
@@ -1436,11 +1545,12 @@
cls.object_client = cls.os_operator.object_client
def get_swift_stat(self):
- """get swift status for our user account."""
+ """Get swift status for our user account."""
self.account_client.list_account_containers()
LOG.debug('Swift status information obtained successfully')
def create_container(self, container_name=None):
+ """Creates container"""
name = container_name or data_utils.rand_name(
'swift-scenario-container')
self.container_client.update_container(name)
@@ -1453,10 +1563,12 @@
return name
def delete_container(self, container_name):
+ """Deletes container"""
self.container_client.delete_container(container_name)
LOG.debug('Container %s deleted', container_name)
def upload_object_to_container(self, container_name, obj_name=None):
+ """Uploads object to container"""
obj_name = obj_name or data_utils.rand_name('swift-scenario-object')
obj_data = data_utils.random_bytes()
self.object_client.create_object(container_name, obj_name, obj_data)
@@ -1467,6 +1579,7 @@
return obj_name, obj_data
def delete_object(self, container_name, filename):
+ """Deletes object"""
self.object_client.delete_object(container_name, filename)
self.list_and_check_container_objects(container_name,
not_present_obj=[filename])
@@ -1474,8 +1587,13 @@
def list_and_check_container_objects(self, container_name,
present_obj=None,
not_present_obj=None):
- # List objects for a given container and assert which are present and
- # which are not.
+ """List and verify objects for a given container
+
+ This utility lists objects for a given container
+ and asserts which are present and
+ which are not
+ """
+
if present_obj is None:
present_obj = []
if not_present_obj is None:
@@ -1490,5 +1608,6 @@
self.assertNotIn(obj, object_list)
def download_and_verify(self, container_name, obj_name, expected_data):
+ """Asserts the object and expected data to verify if they are same"""
_, obj = self.object_client.get_object(container_name, obj_name)
self.assertEqual(obj, expected_data)
diff --git a/tempest/scenario/test_encrypted_cinder_volumes.py b/tempest/scenario/test_encrypted_cinder_volumes.py
index 008d1ae..fc93a5e 100644
--- a/tempest/scenario/test_encrypted_cinder_volumes.py
+++ b/tempest/scenario/test_encrypted_cinder_volumes.py
@@ -44,7 +44,7 @@
raise cls.skipException('Encrypted volume attach is not supported')
def launch_instance(self):
- image = self.glance_image_create()
+ image = self.image_create()
keypair = self.create_keypair()
return self.create_server(image_id=image, key_name=keypair['name'])
diff --git a/tempest/scenario/test_minimum_basic.py b/tempest/scenario/test_minimum_basic.py
index 4cd860d..fe42583 100644
--- a/tempest/scenario/test_minimum_basic.py
+++ b/tempest/scenario/test_minimum_basic.py
@@ -106,7 +106,7 @@
@decorators.idempotent_id('bdbb5441-9204-419d-a225-b4fdbfb1a1a8')
@utils.services('compute', 'volume', 'image', 'network')
def test_minimum_basic_scenario(self):
- image = self.glance_image_create()
+ image = self.image_create()
keypair = self.create_keypair()
server = self.create_server(image_id=image, key_name=keypair['name'])
diff --git a/tempest/scenario/test_network_v6.py b/tempest/scenario/test_network_v6.py
index 8de6614..14f24c7 100644
--- a/tempest/scenario/test_network_v6.py
+++ b/tempest/scenario/test_network_v6.py
@@ -130,7 +130,7 @@
key_name=self.keypair['name'],
security_groups=[{'name': self.sec_grp['name']}],
networks=[{'uuid': n['id']} for n in networks])
- fip = self.create_floating_ip(thing=srv)
+ fip = self.create_floating_ip(server=srv)
ips = self.define_server_ips(srv=srv)
ssh = self.get_remote_client(
ip_address=fip['floating_ip_address'],
diff --git a/tempest/tests/cmd/sample_streams/calls.subunit b/tempest/tests/cmd/subunit_describe_calls_data/calls.subunit
similarity index 100%
rename from tempest/tests/cmd/sample_streams/calls.subunit
rename to tempest/tests/cmd/subunit_describe_calls_data/calls.subunit
Binary files differ
diff --git a/tempest/tests/cmd/subunit_describe_calls_data/calls_subunit_expected.json b/tempest/tests/cmd/subunit_describe_calls_data/calls_subunit_expected.json
new file mode 100644
index 0000000..53976ee
--- /dev/null
+++ b/tempest/tests/cmd/subunit_describe_calls_data/calls_subunit_expected.json
@@ -0,0 +1,87 @@
+{"bar":[
+ {
+ "name":"AgentsAdminTestJSON:setUp",
+ "request_body":"{\"agent\": {\"url\": \"xxx://xxxx/xxx/xxx\", \"hypervisor\": \"common\", \"md5hash\": \"add6bb58e139be103324d04d82d8f545\", \"version\": \"7.0\", \"architecture\": \"tempest-x86_64-424013832\", \"os\": \"linux\"}}",
+ "request_headers":"{'Content-Type': 'application/json', 'Accept': 'application/json', 'X-Auth-Token': '<omitted>'}",
+ "response_body":"{\"agent\": {\"url\": \"xxx://xxxx/xxx/xxx\", \"hypervisor\": \"common\", \"md5hash\": \"add6bb58e139be103324d04d82d8f545\", \"version\": \"7.0\", \"architecture\": \"tempest-x86_64-424013832\", \"os\": \"linux\", \"agent_id\": 1}}",
+ "response_headers":"{'status': '200', 'content-length': '203', 'x-compute-request-id': 'req-25ddaae2-0ef1-40d1-8228-59bd64a7e75b', 'vary': 'X-OpenStack-Nova-API-Version', 'connection': 'close', 'x-openstack-nova-api-version': '2.1', 'date': 'Tue, 02 Feb 2016 03:27:00 GMT', 'content-type': 'application/json'}",
+ "service":"Nova",
+ "status_code":"200",
+ "url":"v2.1/<id>/os-agents",
+ "verb":"POST"
+},
+ {
+ "name":"AgentsAdminTestJSON:test_create_agent",
+ "request_body":"{\"agent\": {\"url\": \"xxx://xxxx/xxx/xxx\", \"hypervisor\": \"kvm\", \"md5hash\": \"add6bb58e139be103324d04d82d8f545\", \"version\": \"7.0\", \"architecture\": \"tempest-x86-252246646\", \"os\": \"win\"}}",
+ "request_headers":"{'Content-Type': 'application/json', 'Accept': 'application/json', 'X-Auth-Token': '<omitted>'}",
+ "response_body":"{\"agent\": {\"url\": \"xxx://xxxx/xxx/xxx\", \"hypervisor\": \"kvm\", \"md5hash\": \"add6bb58e139be103324d04d82d8f545\", \"version\": \"7.0\", \"architecture\": \"tempest-x86-252246646\", \"os\": \"win\", \"agent_id\": 2}}",
+ "response_headers":"{'status': '200', 'content-length': '195', 'x-compute-request-id': 'req-b4136f06-c015-4e7e-995f-c43831e3ecce', 'vary': 'X-OpenStack-Nova-API-Version', 'connection': 'close', 'x-openstack-nova-api-version': '2.1', 'date': 'Tue, 02 Feb 2016 03:27:00 GMT', 'content-type': 'application/json'}",
+ "service":"Nova",
+ "status_code":"200",
+ "url":"v2.1/<id>/os-agents",
+ "verb":"POST"
+},
+ {
+ "name":"AgentsAdminTestJSON:tearDown",
+ "request_body":"None",
+ "request_headers":"{'Content-Type': 'application/json', 'Accept': 'application/json', 'X-Auth-Token': '<omitted>'}",
+ "response_body":"",
+ "response_headers":"{'status': '200', 'content-length': '0', 'x-compute-request-id': 'req-ee905fd6-a5b5-4da4-8c37-5363cb25bd9d', 'vary': 'X-OpenStack-Nova-API-Version', 'connection': 'close', 'x-openstack-nova-api-version': '2.1', 'date': 'Tue, 02 Feb 2016 03:27:00 GMT', 'content-type': 'application/json'}",
+ "service":"Nova",
+ "status_code":"200",
+ "url":"v2.1/<id>/os-agents/1",
+ "verb":"DELETE"
+},
+ {
+ "name":"AgentsAdminTestJSON:_run_cleanups",
+ "request_body":"None",
+ "request_headers":"{'Content-Type': 'application/json', 'Accept': 'application/json', 'X-Auth-Token': '<omitted>'}",
+ "response_headers":"{'status': '200', 'content-length': '0', 'x-compute-request-id': 'req-e912cac0-63e0-4679-a68a-b6d18ddca074', 'vary': 'X-OpenStack-Nova-API-Version', 'connection': 'close', 'x-openstack-nova-api-version': '2.1', 'date': 'Tue, 02 Feb 2016 03:27:00 GMT', 'content-type': 'application/json'}",
+ "service":"Nova",
+ "status_code":"200",
+ "url":"v2.1/<id>/os-agents/2",
+ "verb":"DELETE"
+}], "foo":[
+ {
+ "name":"AgentsAdminTestJSON:setUp",
+ "request_body":"{\"agent\": {\"url\": \"xxx://xxxx/xxx/xxx\", \"hypervisor\": \"common\", \"md5hash\": \"add6bb58e139be103324d04d82d8f545\", \"version\": \"7.0\", \"architecture\": \"tempest-x86_64-948635295\", \"os\": \"linux\"}}",
+ "request_headers":"{'Content-Type': 'application/json', 'Accept': 'application/json', 'X-Auth-Token': '<omitted>'}",
+ "response_body":"{\"agent\": {\"url\": \"xxx://xxxx/xxx/xxx\", \"hypervisor\": \"common\", \"md5hash\": \"add6bb58e139be103324d04d82d8f545\", \"version\": \"7.0\", \"architecture\": \"tempest-x86_64-948635295\", \"os\": \"linux\", \"agent_id\": 3}}",
+ "response_headers":"{'status': '200', 'content-length': '203', 'x-compute-request-id': 'req-ccd2116d-04b1-4ffe-ae32-fb623f68bf1c', 'vary': 'X-OpenStack-Nova-API-Version', 'connection': 'close', 'x-openstack-nova-api-version': '2.1', 'date': 'Tue, 02 Feb 2016 03:27:01 GMT', 'content-type': 'application/json'}",
+ "service":"Nova",
+ "status_code":"200",
+ "url":"v2.1/<id>/os-agents",
+ "verb":"POST"
+},
+ {
+ "name":"AgentsAdminTestJSON:test_delete_agent",
+ "request_body":"None",
+ "request_headers":"{'Content-Type': 'application/json', 'Accept': 'application/json', 'X-Auth-Token': '<omitted>'}",
+ "response_body":"",
+ "response_headers":"{'status': '200', 'content-length': '0', 'x-compute-request-id': 'req-6e7fa28f-ae61-4388-9a78-947c58bc0588', 'vary': 'X-OpenStack-Nova-API-Version', 'connection': 'close', 'x-openstack-nova-api-version': '2.1', 'date': 'Tue, 02 Feb 2016 03:27:01 GMT', 'content-type': 'application/json'}",
+ "service":"Nova",
+ "status_code":"200",
+ "url":"v2.1/<id>/os-agents/3",
+ "verb":"DELETE"
+},
+ {
+ "name":"AgentsAdminTestJSON:test_delete_agent",
+ "request_body":"None",
+ "request_headers":"{'Content-Type': 'application/json', 'Accept': 'application/json', 'X-Auth-Token': '<omitted>'}",
+ "response_body":"{\"agents\": []}",
+ "response_headers":"{'status': '200', 'content-length': '14', 'content-location': 'http://23.253.76.97:8774/v2.1/cf6b1933fe5b476fbbabb876f6d1b924/os-agents', 'x-compute-request-id': 'req-e41aa9b4-41a6-4138-ae04-220b768eb644', 'vary': 'X-OpenStack-Nova-API-Version', 'connection': 'close', 'x-openstack-nova-api-version': '2.1', 'date': 'Tue, 02 Feb 2016 03:27:01 GMT', 'content-type': 'application/json'}",
+ "service":"Nova",
+ "status_code":"200",
+ "url":"v2.1/<id>/os-agents",
+ "verb":"GET"
+},
+ {
+ "name":"AgentsAdminTestJSON:tearDown",
+ "request_body":"None",
+ "request_headers":"{'Content-Type': 'application/json', 'Accept': 'application/json', 'X-Auth-Token': '<omitted>'}",
+ "response_headers":"{'status': '404', 'content-length': '82', 'x-compute-request-id': 'req-e297aeea-91cf-4f26-b49c-8f46b1b7a926', 'vary': 'X-OpenStack-Nova-API-Version', 'connection': 'close', 'x-openstack-nova-api-version': '2.1', 'date': 'Tue, 02 Feb 2016 03:27:02 GMT', 'content-type': 'application/json; charset=UTF-8'}",
+ "service":"Nova",
+ "status_code":"404",
+ "url":"v2.1/<id>/os-agents/3",
+ "verb":"DELETE"
+}]}
\ No newline at end of file
diff --git a/tempest/tests/cmd/test_cleanup_services.py b/tempest/tests/cmd/test_cleanup_services.py
index 7bf7315..fc44793 100644
--- a/tempest/tests/cmd/test_cleanup_services.py
+++ b/tempest/tests/cmd/test_cleanup_services.py
@@ -511,7 +511,8 @@
},
{
"id": "aa77asdf-1234",
- "name": "saved-volume"
+ "name": "saved-volume",
+ "links": [],
}
]
}
diff --git a/tempest/tests/cmd/test_subunit_describe_calls.py b/tempest/tests/cmd/test_subunit_describe_calls.py
index cb34ba6..4fed84a 100644
--- a/tempest/tests/cmd/test_subunit_describe_calls.py
+++ b/tempest/tests/cmd/test_subunit_describe_calls.py
@@ -14,220 +14,422 @@
# License for the specific language governing permissions and limitations
# under the License.
+import argparse
+from io import StringIO
import os
+import shutil
import subprocess
+import sys
import tempfile
+from unittest import mock
+from unittest.mock import patch
+
+from oslo_serialization import jsonutils as json
from tempest.cmd import subunit_describe_calls
from tempest.tests import base
-class TestSubunitDescribeCalls(base.TestCase):
- def test_return_code(self):
- subunit_file = os.path.join(
- os.path.dirname(os.path.abspath(__file__)),
- 'sample_streams/calls.subunit')
- p = subprocess.Popen([
- 'subunit-describe-calls', '-s', subunit_file,
- '-o', tempfile.mkstemp()[1]], stdin=subprocess.PIPE)
- p.communicate()
- self.assertEqual(0, p.returncode)
+class TestArgumentParser(base.TestCase):
+ def test_init(self):
+ test_object = subunit_describe_calls.ArgumentParser()
+ self.assertEqual("subunit-describe-calls", test_object.prog)
+ self.assertEqual(subunit_describe_calls.DESCRIPTION,
+ test_object.description)
- def test_verbose(self):
- subunit_file = os.path.join(
- os.path.dirname(os.path.abspath(__file__)),
- 'sample_streams/calls.subunit')
- p = subprocess.Popen([
- 'subunit-describe-calls', '-s', subunit_file,
- '-v'], stdin=subprocess.PIPE, stdout=subprocess.PIPE)
- stdout = p.communicate()
- self.assertEqual(0, p.returncode)
- self.assertIn(b'- request headers:', stdout[0])
- self.assertIn(b'- request body:', stdout[0])
- self.assertIn(b'- response headers:', stdout[0])
- self.assertIn(b'- response body:', stdout[0])
- def test_return_code_no_output(self):
- subunit_file = os.path.join(
+class TestUrlParser(base.TestCase):
+ services_custom_ports = {
+ "18776": "Block Storage",
+ "18774": "Nova",
+ "18773": "Nova-API",
+ "18386": "Sahara",
+ "35358": "Keystone",
+ "19292": "Glance",
+ "19696": "Neutron",
+ "16000": "Swift",
+ "18004": "Heat",
+ "18777": "Ceilometer",
+ "10080": "Horizon",
+ "18080": "Swift",
+ "1873": "rsync",
+ "13260": "iSCSI",
+ "13306": "MySQL",
+ "15672": "AMQP",
+ "18082": "murano"}
+
+ def setUp(self):
+ super(TestUrlParser, self).setUp()
+ self.test_object = subunit_describe_calls.UrlParser()
+
+ def test_get_service_default_ports(self):
+ base_url = "http://site.something.com:"
+ for port in self.test_object.services:
+ url = base_url + port + "/v2/action"
+ service = self.test_object.services[port]
+ self.assertEqual(service, self.test_object.get_service(url))
+
+ def test_get_service_custom_ports(self):
+ self.test_object = subunit_describe_calls.\
+ UrlParser(services=self.services_custom_ports)
+ base_url = "http://site.something.com:"
+ for port in self.services_custom_ports:
+ url = base_url + port + "/v2/action"
+ service = self.services_custom_ports[port]
+ self.assertEqual(service, self.test_object.get_service(url))
+
+ def test_get_service_port_not_found(self):
+ url = "https://site.somewhere.com:1234/v2/action"
+ self.assertEqual("Unknown", self.test_object.get_service(url))
+ self.assertEqual("Unknown", self.test_object.get_service(""))
+
+ def test_parse_details_none(self):
+ self.assertIsNone(self.test_object.parse_details(None))
+
+ def test_url_path_ports(self):
+ uuid_sample1 = "3715e0bb-b1b3-4291-aa13-2c86c3b9ec93"
+ uuid_sample2 = "2715e0bb-b1b4-4291-aa13-2c86c3b9ec88"
+
+ # test http url
+ host = "http://host.company.com"
+ url = host + ":8776/v3/" + uuid_sample1 + "/types/" + \
+ uuid_sample2 + "/extra_specs"
+ self.assertEqual("v3/<uuid>/types/<uuid>/extra_specs",
+ self.test_object.url_path(url))
+ url = host + ":8774/v2.1/servers/" + uuid_sample1
+ self.assertEqual("v2.1/servers/<uuid>",
+ self.test_object.url_path(url))
+ # test https url
+ host = "https://host.company.com"
+ url = host + ":8776/v3/" + uuid_sample1 + "/types/" + \
+ uuid_sample2 + "/extra_specs"
+ self.assertEqual("v3/<uuid>/types/<uuid>/extra_specs",
+ self.test_object.url_path(url))
+ url = host + ":8774/v2.1/servers/" + uuid_sample1
+ self.assertEqual("v2.1/servers/<uuid>",
+ self.test_object.url_path(url))
+
+ def test_url_path_no_match(self):
+ host_port = 'https://host.company.com:1234/'
+ url = 'v2/action/no/special/data'
+ self.assertEqual(url, self.test_object.url_path(host_port + url))
+ url = 'data'
+ self.assertEqual(url, self.test_object.url_path(url))
+
+
+class TestCliBase(base.TestCase):
+ """Base class for share code on all CLI sub-process testing"""
+
+ def setUp(self):
+ super(TestCliBase, self).setUp()
+ self._subunit_file = os.path.join(
os.path.dirname(os.path.abspath(__file__)),
- 'sample_streams/calls.subunit')
+ 'subunit_describe_calls_data', 'calls.subunit')
+
+ def _bytes_to_string(self, data):
+ if isinstance(data, (bytes, bytearray)):
+ data = str(data, 'utf-8')
+ return data
+
+ def _assert_cli_message(self, data):
+ data = self._bytes_to_string(data)
+ self.assertIn("Running subunit_describe_calls ...", data)
+
+ def _assert_deprecated_warning(self, stdout):
+ self.assertIn(
+ b"Use of: 'subunit-describe-calls' is deprecated, "
+ b"please use: 'tempest subunit-describe-calls'", stdout)
+
+ def _assert_expect_json(self, json_data):
+ expected_file_name = os.path.join(
+ os.path.dirname(os.path.abspath(__file__)),
+ 'subunit_describe_calls_data', 'calls_subunit_expected.json')
+ with open(expected_file_name, "rb") as read_file:
+ expected_result = json.load(read_file)
+ self.assertDictEqual(expected_result, json_data)
+
+ def _assert_headers_and_bodies(self, data):
+ data = self._bytes_to_string(data)
+ self.assertIn('- request headers:', data)
+ self.assertIn('- request body:', data)
+ self.assertIn('- response headers:', data)
+ self.assertIn('- response body:', data)
+
+ def _assert_methods_details(self, data):
+ data = self._bytes_to_string(data)
+ self.assertIn('foo', data)
+ self.assertIn('- 200 POST request for Nova to v2.1/<id>/',
+ data)
+ self.assertIn('- 200 DELETE request for Nova to v2.1/<id>/',
+ data)
+ self.assertIn('- 200 GET request for Nova to v2.1/<id>/',
+ data)
+ self.assertIn('- 404 DELETE request for Nova to v2.1/<id>/',
+ data)
+
+ def _assert_mutual_exclusive_message(self, stderr):
+ self.assertIn(b"usage: subunit-describe-calls "
+ b"[-h] [-s [<subunit file>]]", stderr)
+ self.assertIn(b"[-n <non subunit name>] [-o <output file>]",
+ stderr)
+ self.assertIn(b"[-p <ports file>] [-v | -a]", stderr)
+ self.assertIn(
+ b"subunit-describe-calls: error: argument -v/--verbose: "
+ b"not allowed with argument -a/--all-stdout", stderr)
+
+ def _assert_no_headers_and_bodies(self, data):
+ data = self._bytes_to_string(data)
+ self.assertNotIn('- request headers:', data)
+ self.assertNotIn('- request body:', data)
+ self.assertNotIn('- response headers:', data)
+ self.assertNotIn('- response body:', data)
+
+
+class TestMainCli(TestCliBase):
+ """Test cases that use subunit_describe_calls module main interface
+
+ via subprocess calls to make sure the total user experience
+ is well defined and tested. This interface is deprecated.
+ Note: these test do not affect code coverage percentages.
+ """
+
+ def test_main_output_file(self):
+ temp_file = tempfile.mkstemp()[1]
p = subprocess.Popen([
- 'subunit-describe-calls', '-s', subunit_file],
- stdin=subprocess.PIPE, stdout=subprocess.PIPE)
- stdout = p.communicate()
+ 'subunit-describe-calls', '-s', self._subunit_file,
+ '-o', temp_file], stdin=subprocess.PIPE,
+ stdout=subprocess.PIPE, stderr=subprocess.PIPE)
+ stdout, stderr = p.communicate()
self.assertEqual(0, p.returncode)
- self.assertIn(b'foo', stdout[0])
- self.assertIn(b'- 200 POST request for Nova to v2.1/<id>/',
- stdout[0])
- self.assertIn(b'- 200 DELETE request for Nova to v2.1/<id>/',
- stdout[0])
- self.assertIn(b'- 200 GET request for Nova to v2.1/<id>/',
- stdout[0])
- self.assertIn(b'- 404 DELETE request for Nova to v2.1/<id>/',
- stdout[0])
- self.assertNotIn(b'- request headers:', stdout[0])
- self.assertNotIn(b'- request body:', stdout[0])
- self.assertNotIn(b'- response headers:', stdout[0])
- self.assertNotIn(b'- response body:', stdout[0])
+ self._assert_cli_message(stdout)
+ self._assert_deprecated_warning(stdout)
+ with open(temp_file, 'r') as file:
+ data = json.loads(file.read())
+ self._assert_expect_json(data)
+
+ def test_main_verbose(self):
+ p = subprocess.Popen([
+ 'subunit-describe-calls', '-s', self._subunit_file,
+ '-v'], stdin=subprocess.PIPE, stdout=subprocess.PIPE,
+ stderr=subprocess.PIPE)
+ stdout, stderr = p.communicate()
+ self.assertEqual(0, p.returncode)
+ self._assert_cli_message(stdout)
+ self._assert_deprecated_warning(stdout)
+ self._assert_methods_details(stdout)
+ self._assert_headers_and_bodies(stdout)
+
+ def test_main_all_stdout(self):
+ p = subprocess.Popen([
+ 'subunit-describe-calls', '-s', self._subunit_file,
+ '--all-stdout'], stdin=subprocess.PIPE, stdout=subprocess.PIPE,
+ stderr=subprocess.PIPE)
+ stdout, stderr = p.communicate()
+ self.assertEqual(0, p.returncode)
+ self._assert_cli_message(stdout)
+ self._assert_deprecated_warning(stdout)
+ self._assert_methods_details(stdout)
+ self._assert_headers_and_bodies(stdout)
+
+ def test_main(self):
+ p = subprocess.Popen([
+ 'subunit-describe-calls', '-s', self._subunit_file],
+ stdin=subprocess.PIPE, stdout=subprocess.PIPE,
+ stderr=subprocess.PIPE)
+ stdout, stderr = p.communicate()
+ self.assertEqual(0, p.returncode)
+ self._assert_cli_message(stdout)
+ self._assert_deprecated_warning(stdout)
+ self._assert_methods_details(stdout)
+ self._assert_no_headers_and_bodies(stdout)
+
+ def test_main_verbose_and_all_stdout(self):
+ p = subprocess.Popen([
+ 'subunit-describe-calls', '-s', self._subunit_file,
+ '-a', '-v'],
+ stdin=subprocess.PIPE, stdout=subprocess.PIPE,
+ stderr=subprocess.PIPE)
+ stdout, stderr = p.communicate()
+ self.assertEqual(2, p.returncode)
+ self._assert_cli_message(stdout)
+ self._assert_deprecated_warning(stdout)
+ self._assert_mutual_exclusive_message(stderr)
+
+
+class TestCli(TestCliBase):
+ """Test cases that use tempest subunit_describe_calls cliff interface
+
+ via subprocess calls to make sure the total user experience
+ is well defined and tested.
+ Note: these test do not affect code coverage percentages.
+ """
+
+ def _assert_cliff_verbose(self, stdout):
+ self.assertIn(b'tempest initialize_app', stdout)
+ self.assertIn(b'prepare_to_run_command TempestSubunitDescribeCalls',
+ stdout)
+ self.assertIn(b'tempest clean_up TempestSubunitDescribeCalls',
+ stdout)
+
+ def test_run_all_stdout(self):
+ p = subprocess.Popen(['tempest', 'subunit-describe-calls',
+ '-s', self._subunit_file, '-a'],
+ stdin=subprocess.PIPE,
+ stdout=subprocess.PIPE,
+ stderr=subprocess.PIPE)
+ stdout, stderr = p.communicate()
+ self.assertEqual(0, p.returncode)
+ self._assert_cli_message(stdout)
+ self._assert_methods_details(stdout)
+ self._assert_headers_and_bodies(stdout)
+
+ def test_run_verbose(self):
+ p = subprocess.Popen(['tempest', 'subunit-describe-calls',
+ '-s', self._subunit_file, '-v'],
+ stdin=subprocess.PIPE,
+ stdout=subprocess.PIPE,
+ stderr=subprocess.PIPE)
+ stdout, stderr = p.communicate()
+ self.assertEqual(0, p.returncode)
+ self._assert_cli_message(stdout)
+ self._assert_methods_details(stdout)
+ self._assert_no_headers_and_bodies(stdout)
+ self._assert_cliff_verbose(stderr)
+
+ def test_run_min(self):
+ p = subprocess.Popen(['tempest', 'subunit-describe-calls',
+ '-s', self._subunit_file],
+ stdin=subprocess.PIPE,
+ stdout=subprocess.PIPE,
+ stderr=subprocess.PIPE)
+ stdout, stderr = p.communicate()
+ self.assertEqual(0, p.returncode)
+ self._assert_cli_message(stdout)
+ self._assert_methods_details(stdout)
+ self._assert_no_headers_and_bodies(stdout)
+
+ def test_run_verbose_all_stdout(self):
+ """Test Cliff -v argument
+
+ Since Cliff framework has a argument at the
+ abstract command level the -v or --verbose for
+ this command is not processed as a boolean.
+ So the use of verbose only exists for the
+ deprecated main CLI interface. When the
+ main is deleted this test would not be needed.
+ """
+ p = subprocess.Popen(['tempest', 'subunit-describe-calls',
+ '-s', self._subunit_file, '-a', '-v'],
+ stdin=subprocess.PIPE,
+ stdout=subprocess.PIPE,
+ stderr=subprocess.PIPE)
+ stdout, stderr = p.communicate()
+ self.assertEqual(0, p.returncode)
+ self._assert_cli_message(stdout)
+ self._assert_cliff_verbose(stderr)
+ self._assert_methods_details(stdout)
+
+
+class TestSubunitDescribeCalls(TestCliBase):
+ """Test cases use the subunit_describe_calls module interface
+
+ and effect code coverage reporting
+ """
+
+ def setUp(self):
+ super(TestSubunitDescribeCalls, self).setUp()
+ self.test_object = subunit_describe_calls.TempestSubunitDescribeCalls(
+ app=mock.Mock(),
+ app_args=mock.Mock(spec=argparse.Namespace))
def test_parse(self):
- subunit_file = os.path.join(
- os.path.dirname(os.path.abspath(__file__)),
- 'sample_streams/calls.subunit')
- parser = subunit_describe_calls.parse(
- open(subunit_file), "pythonlogging", None)
- expected_result = {
- 'bar': [{
- 'name': 'AgentsAdminTestJSON:setUp',
- 'request_body': '{"agent": {"url": "xxx://xxxx/xxx/xxx", '
- '"hypervisor": "common", "md5hash": '
- '"add6bb58e139be103324d04d82d8f545", "version": "7.0", '
- '"architecture": "tempest-x86_64-424013832", "os": "linux"}}',
- 'request_headers': "{'Content-Type': 'application/json', "
- "'Accept': 'application/json', 'X-Auth-Token': '<omitted>'}",
- 'response_body': '{"agent": {"url": "xxx://xxxx/xxx/xxx", '
- '"hypervisor": "common", "md5hash": '
- '"add6bb58e139be103324d04d82d8f545", "version": "7.0", '
- '"architecture": "tempest-x86_64-424013832", "os": "linux", '
- '"agent_id": 1}}',
- 'response_headers': "{'status': '200', 'content-length': "
- "'203', 'x-compute-request-id': "
- "'req-25ddaae2-0ef1-40d1-8228-59bd64a7e75b', 'vary': "
- "'X-OpenStack-Nova-API-Version', 'connection': 'close', "
- "'x-openstack-nova-api-version': '2.1', 'date': "
- "'Tue, 02 Feb 2016 03:27:00 GMT', 'content-type': "
- "'application/json'}",
- 'service': 'Nova',
- 'status_code': '200',
- 'url': 'v2.1/<id>/os-agents',
- 'verb': 'POST'}, {
- 'name': 'AgentsAdminTestJSON:test_create_agent',
- 'request_body': '{"agent": {"url": "xxx://xxxx/xxx/xxx", '
- '"hypervisor": "kvm", "md5hash": '
- '"add6bb58e139be103324d04d82d8f545", "version": "7.0", '
- '"architecture": "tempest-x86-252246646", "os": "win"}}',
- 'request_headers': "{'Content-Type': 'application/json', "
- "'Accept': 'application/json', 'X-Auth-Token': '<omitted>'}",
- 'response_body': '{"agent": {"url": "xxx://xxxx/xxx/xxx", '
- '"hypervisor": "kvm", "md5hash": '
- '"add6bb58e139be103324d04d82d8f545", "version": "7.0", '
- '"architecture": "tempest-x86-252246646", "os": "win", '
- '"agent_id": 2}}',
- 'response_headers': "{'status': '200', 'content-length': "
- "'195', 'x-compute-request-id': "
- "'req-b4136f06-c015-4e7e-995f-c43831e3ecce', 'vary': "
- "'X-OpenStack-Nova-API-Version', 'connection': 'close', "
- "'x-openstack-nova-api-version': '2.1', 'date': "
- "'Tue, 02 Feb 2016 03:27:00 GMT', 'content-type': "
- "'application/json'}",
- 'service': 'Nova',
- 'status_code': '200',
- 'url': 'v2.1/<id>/os-agents',
- 'verb': 'POST'}, {
- 'name': 'AgentsAdminTestJSON:tearDown',
- 'request_body': 'None',
- 'request_headers': "{'Content-Type': 'application/json', "
- "'Accept': 'application/json', 'X-Auth-Token': '<omitted>'}",
- 'response_body': '',
- 'response_headers': "{'status': '200', 'content-length': "
- "'0', 'x-compute-request-id': "
- "'req-ee905fd6-a5b5-4da4-8c37-5363cb25bd9d', 'vary': "
- "'X-OpenStack-Nova-API-Version', 'connection': 'close', "
- "'x-openstack-nova-api-version': '2.1', 'date': "
- "'Tue, 02 Feb 2016 03:27:00 GMT', 'content-type': "
- "'application/json'}",
- 'service': 'Nova',
- 'status_code': '200',
- 'url': 'v2.1/<id>/os-agents/1',
- 'verb': 'DELETE'}, {
- 'name': 'AgentsAdminTestJSON:_run_cleanups',
- 'request_body': 'None',
- 'request_headers': "{'Content-Type': 'application/json', "
- "'Accept': 'application/json', 'X-Auth-Token': '<omitted>'}",
- 'response_headers': "{'status': '200', 'content-length': "
- "'0', 'x-compute-request-id': "
- "'req-e912cac0-63e0-4679-a68a-b6d18ddca074', 'vary': "
- "'X-OpenStack-Nova-API-Version', 'connection': 'close', "
- "'x-openstack-nova-api-version': '2.1', 'date': "
- "'Tue, 02 Feb 2016 03:27:00 GMT', 'content-type': "
- "'application/json'}",
- 'service': 'Nova',
- 'status_code': '200',
- 'url': 'v2.1/<id>/os-agents/2',
- 'verb': 'DELETE'}],
- 'foo': [{
- 'name': 'AgentsAdminTestJSON:setUp',
- 'request_body': '{"agent": {"url": "xxx://xxxx/xxx/xxx", '
- '"hypervisor": "common", "md5hash": '
- '"add6bb58e139be103324d04d82d8f545", "version": "7.0", '
- '"architecture": "tempest-x86_64-948635295", "os": "linux"}}',
- 'request_headers': "{'Content-Type': 'application/json', "
- "'Accept': 'application/json', 'X-Auth-Token': '<omitted>'}",
- 'response_body': '{"agent": {"url": "xxx://xxxx/xxx/xxx", '
- '"hypervisor": "common", "md5hash": '
- '"add6bb58e139be103324d04d82d8f545", "version": "7.0", '
- '"architecture": "tempest-x86_64-948635295", "os": "linux", '
- '"agent_id": 3}}',
- 'response_headers': "{'status': '200', 'content-length': "
- "'203', 'x-compute-request-id': "
- "'req-ccd2116d-04b1-4ffe-ae32-fb623f68bf1c', 'vary': "
- "'X-OpenStack-Nova-API-Version', 'connection': 'close', "
- "'x-openstack-nova-api-version': '2.1', 'date': "
- "'Tue, 02 Feb 2016 03:27:01 GMT', 'content-type': "
- "'application/json'}",
- 'service': 'Nova',
- 'status_code': '200',
- 'url': 'v2.1/<id>/os-agents',
- 'verb': 'POST'}, {
- 'name': 'AgentsAdminTestJSON:test_delete_agent',
- 'request_body': 'None',
- 'request_headers': "{'Content-Type': 'application/json', "
- "'Accept': 'application/json', 'X-Auth-Token': '<omitted>'}",
- 'response_body': '',
- 'response_headers': "{'status': '200', 'content-length': "
- "'0', 'x-compute-request-id': "
- "'req-6e7fa28f-ae61-4388-9a78-947c58bc0588', 'vary': "
- "'X-OpenStack-Nova-API-Version', 'connection': 'close', "
- "'x-openstack-nova-api-version': '2.1', 'date': "
- "'Tue, 02 Feb 2016 03:27:01 GMT', 'content-type': "
- "'application/json'}",
- 'service': 'Nova',
- 'status_code': '200',
- 'url': 'v2.1/<id>/os-agents/3',
- 'verb': 'DELETE'}, {
- 'name': 'AgentsAdminTestJSON:test_delete_agent',
- 'request_body': 'None',
- 'request_headers': "{'Content-Type': 'application/json', "
- "'Accept': 'application/json', 'X-Auth-Token': '<omitted>'}",
- 'response_body': '{"agents": []}',
- 'response_headers': "{'status': '200', 'content-length': "
- "'14', 'content-location': "
- "'http://23.253.76.97:8774/v2.1/"
- "cf6b1933fe5b476fbbabb876f6d1b924/os-agents', "
- "'x-compute-request-id': "
- "'req-e41aa9b4-41a6-4138-ae04-220b768eb644', 'vary': "
- "'X-OpenStack-Nova-API-Version', 'connection': 'close', "
- "'x-openstack-nova-api-version': '2.1', 'date': "
- "'Tue, 02 Feb 2016 03:27:01 GMT', 'content-type': "
- "'application/json'}",
- 'service': 'Nova',
- 'status_code': '200',
- 'url': 'v2.1/<id>/os-agents',
- 'verb': 'GET'}, {
- 'name': 'AgentsAdminTestJSON:tearDown',
- 'request_body': 'None',
- 'request_headers': "{'Content-Type': 'application/json', "
- "'Accept': 'application/json', 'X-Auth-Token': '<omitted>'}",
- 'response_headers': "{'status': '404', 'content-length': "
- "'82', 'x-compute-request-id': "
- "'req-e297aeea-91cf-4f26-b49c-8f46b1b7a926', 'vary': "
- "'X-OpenStack-Nova-API-Version', 'connection': 'close', "
- "'x-openstack-nova-api-version': '2.1', 'date': "
- "'Tue, 02 Feb 2016 03:27:02 GMT', 'content-type': "
- "'application/json; charset=UTF-8'}",
- 'service': 'Nova',
- 'status_code': '404',
- 'url': 'v2.1/<id>/os-agents/3',
- 'verb': 'DELETE'}]}
+ with open(self._subunit_file, 'r') as read_file:
+ parser = subunit_describe_calls.parse(
+ read_file, "pythonlogging", None)
+ self._assert_expect_json(parser.test_logs)
- self.assertEqual(expected_result, parser.test_logs)
+ def test_get_description(self):
+ self.assertEqual(subunit_describe_calls.DESCRIPTION,
+ self.test_object.get_description())
+
+ def test_get_parser_default_min(self):
+ parser = self.test_object.get_parser('NAME')
+ parsed_args = parser.parse_args([])
+ self.assertIsNone(parsed_args.output_file)
+ self.assertIsNone(parsed_args.ports)
+ self.assertFalse(parsed_args.all_stdout)
+ self.assertEqual(parsed_args.subunit, sys.stdin)
+
+ def test_get_parser_default_max(self):
+ temp_dir = tempfile.mkdtemp(prefix="parser")
+ self.addCleanup(shutil.rmtree, temp_dir, ignore_errors=True)
+ outfile_name = os.path.join(temp_dir, 'output.json')
+ open(outfile_name, 'a').close()
+ portfile_name = os.path.join(temp_dir, 'ports.json')
+ open(portfile_name, 'a').close()
+
+ parser = self.test_object.get_parser('NAME')
+ parsed_args = parser.parse_args(["-a", "-o " + outfile_name,
+ "-p " + portfile_name])
+
+ self.assertIsNotNone(parsed_args.output_file)
+ self.assertIsNotNone(parsed_args.ports)
+ self.assertTrue(parsed_args.all_stdout)
+ self.assertEqual(parsed_args.subunit, sys.stdin)
+
+ def test_take_action_min(self):
+ parser = self.test_object.get_parser('NAME')
+ parsed_args = parser.parse_args(["-s" + self._subunit_file],)
+ with patch('sys.stdout', new=StringIO()) as mock_stdout:
+ self.test_object.take_action(parsed_args)
+
+ stdout_data = mock_stdout.getvalue()
+ self._assert_methods_details(stdout_data)
+ self._assert_no_headers_and_bodies(stdout_data)
+
+ def test_take_action_all_stdout(self):
+ parser = self.test_object.get_parser('NAME')
+ parsed_args = parser.parse_args(["-as" + self._subunit_file],)
+ with patch('sys.stdout', new=StringIO()) as mock_stdout:
+ self.test_object.take_action(parsed_args)
+
+ stdout_data = mock_stdout.getvalue()
+ self._assert_methods_details(stdout_data)
+ self._assert_headers_and_bodies(stdout_data)
+
+ def test_take_action_outfile_files(self):
+ temp_file = tempfile.mkstemp()[1]
+ parser = self.test_object.get_parser('NAME')
+ parsed_args = parser.parse_args(
+ ["-as" + self._subunit_file, '-o', temp_file], )
+ with patch('sys.stdout', new=StringIO()) as mock_stdout:
+ self.test_object.take_action(parsed_args)
+ stdout_data = mock_stdout.getvalue()
+ self._assert_cli_message(stdout_data)
+ with open(temp_file, 'r') as file:
+ data = json.loads(file.read())
+ self._assert_expect_json(data)
+
+ def test_take_action_no_items(self):
+ temp_file = tempfile.mkstemp()[1]
+ parser = self.test_object.get_parser('NAME')
+ parsed_args = parser.parse_args(
+ ["-as" + temp_file], )
+ with patch('sys.stdout', new=StringIO()) as mock_stdout:
+ self.test_object.take_action(parsed_args)
+ stdout_data = mock_stdout.getvalue()
+ self._assert_cli_message(stdout_data)
+
+ def test_take_action_exception(self):
+ parser = self.test_object.get_parser('NAME')
+ parsed_args = parser.parse_args(["-s" + self._subunit_file],)
+ with patch('sys.stderr', new=StringIO()) as mock_stderr:
+ with patch('tempest.cmd.subunit_describe_calls.entry_point') \
+ as mock_method:
+ mock_method.side_effect = OSError()
+ self.assertRaises(OSError, self.test_object.take_action,
+ parsed_args)
+ stderr_data = mock_stderr.getvalue()
+
+ self.assertIn("Traceback (most recent call last):", stderr_data)
+ self.assertIn("entry_point(parsed_args)", stderr_data)
diff --git a/tempest/tests/lib/cmd/test_check_uuid.py b/tempest/tests/lib/cmd/test_check_uuid.py
index 130f90a..28ebca1 100644
--- a/tempest/tests/lib/cmd/test_check_uuid.py
+++ b/tempest/tests/lib/cmd/test_check_uuid.py
@@ -11,13 +11,56 @@
# under the License.
import importlib
+import os
+import sys
import tempfile
from unittest import mock
+import fixtures
+
from tempest.lib.cmd import check_uuid
from tempest.tests import base
+class TestCLInterface(base.TestCase):
+ CODE = "import unittest\n" \
+ "class TestClass(unittest.TestCase):\n" \
+ " def test_tests(self):\n" \
+ " pass"
+
+ def create_tests_file(self, directory):
+ with open(directory + "/__init__.py", "w"):
+ pass
+
+ tests_file = directory + "/tests.py"
+ with open(tests_file, "w") as fake_file:
+ fake_file.write(TestCLInterface.CODE)
+
+ return tests_file
+
+ def test_fix_argument_no(self):
+ temp_dir = self.useFixture(fixtures.TempDir(rootdir="."))
+ tests_file = self.create_tests_file(temp_dir.path)
+
+ sys.argv = [sys.argv[0]] + ["--package",
+ os.path.relpath(temp_dir.path)]
+
+ self.assertRaises(SystemExit, check_uuid.run)
+ with open(tests_file, "r") as f:
+ self.assertTrue(TestCLInterface.CODE == f.read())
+
+ def test_fix_argument_yes(self):
+ temp_dir = self.useFixture(fixtures.TempDir(rootdir="."))
+ tests_file = self.create_tests_file(temp_dir.path)
+
+ sys.argv = [sys.argv[0]] + ["--fix", "--package",
+ os.path.relpath(temp_dir.path)]
+
+ check_uuid.run()
+ with open(tests_file, "r") as f:
+ self.assertTrue(TestCLInterface.CODE != f.read())
+
+
class TestSourcePatcher(base.TestCase):
def test_add_patch(self):
patcher = check_uuid.SourcePatcher()
diff --git a/tempest/tests/lib/services/volume/v3/test_volumes_client.py b/tempest/tests/lib/services/volume/v3/test_volumes_client.py
index 56c1a35..6bd75d9 100644
--- a/tempest/tests/lib/services/volume/v3/test_volumes_client.py
+++ b/tempest/tests/lib/services/volume/v3/test_volumes_client.py
@@ -26,10 +26,6 @@
"volume-summary": {
"total_size": 4,
"total_count": 4,
- "metadata": {
- "key1": ["value1", "value2"],
- "key2": ["value2"]
- }
}
}
diff --git a/tools/generate-tempest-plugins-list.py b/tools/generate-tempest-plugins-list.py
index 530ce5e..618c388 100644
--- a/tools/generate-tempest-plugins-list.py
+++ b/tools/generate-tempest-plugins-list.py
@@ -41,7 +41,7 @@
'x/intel-nfv-ci-tests', # https://review.opendev.org/#/c/634640/
'openstack/networking-generic-switch',
# https://review.opendev.org/#/c/634846/
- 'openstack/networking-l2gw-tempest-plugin',
+ 'x/networking-l2gw-tempest-plugin',
# https://review.opendev.org/#/c/635093/
'openstack/networking-midonet', # https://review.opendev.org/#/c/635096/
'x/networking-plumgrid', # https://review.opendev.org/#/c/635096/