Merge "add py27 to default tox"
diff --git a/etc/tempest.conf.sample b/etc/tempest.conf.sample
index bf0b308..2ed2582 100644
--- a/etc/tempest.conf.sample
+++ b/etc/tempest.conf.sample
@@ -740,6 +740,11 @@
# all which indicates every extension is enabled (list value)
#api_extensions=all
+# Allow the execution of IPv6 subnet tests that use the
+# extended IPv6 attributes ipv6_ra_mode and ipv6_address_mode
+# (boolean value)
+#ipv6_subnet_attributes=false
+
[object-storage]
@@ -845,6 +850,28 @@
# queues (integer value)
#max_queues_per_page=20
+# The maximum metadata size for a queue (integer value)
+#max_queue_metadata=65536
+
+# The maximum number of queue message per page when listing
+# (or) posting messages (integer value)
+#max_messages_per_page=20
+
+# The maximum size of a message body (integer value)
+#max_message_size=262144
+
+# The maximum number of messages per claim (integer value)
+#max_messages_per_claim=20
+
+# The maximum ttl for a message (integer value)
+#max_message_ttl=1209600
+
+# The maximum ttl for a claim (integer value)
+#max_claim_ttl=43200
+
+# The maximum grace period for a claim (integer value)
+#max_claim_grace=43200
+
[scenario]
diff --git a/tempest/api/compute/admin/test_quotas_negative.py b/tempest/api/compute/admin/test_quotas_negative.py
index e1dc685..f147b9c 100644
--- a/tempest/api/compute/admin/test_quotas_negative.py
+++ b/tempest/api/compute/admin/test_quotas_negative.py
@@ -44,6 +44,7 @@
# TODO(afazekas): Add dedicated tenant to the skiped quota tests
# it can be moved into the setUpClass as well
+ @test.skip_because(bug="1298131")
@test.attr(type=['negative', 'gate'])
def test_create_server_when_cpu_quota_is_full(self):
# Disallow server creation when tenant's vcpu quota is full
@@ -57,8 +58,9 @@
self.addCleanup(self.adm_client.update_quota_set, self.demo_tenant_id,
cores=default_vcpu_quota)
- self.assertRaises(exceptions.OverLimit, self.create_test_server)
+ self.assertRaises(exceptions.Unauthorized, self.create_test_server)
+ @test.skip_because(bug="1298131")
@test.attr(type=['negative', 'gate'])
def test_create_server_when_memory_quota_is_full(self):
# Disallow server creation when tenant's memory quota is full
@@ -72,8 +74,9 @@
self.addCleanup(self.adm_client.update_quota_set, self.demo_tenant_id,
ram=default_mem_quota)
- self.assertRaises(exceptions.OverLimit, self.create_test_server)
+ self.assertRaises(exceptions.Unauthorized, self.create_test_server)
+ @test.skip_because(bug="1298131")
@test.attr(type=['negative', 'gate'])
def test_create_server_when_instances_quota_is_full(self):
# Once instances quota limit is reached, disallow server creation
@@ -86,7 +89,7 @@
instances=instances_quota)
self.addCleanup(self.adm_client.update_quota_set, self.demo_tenant_id,
instances=default_instances_quota)
- self.assertRaises(exceptions.OverLimit, self.create_test_server)
+ self.assertRaises(exceptions.Unauthorized, self.create_test_server)
@test.skip_because(bug="1186354",
condition=CONF.service_available.neutron)
diff --git a/tempest/api/compute/admin/test_servers_negative.py b/tempest/api/compute/admin/test_servers_negative.py
index 8b3a0b5..15025ba 100644
--- a/tempest/api/compute/admin/test_servers_negative.py
+++ b/tempest/api/compute/admin/test_servers_negative.py
@@ -54,6 +54,7 @@
flavor_id = data_utils.rand_int_id(start=1000)
return flavor_id
+ @test.skip_because(bug="1298131")
@test.attr(type=['negative', 'gate'])
def test_resize_server_using_overlimit_ram(self):
flavor_name = data_utils.rand_name("flavor-")
@@ -67,11 +68,12 @@
ram, vcpus, disk,
flavor_id)
self.addCleanup(self.flavors_client.delete_flavor, flavor_id)
- self.assertRaises(exceptions.OverLimit,
+ self.assertRaises(exceptions.Unauthorized,
self.client.resize,
self.servers[0]['id'],
flavor_ref['id'])
+ @test.skip_because(bug="1298131")
@test.attr(type=['negative', 'gate'])
def test_resize_server_using_overlimit_vcpus(self):
flavor_name = data_utils.rand_name("flavor-")
@@ -85,7 +87,7 @@
ram, vcpus, disk,
flavor_id)
self.addCleanup(self.flavors_client.delete_flavor, flavor_id)
- self.assertRaises(exceptions.OverLimit,
+ self.assertRaises(exceptions.Unauthorized,
self.client.resize,
self.servers[0]['id'],
flavor_ref['id'])
diff --git a/tempest/api/compute/base.py b/tempest/api/compute/base.py
index 7c70aec..a1aaa95 100644
--- a/tempest/api/compute/base.py
+++ b/tempest/api/compute/base.py
@@ -270,7 +270,7 @@
if not policy:
policy = ['affinity']
resp, body = cls.servers_client.create_server_group(name, policy)
- cls.server_groups.append(body)
+ cls.server_groups.append(body['id'])
return resp, body
def wait_for(self, condition):
diff --git a/tempest/api/compute/servers/test_create_server.py b/tempest/api/compute/servers/test_create_server.py
index e135eca..279dc51 100644
--- a/tempest/api/compute/servers/test_create_server.py
+++ b/tempest/api/compute/servers/test_create_server.py
@@ -102,6 +102,28 @@
self.password)
self.assertTrue(linux_client.hostname_equals_servername(self.name))
+ @test.skip_because(bug="1306367", interface="xml")
+ @test.attr(type='gate')
+ def test_create_server_with_scheduler_hint_group(self):
+ # Create a server with the scheduler hint "group".
+ name = data_utils.rand_name('server_group')
+ policies = ['affinity']
+ resp, body = self.client.create_server_group(name=name,
+ policies=policies)
+ self.assertEqual(200, resp.status)
+ group_id = body['id']
+ self.addCleanup(self.client.delete_server_group, group_id)
+
+ hints = {'group': group_id}
+ resp, server = self.create_test_server(sched_hints=hints,
+ wait_until='ACTIVE')
+ self.assertEqual(202, resp.status)
+
+ # Check a server is in the group
+ resp, server_group = self.client.get_server_group(group_id)
+ self.assertEqual(200, resp.status)
+ self.assertIn(server['id'], server_group['members'])
+
class ServersWithSpecificFlavorTestJSON(base.BaseV2ComputeAdminTest):
disk_config = 'AUTO'
diff --git a/tempest/api/compute/servers/test_server_actions.py b/tempest/api/compute/servers/test_server_actions.py
index d0fd876..71fcbff 100644
--- a/tempest/api/compute/servers/test_server_actions.py
+++ b/tempest/api/compute/servers/test_server_actions.py
@@ -14,6 +14,7 @@
# under the License.
import base64
+import logging
import testtools
import urlparse
@@ -27,6 +28,8 @@
CONF = config.CONF
+LOG = logging.getLogger(__name__)
+
class ServerActionsTestJSON(base.BaseV2ComputeTest):
run_ssh = CONF.compute.run_ssh
@@ -267,7 +270,14 @@
# the oldest one should be deleted automatically in this test
def _clean_oldest_backup(oldest_backup):
if oldest_backup_exist:
- self.os.image_client.delete_image(oldest_backup)
+ try:
+ self.os.image_client.delete_image(oldest_backup)
+ except exceptions.NotFound:
+ pass
+ else:
+ LOG.warning("Deletion of oldest backup %s should not have "
+ "been successful as it should have been "
+ "deleted during rotation." % oldest_backup)
image1_id = data_utils.parse_image_id(resp['location'])
self.addCleanup(_clean_oldest_backup, image1_id)
diff --git a/tempest/api/compute/servers/test_server_metadata.py b/tempest/api/compute/servers/test_server_metadata.py
index 448b8ff..01ff6b9 100644
--- a/tempest/api/compute/servers/test_server_metadata.py
+++ b/tempest/api/compute/servers/test_server_metadata.py
@@ -24,10 +24,7 @@
super(ServerMetadataTestJSON, cls).setUpClass()
cls.client = cls.servers_client
cls.quotas = cls.quotas_client
- cls.admin_client = cls._get_identity_admin_client()
- resp, tenants = cls.admin_client.list_tenants()
resp, server = cls.create_test_server(meta={}, wait_until='ACTIVE')
-
cls.server_id = server['id']
def setUp(self):
diff --git a/tempest/api/compute/v3/admin/test_quotas_negative.py b/tempest/api/compute/v3/admin/test_quotas_negative.py
index 307462f..7739f09 100644
--- a/tempest/api/compute/v3/admin/test_quotas_negative.py
+++ b/tempest/api/compute/v3/admin/test_quotas_negative.py
@@ -34,6 +34,7 @@
# TODO(afazekas): Add dedicated tenant to the skiped quota tests
# it can be moved into the setUpClass as well
+ @test.skip_because(bug="1298131")
@test.attr(type=['negative', 'gate'])
def test_create_server_when_cpu_quota_is_full(self):
# Disallow server creation when tenant's vcpu quota is full
@@ -47,8 +48,9 @@
self.addCleanup(self.adm_client.update_quota_set, self.demo_tenant_id,
cores=default_vcpu_quota)
- self.assertRaises(exceptions.OverLimit, self.create_test_server)
+ self.assertRaises(exceptions.Unauthorized, self.create_test_server)
+ @test.skip_because(bug="1298131")
@test.attr(type=['negative', 'gate'])
def test_create_server_when_memory_quota_is_full(self):
# Disallow server creation when tenant's memory quota is full
@@ -62,7 +64,7 @@
self.addCleanup(self.adm_client.update_quota_set, self.demo_tenant_id,
ram=default_mem_quota)
- self.assertRaises(exceptions.OverLimit, self.create_test_server)
+ self.assertRaises(exceptions.Unauthorized, self.create_test_server)
@test.attr(type=['negative', 'gate'])
def test_update_quota_normal_user(self):
@@ -71,6 +73,7 @@
self.demo_tenant_id,
ram=0)
+ @test.skip_because(bug="1298131")
@test.attr(type=['negative', 'gate'])
def test_create_server_when_instances_quota_is_full(self):
# Once instances quota limit is reached, disallow server creation
@@ -83,4 +86,4 @@
instances=instances_quota)
self.addCleanup(self.adm_client.update_quota_set, self.demo_tenant_id,
instances=default_instances_quota)
- self.assertRaises(exceptions.OverLimit, self.create_test_server)
+ self.assertRaises(exceptions.Unauthorized, self.create_test_server)
diff --git a/tempest/api/compute/v3/servers/test_server_actions.py b/tempest/api/compute/v3/servers/test_server_actions.py
index e098311..3ee8050 100644
--- a/tempest/api/compute/v3/servers/test_server_actions.py
+++ b/tempest/api/compute/v3/servers/test_server_actions.py
@@ -13,6 +13,8 @@
# License for the specific language governing permissions and limitations
# under the License.
+import logging
+
import testtools
import urlparse
@@ -25,6 +27,8 @@
CONF = config.CONF
+LOG = logging.getLogger(__name__)
+
class ServerActionsV3Test(base.BaseV3ComputeTest):
run_ssh = CONF.compute.run_ssh
@@ -260,7 +264,14 @@
# the oldest one should be deleted automatically in this test
def _clean_oldest_backup(oldest_backup):
if oldest_backup_exist:
- self.images_client.delete_image(oldest_backup)
+ try:
+ self.images_client.delete_image(oldest_backup)
+ except exceptions.NotFound:
+ pass
+ else:
+ LOG.warning("Deletion of oldest backup %s should not have "
+ "been successful as it should have been "
+ "deleted during rotation." % oldest_backup)
image1_id = data_utils.parse_image_id(resp['location'])
self.addCleanup(_clean_oldest_backup, image1_id)
diff --git a/tempest/api/compute/v3/servers/test_server_metadata.py b/tempest/api/compute/v3/servers/test_server_metadata.py
index 298cd3c..c5443ee 100644
--- a/tempest/api/compute/v3/servers/test_server_metadata.py
+++ b/tempest/api/compute/v3/servers/test_server_metadata.py
@@ -24,9 +24,7 @@
super(ServerMetadataV3Test, cls).setUpClass()
cls.client = cls.servers_client
cls.quotas = cls.quotas_client
- cls.tenant_id = cls.client.tenant_id
resp, server = cls.create_test_server(meta={}, wait_until='ACTIVE')
-
cls.server_id = server['id']
def setUp(self):
diff --git a/tempest/api/network/base.py b/tempest/api/network/base.py
index cc768fd..087b87a 100644
--- a/tempest/api/network/base.py
+++ b/tempest/api/network/base.py
@@ -149,7 +149,8 @@
return network
@classmethod
- def create_subnet(cls, network, gateway=None, cidr=None, mask_bits=None):
+ def create_subnet(cls, network, gateway=None, cidr=None, mask_bits=None,
+ **kwargs):
"""Wrapper utility that returns a test subnet."""
# The cidr and mask_bits depend on the ip version.
if cls._ip_version == 4:
@@ -168,7 +169,8 @@
network_id=network['id'],
cidr=str(subnet_cidr),
ip_version=cls._ip_version,
- gateway_ip=gateway)
+ gateway_ip=gateway,
+ **kwargs)
break
except exceptions.BadRequest as e:
is_overlapping_cidr = 'overlaps with another subnet' in str(e)
diff --git a/tempest/api/network/test_networks.py b/tempest/api/network/test_networks.py
index ac0fd11..78cb221 100644
--- a/tempest/api/network/test_networks.py
+++ b/tempest/api/network/test_networks.py
@@ -14,6 +14,7 @@
# under the License.
import netaddr
+import testtools
from tempest.api.network import base
from tempest.common.utils import data_utils
@@ -433,6 +434,29 @@
self.assertEqual('204', resp['status'])
self.subnets.pop()
+ @testtools.skipUnless(CONF.network_feature_enabled.ipv6_subnet_attributes,
+ "IPv6 extended attributes for subnets not "
+ "available")
+ @test.attr(type='smoke')
+ def test_create_delete_subnet_with_v6_attributes(self):
+ name = data_utils.rand_name('network-')
+ resp, body = self.client.create_network(name=name)
+ self.assertEqual('201', resp['status'])
+ network = body['network']
+ net_id = network['id']
+ subnet = self.create_subnet(network,
+ gateway='fe80::1',
+ ipv6_ra_mode='slaac',
+ ipv6_address_mode='slaac')
+ # Verifies Subnet GW in IPv6
+ self.assertEqual(subnet['gateway_ip'], 'fe80::1')
+ self.assertEqual(subnet['ipv6_ra_mode'], 'slaac')
+ self.assertEqual(subnet['ipv6_address_mode'], 'slaac')
+ # Delete network and subnet
+ resp, body = self.client.delete_network(net_id)
+ self.assertEqual('204', resp['status'])
+ self.subnets.pop()
+
class NetworksIpV6TestXML(NetworksIpV6TestJSON):
_interface = 'xml'
diff --git a/tempest/api/object_storage/base.py b/tempest/api/object_storage/base.py
index 6b18182..ccc0067 100644
--- a/tempest/api/object_storage/base.py
+++ b/tempest/api/object_storage/base.py
@@ -44,9 +44,6 @@
cls.isolated_creds.get_admin_creds())
# Get isolated creds for alt user
cls.os_alt = clients.Manager(cls.isolated_creds.get_alt_creds())
- # Add isolated users to operator role so that they can create a
- # container in swift.
- cls._assign_member_role()
else:
cls.os = clients.Manager()
cls.os_admin = clients.AdminManager()
@@ -80,22 +77,6 @@
super(BaseObjectTest, cls).tearDownClass()
@classmethod
- def _assign_member_role(cls):
- primary_creds = cls.isolated_creds.get_primary_creds()
- alt_creds = cls.isolated_creds.get_alt_creds()
- swift_role = CONF.object_storage.operator_role
- try:
- resp, roles = cls.os_admin.identity_client.list_roles()
- role = next(r for r in roles if r['name'] == swift_role)
- except StopIteration:
- msg = "No role named %s found" % swift_role
- raise exceptions.NotFound(msg)
- for creds in [primary_creds, alt_creds]:
- cls.os_admin.identity_client.assign_user_role(creds.tenant_id,
- creds.user_id,
- role['id'])
-
- @classmethod
def delete_containers(cls, containers, container_client=None,
object_client=None):
"""Remove given containers and all objects in them.
diff --git a/tempest/api/object_storage/test_object_services.py b/tempest/api/object_storage/test_object_services.py
index 06e63a4..1ef9aa1 100644
--- a/tempest/api/object_storage/test_object_services.py
+++ b/tempest/api/object_storage/test_object_services.py
@@ -13,11 +13,13 @@
# License for the specific language governing permissions and limitations
# under the License.
+import cStringIO as StringIO
import hashlib
import random
import re
from six import moves
import time
+import zlib
from tempest.api.object_storage import base
from tempest.common import custom_matchers
@@ -61,7 +63,7 @@
return object_name, data_segments
- @test.attr(type='smoke')
+ @test.attr(type='gate')
def test_create_object(self):
# create object
object_name = data_utils.rand_name(name='TestObject')
@@ -76,7 +78,242 @@
self.assertEqual(resp['status'], '201')
self.assertHeaders(resp, 'Object', 'PUT')
- @test.attr(type='smoke')
+ # check uploaded content
+ _, body = self.object_client.get_object(self.container_name,
+ object_name)
+ self.assertEqual(data, body)
+
+ @test.attr(type='gate')
+ def test_create_object_with_content_disposition(self):
+ # create object with content_disposition
+ object_name = data_utils.rand_name(name='TestObject')
+ data = data_utils.arbitrary_string()
+ metadata = {}
+ metadata['content-disposition'] = 'inline'
+ resp, _ = self.object_client.create_object(
+ self.container_name,
+ object_name,
+ data,
+ metadata=metadata)
+ self.assertEqual(resp['status'], '201')
+ self.assertHeaders(resp, 'Object', 'PUT')
+
+ resp, body = self.object_client.get_object(
+ self.container_name,
+ object_name,
+ metadata=None)
+ self.assertIn('content-disposition', resp)
+ self.assertEqual(resp['content-disposition'], 'inline')
+ self.assertEqual(body, data)
+
+ @test.attr(type='gate')
+ def test_create_object_with_content_encoding(self):
+ # create object with content_encoding
+ object_name = data_utils.rand_name(name='TestObject')
+
+ # put compressed string
+ data_before = 'x' * 2000
+ data = zlib.compress(data_before)
+ metadata = {}
+ metadata['content-encoding'] = 'deflate'
+
+ resp, _ = self.object_client.create_object(
+ self.container_name,
+ object_name,
+ data,
+ metadata=metadata)
+ self.assertEqual(resp['status'], '201')
+ self.assertHeaders(resp, 'Object', 'PUT')
+
+ # download compressed object
+ metadata = {}
+ metadata['accept-encoding'] = 'deflate'
+ resp, body = self.object_client.get_object(
+ self.container_name,
+ object_name,
+ metadata=metadata)
+ self.assertEqual(body, data_before)
+
+ @test.attr(type='gate')
+ def test_create_object_with_etag(self):
+ # create object with etag
+ object_name = data_utils.rand_name(name='TestObject')
+ data = data_utils.arbitrary_string()
+ md5 = hashlib.md5(data).hexdigest()
+ metadata = {'Etag': md5}
+ resp, _ = self.object_client.create_object(
+ self.container_name,
+ object_name,
+ data,
+ metadata=metadata)
+ self.assertEqual(resp['status'], '201')
+ self.assertHeaders(resp, 'Object', 'PUT')
+
+ # check uploaded content
+ _, body = self.object_client.get_object(self.container_name,
+ object_name)
+ self.assertEqual(data, body)
+
+ @test.attr(type='gate')
+ def test_create_object_with_expect_continue(self):
+ # create object with expect_continue
+ object_name = data_utils.rand_name(name='TestObject')
+ data = data_utils.arbitrary_string()
+ metadata = {'Expect': '100-continue'}
+ resp = self.custom_object_client.create_object_continue(
+ self.container_name,
+ object_name,
+ data,
+ metadata=metadata)
+
+ self.assertIn('status', resp)
+ self.assertEqual(resp['status'], '100')
+
+ self.custom_object_client.create_object_continue(
+ self.container_name,
+ object_name,
+ data,
+ metadata=None)
+
+ # check uploaded content
+ _, body = self.object_client.get_object(self.container_name,
+ object_name)
+ self.assertEqual(data, body)
+
+ @test.attr(type='gate')
+ def test_create_object_with_transfer_encoding(self):
+ # create object with transfer_encoding
+ object_name = data_utils.rand_name(name='TestObject')
+ data = data_utils.arbitrary_string(1024)
+ status, _, resp_headers = self.object_client.put_object_with_chunk(
+ container=self.container_name,
+ name=object_name,
+ contents=StringIO.StringIO(data),
+ chunk_size=512)
+ self.assertEqual(status, 201)
+ self.assertHeaders(resp_headers, 'Object', 'PUT')
+
+ # check uploaded content
+ _, body = self.object_client.get_object(self.container_name,
+ object_name)
+ self.assertEqual(data, body)
+
+ @test.attr(type='gate')
+ def test_create_object_with_x_fresh_metadata(self):
+ # create object with x_fresh_metadata
+ object_name_base = data_utils.rand_name(name='TestObject')
+ data = data_utils.arbitrary_string()
+ metadata_1 = {'X-Object-Meta-test-meta': 'Meta'}
+ self.object_client.create_object(self.container_name,
+ object_name_base,
+ data,
+ metadata=metadata_1)
+ object_name = data_utils.rand_name(name='TestObject')
+ metadata_2 = {'X-Copy-From': '%s/%s' % (self.container_name,
+ object_name_base),
+ 'X-Fresh-Metadata': 'true'}
+ resp, _ = self.object_client.create_object(
+ self.container_name,
+ object_name,
+ '',
+ metadata=metadata_2)
+ self.assertEqual(resp['status'], '201')
+ self.assertHeaders(resp, 'Object', 'PUT')
+
+ resp, body = self.object_client.get_object(self.container_name,
+ object_name)
+ self.assertNotIn('x-object-meta-test-meta', resp)
+ self.assertEqual(data, body)
+
+ @test.attr(type='gate')
+ def test_create_object_with_x_object_meta(self):
+ # create object with object_meta
+ object_name = data_utils.rand_name(name='TestObject')
+ data = data_utils.arbitrary_string()
+ metadata = {'X-Object-Meta-test-meta': 'Meta'}
+ resp, _ = self.object_client.create_object(
+ self.container_name,
+ object_name,
+ data,
+ metadata=metadata)
+ self.assertEqual(resp['status'], '201')
+ self.assertHeaders(resp, 'Object', 'PUT')
+
+ resp, body = self.object_client.get_object(self.container_name,
+ object_name)
+ self.assertIn('x-object-meta-test-meta', resp)
+ self.assertEqual(resp['x-object-meta-test-meta'], 'Meta')
+ self.assertEqual(data, body)
+
+ @test.attr(type='gate')
+ def test_create_object_with_x_object_metakey(self):
+ # create object with the blank value of metadata
+ object_name = data_utils.rand_name(name='TestObject')
+ data = data_utils.arbitrary_string()
+ metadata = {'X-Object-Meta-test-meta': ''}
+ resp, _ = self.object_client.create_object(
+ self.container_name,
+ object_name,
+ data,
+ metadata=metadata)
+ self.assertEqual(resp['status'], '201')
+ self.assertHeaders(resp, 'Object', 'PUT')
+
+ resp, body = self.object_client.get_object(self.container_name,
+ object_name)
+ self.assertIn('x-object-meta-test-meta', resp)
+ self.assertEqual(resp['x-object-meta-test-meta'], '')
+ self.assertEqual(data, body)
+
+ @test.attr(type='gate')
+ def test_create_object_with_x_remove_object_meta(self):
+ # create object with x_remove_object_meta
+ object_name = data_utils.rand_name(name='TestObject')
+ data = data_utils.arbitrary_string()
+ metadata_add = {'X-Object-Meta-test-meta': 'Meta'}
+ self.object_client.create_object(self.container_name,
+ object_name,
+ data,
+ metadata=metadata_add)
+ metadata_remove = {'X-Remove-Object-Meta-test-meta': 'Meta'}
+ resp, _ = self.object_client.create_object(
+ self.container_name,
+ object_name,
+ data,
+ metadata=metadata_remove)
+ self.assertEqual(resp['status'], '201')
+ self.assertHeaders(resp, 'Object', 'PUT')
+
+ resp, body = self.object_client.get_object(self.container_name,
+ object_name)
+ self.assertNotIn('x-object-meta-test-meta', resp)
+ self.assertEqual(data, body)
+
+ @test.attr(type='gate')
+ def test_create_object_with_x_remove_object_metakey(self):
+ # create object with the blank value of remove metadata
+ object_name = data_utils.rand_name(name='TestObject')
+ data = data_utils.arbitrary_string()
+ metadata_add = {'X-Object-Meta-test-meta': 'Meta'}
+ self.object_client.create_object(self.container_name,
+ object_name,
+ data,
+ metadata=metadata_add)
+ metadata_remove = {'X-Remove-Object-Meta-test-meta': ''}
+ resp, _ = self.object_client.create_object(
+ self.container_name,
+ object_name,
+ data,
+ metadata=metadata_remove)
+ self.assertEqual(resp['status'], '201')
+ self.assertHeaders(resp, 'Object', 'PUT')
+
+ resp, body = self.object_client.get_object(self.container_name,
+ object_name)
+ self.assertNotIn('x-object-meta-test-meta', resp)
+ self.assertEqual(data, body)
+
+ @test.attr(type='gate')
def test_delete_object(self):
# create object
object_name = data_utils.rand_name(name='TestObject')
diff --git a/tempest/api/queuing/base.py b/tempest/api/queuing/base.py
index 5649619..f4ff7f1 100644
--- a/tempest/api/queuing/base.py
+++ b/tempest/api/queuing/base.py
@@ -13,6 +13,7 @@
# See the License for the specific language governing permissions and
# limitations under the License.
+from tempest.common.utils import data_utils
from tempest import config
from tempest.openstack.common import log as logging
from tempest import test
@@ -89,3 +90,80 @@
"""Wrapper utility that sets the metadata of a queue."""
resp, body = cls.client.set_queue_metadata(queue_name, rbody)
return resp, body
+
+ @classmethod
+ def post_messages(cls, queue_name, rbody):
+ '''Wrapper utility that posts messages to a queue.'''
+ resp, body = cls.client.post_messages(queue_name, rbody)
+
+ return resp, body
+
+ @classmethod
+ def list_messages(cls, queue_name):
+ '''Wrapper utility that lists the messages in a queue.'''
+ resp, body = cls.client.list_messages(queue_name)
+
+ return resp, body
+
+ @classmethod
+ def get_single_message(cls, message_uri):
+ '''Wrapper utility that gets a single message.'''
+ resp, body = cls.client.get_single_message(message_uri)
+
+ return resp, body
+
+ @classmethod
+ def get_multiple_messages(cls, message_uri):
+ '''Wrapper utility that gets multiple messages.'''
+ resp, body = cls.client.get_multiple_messages(message_uri)
+
+ return resp, body
+
+ @classmethod
+ def delete_messages(cls, message_uri):
+ '''Wrapper utility that deletes messages.'''
+ resp, body = cls.client.delete_messages(message_uri)
+
+ return resp, body
+
+ @classmethod
+ def post_claims(cls, queue_name, rbody, url_params=False):
+ '''Wrapper utility that claims messages.'''
+ resp, body = cls.client.post_claims(
+ queue_name, rbody, url_params=False)
+
+ return resp, body
+
+ @classmethod
+ def query_claim(cls, claim_uri):
+ '''Wrapper utility that gets a claim.'''
+ resp, body = cls.client.query_claim(claim_uri)
+
+ return resp, body
+
+ @classmethod
+ def update_claim(cls, claim_uri, rbody):
+ '''Wrapper utility that updates a claim.'''
+ resp, body = cls.client.update_claim(claim_uri, rbody)
+
+ return resp, body
+
+ @classmethod
+ def release_claim(cls, claim_uri):
+ '''Wrapper utility that deletes a claim.'''
+ resp, body = cls.client.release_claim(claim_uri)
+
+ return resp, body
+
+ @classmethod
+ def generate_message_body(cls, repeat=1):
+ '''Wrapper utility that sets the metadata of a queue.'''
+ message_ttl = data_utils.rand_int_id(start=60,
+ end=CONF.queuing.max_message_ttl)
+
+ key = data_utils.arbitrary_string(size=20, base_text='QueuingKey')
+ value = data_utils.arbitrary_string(size=20, base_text='QueuingValue')
+ message_body = {key: value}
+
+ rbody = ([{'body': message_body, 'ttl': message_ttl}] * repeat)
+ return rbody
diff --git a/tempest/api/queuing/test_claims.py b/tempest/api/queuing/test_claims.py
new file mode 100644
index 0000000..a306623
--- /dev/null
+++ b/tempest/api/queuing/test_claims.py
@@ -0,0 +1,123 @@
+# Copyright (c) 2014 Rackspace, Inc.
+#
+# Licensed under the Apache License, Version 2.0 (the "License");
+# you may not use this file except in compliance with the License.
+# You may obtain a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
+# implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+
+import logging
+import urlparse
+
+from tempest.api.queuing import base
+from tempest.common.utils import data_utils
+from tempest import config
+from tempest import test
+
+
+LOG = logging.getLogger(__name__)
+CONF = config.CONF
+
+
+class TestClaims(base.BaseQueuingTest):
+ _interface = 'json'
+
+ @classmethod
+ def setUpClass(cls):
+ super(TestClaims, cls).setUpClass()
+ cls.queue_name = data_utils.rand_name('Queues-Test')
+ # Create Queue
+ cls.create_queue(cls.queue_name)
+
+ def _post_and_claim_messages(self, queue_name, repeat=1):
+ # Post Messages
+ message_body = self.generate_message_body(repeat=repeat)
+ self.client.post_messages(queue_name=self.queue_name,
+ rbody=message_body)
+
+ # Post Claim
+ claim_ttl = data_utils.rand_int_id(start=60,
+ end=CONF.queuing.max_claim_ttl)
+ claim_grace = data_utils.rand_int_id(start=60,
+ end=CONF.queuing.max_claim_grace)
+ claim_body = {"ttl": claim_ttl, "grace": claim_grace}
+ resp, body = self.client.post_claims(queue_name=self.queue_name,
+ rbody=claim_body)
+
+ return resp, body
+
+ @test.attr(type='smoke')
+ def test_post_claim(self):
+ _, body = self._post_and_claim_messages(queue_name=self.queue_name)
+ claimed_message_uri = body[0]['href']
+
+ # Skipping this step till bug-1331517 is fixed
+ # Get posted claim
+ # self.client.query_claim(claimed_message_uri)
+
+ # Delete Claimed message
+ self.client.delete_messages(claimed_message_uri)
+
+ @test.skip_because(bug="1331517")
+ @test.attr(type='smoke')
+ def test_query_claim(self):
+ # Post a Claim
+ resp, body = self._post_and_claim_messages(queue_name=self.queue_name)
+
+ # Query Claim
+ claim_uri = resp['location']
+ self.client.query_claim(claim_uri)
+
+ # Delete Claimed message
+ claimed_message_uri = body[0]['href']
+ self.delete_messages(claimed_message_uri)
+
+ @test.skip_because(bug="1328111")
+ @test.attr(type='smoke')
+ def test_update_claim(self):
+ # Post a Claim
+ resp, body = self._post_and_claim_messages(queue_name=self.queue_name)
+
+ claim_uri = resp['location']
+ claimed_message_uri = body[0]['href']
+
+ # Update Claim
+ claim_ttl = data_utils.rand_int_id(start=60,
+ end=CONF.queuing.max_claim_ttl)
+ update_rbody = {"ttl": claim_ttl}
+
+ self.client.update_claim(claim_uri, rbody=update_rbody)
+
+ # Verify claim ttl >= updated ttl value
+ _, body = self.client.query_claim(claim_uri)
+ updated_claim_ttl = body["ttl"]
+ self.assertTrue(updated_claim_ttl >= claim_ttl)
+
+ # Delete Claimed message
+ self.client.delete_messages(claimed_message_uri)
+
+ @test.attr(type='smoke')
+ def test_release_claim(self):
+ # Post a Claim
+ resp, body = self._post_and_claim_messages(queue_name=self.queue_name)
+ claim_uri = resp['location']
+
+ # Release Claim
+ self.client.release_claim(claim_uri)
+
+ # Delete Claimed message
+ # This will implicitly verify that the claim is deleted.
+ message_uri = urlparse.urlparse(claim_uri).path
+ self.client.delete_messages(message_uri)
+
+ @classmethod
+ def tearDownClass(cls):
+ cls.delete_queue(cls.queue_name)
+ super(TestClaims, cls).tearDownClass()
diff --git a/tempest/api/queuing/test_messages.py b/tempest/api/queuing/test_messages.py
new file mode 100644
index 0000000..9546c91
--- /dev/null
+++ b/tempest/api/queuing/test_messages.py
@@ -0,0 +1,122 @@
+# Copyright (c) 2014 Rackspace, Inc.
+#
+# Licensed under the Apache License, Version 2.0 (the "License");
+# you may not use this file except in compliance with the License.
+# You may obtain a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
+# implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+
+import logging
+
+from tempest.api.queuing import base
+from tempest.common.utils import data_utils
+from tempest import config
+from tempest import test
+
+
+LOG = logging.getLogger(__name__)
+CONF = config.CONF
+
+
+class TestMessages(base.BaseQueuingTest):
+ _interface = 'json'
+
+ @classmethod
+ def setUpClass(cls):
+ super(TestMessages, cls).setUpClass()
+ cls.queue_name = data_utils.rand_name('Queues-Test')
+ # Create Queue
+ cls.client.create_queue(cls.queue_name)
+
+ def _post_messages(self, repeat=CONF.queuing.max_messages_per_page):
+ message_body = self.generate_message_body(repeat=repeat)
+ resp, body = self.post_messages(queue_name=self.queue_name,
+ rbody=message_body)
+ return resp, body
+
+ @test.attr(type='smoke')
+ def test_post_messages(self):
+ # Post Messages
+ resp, _ = self._post_messages()
+
+ # Get on the posted messages
+ message_uri = resp['location']
+ resp, _ = self.client.get_multiple_messages(message_uri)
+ # The test has an assertion here, because the response cannot be 204
+ # in this case (the client allows 200 or 204 for this API call).
+ self.assertEqual('200', resp['status'])
+
+ @test.attr(type='smoke')
+ def test_list_messages(self):
+ # Post Messages
+ self._post_messages()
+
+ # List Messages
+ resp, _ = self.list_messages(queue_name=self.queue_name)
+ # The test has an assertion here, because the response cannot be 204
+ # in this case (the client allows 200 or 204 for this API call).
+ self.assertEqual('200', resp['status'])
+
+ @test.attr(type='smoke')
+ def test_get_message(self):
+ # Post Messages
+ _, body = self._post_messages()
+ message_uri = body['resources'][0]
+
+ # Get posted message
+ resp, _ = self.client.get_single_message(message_uri)
+ # The test has an assertion here, because the response cannot be 204
+ # in this case (the client allows 200 or 204 for this API call).
+ self.assertEqual('200', resp['status'])
+
+ @test.attr(type='smoke')
+ def test_get_multiple_messages(self):
+ # Post Messages
+ resp, _ = self._post_messages()
+ message_uri = resp['location']
+
+ # Get posted messages
+ resp, _ = self.client.get_multiple_messages(message_uri)
+ # The test has an assertion here, because the response cannot be 204
+ # in this case (the client allows 200 or 204 for this API call).
+ self.assertEqual('200', resp['status'])
+
+ @test.attr(type='smoke')
+ def test_delete_single_message(self):
+ # Post Messages
+ _, body = self._post_messages()
+ message_uri = body['resources'][0]
+
+ # Delete posted message & verify the delete operration
+ self.client.delete_messages(message_uri)
+
+ message_uri = message_uri.replace('/messages/', '/messages?ids=')
+ resp, _ = self.client.get_multiple_messages(message_uri)
+ # The test has an assertion here, because the response has to be 204
+ # in this case (the client allows 200 or 204 for this API call).
+ self.assertEqual('204', resp['status'])
+
+ @test.attr(type='smoke')
+ def test_delete_multiple_messages(self):
+ # Post Messages
+ resp, _ = self._post_messages()
+ message_uri = resp['location']
+
+ # Delete multiple messages
+ self.client.delete_messages(message_uri)
+ resp, _ = self.client.get_multiple_messages(message_uri)
+ # The test has an assertion here, because the response has to be 204
+ # in this case (the client allows 200 or 204 for this API call).
+ self.assertEqual('204', resp['status'])
+
+ @classmethod
+ def tearDownClass(cls):
+ cls.delete_queue(cls.queue_name)
+ super(TestMessages, cls).tearDownClass()
diff --git a/tempest/api_schema/queuing/v1/queues.py b/tempest/api_schema/queuing/v1/queues.py
index 4630e1c..f0b2691 100644
--- a/tempest/api_schema/queuing/v1/queues.py
+++ b/tempest/api_schema/queuing/v1/queues.py
@@ -14,6 +14,7 @@
# See the License for the specific language governing permissions and
# limitations under the License.
+
list_link = {
'type': 'object',
'properties': {
@@ -58,6 +59,11 @@
}
}
+age = {
+ 'type': 'number',
+ 'minimum': 0
+}
+
message_link = {
'type': 'object',
'properties': {
@@ -65,7 +71,7 @@
'type': 'string',
'format': 'uri'
},
- 'age': {'type': 'number'},
+ 'age': age,
'created': {
'type': 'string',
'format': 'date-time'
@@ -96,3 +102,136 @@
'required': ['messages']
}
}
+
+resource_schema = {
+ 'type': 'array',
+ 'items': 'string',
+ 'minItems': 1
+}
+
+post_messages = {
+ 'status_code': [201],
+ 'response_body': {
+ 'type': 'object',
+ 'properties': {
+ 'resources': resource_schema,
+ 'partial': {'type': 'boolean'}
+ }
+ },
+ 'required': ['resources', 'partial']
+}
+
+message_ttl = {
+ 'type': 'number',
+ 'minimum': 1
+}
+
+list_messages_links = {
+ 'type': 'array',
+ 'maxItems': 1,
+ 'minItems': 1,
+ 'items': {
+ 'type': 'object',
+ 'properties': {
+ 'rel': {'type': 'string'},
+ 'href': {'type': 'string'}
+ },
+ 'required': ['rel', 'href']
+ }
+}
+
+list_messages_response = {
+ 'type': 'array',
+ 'minItems': 1,
+ 'items': {
+ 'type': 'object',
+ 'properties': {
+ 'href': {'type': 'string'},
+ 'ttl': message_ttl,
+ 'age': age,
+ 'body': {'type': 'object'}
+ },
+ 'required': ['href', 'ttl', 'age', 'body']
+ }
+}
+
+list_messages = {
+ 'status_code': [200, 204],
+ 'response_body': {
+ 'type': 'object',
+ 'properties': {
+ 'links': list_messages_links,
+ 'messages': list_messages_response
+ }
+ },
+ 'required': ['links', 'messages']
+}
+
+single_message = {
+ 'type': 'object',
+ 'properties': {
+ 'href': {'type': 'string'},
+ 'ttl': message_ttl,
+ 'age': age,
+ 'body': {'type': 'object'}
+ },
+ 'required': ['href', 'ttl', 'age', 'body']
+}
+
+get_single_message = {
+ 'status_code': [200],
+ 'response_body': single_message
+}
+
+get_multiple_messages = {
+ 'status_code': [200],
+ 'response_body': {
+ 'type': 'array',
+ 'items': single_message,
+ 'minItems': 1
+ }
+}
+
+messages_claimed = {
+ 'type': 'object',
+ 'properties': {
+ 'href': {
+ 'type': 'string',
+ 'format': 'uri'
+ },
+ 'ttl': message_ttl,
+ 'age': {'type': 'number'},
+ 'body': {'type': 'object'}
+ },
+ 'required': ['href', 'ttl', 'age', 'body']
+}
+
+claim_messages = {
+ 'status_code': [201, 204],
+ 'response_body': {
+ 'type': 'array',
+ 'items': messages_claimed,
+ 'minItems': 1
+ }
+}
+
+claim_ttl = {
+ 'type': 'number',
+ 'minimum': 1
+}
+
+query_claim = {
+ 'status_code': [200],
+ 'response_body': {
+ 'type': 'object',
+ 'properties': {
+ 'age': {'type': 'number'},
+ 'ttl': claim_ttl,
+ 'messages': {
+ 'type': 'array',
+ 'minItems': 1
+ }
+ },
+ 'required': ['ttl', 'age', 'messages']
+ }
+}
diff --git a/tempest/auth.py b/tempest/auth.py
index 9c51edb..830dca9 100644
--- a/tempest/auth.py
+++ b/tempest/auth.py
@@ -213,7 +213,7 @@
# build authenticated request
# returns new request, it does not touch the original values
_headers = copy.deepcopy(headers) if headers is not None else {}
- _headers['X-Auth-Token'] = token
+ _headers['X-Auth-Token'] = str(token)
if url is None or url == "":
_url = base_url
else:
@@ -223,7 +223,7 @@
parts[2] = re.sub("/{2,}", "/", parts[2])
_url = urlparse.urlunparse(parts)
# no change to method or body
- return _url, _headers, body
+ return str(_url), _headers, body
def _auth_client(self):
raise NotImplementedError
diff --git a/tempest/cmd/javelin.py b/tempest/cmd/javelin.py
index 20ee63e..1d46028 100755
--- a/tempest/cmd/javelin.py
+++ b/tempest/cmd/javelin.py
@@ -35,6 +35,7 @@
from tempest.services.image.v2.json import image_client
from tempest.services.object_storage import container_client
from tempest.services.object_storage import object_client
+from tempest.services.volume.json import volumes_client
OPTS = {}
USERS = {}
@@ -60,6 +61,7 @@
self.containers = container_client.ContainerClient(_auth)
self.images = image_client.ImageClientV2JSON(_auth)
self.flavors = flavors_client.FlavorsClientJSON(_auth)
+ self.volumes = volumes_client.VolumesClientJSON(_auth)
def load_resources(fname):
@@ -190,6 +192,7 @@
self.check_users()
self.check_objects()
self.check_servers()
+ self.check_volumes()
def check_users(self):
"""Check that the users we expect to exist, do.
@@ -235,6 +238,21 @@
"Server %s is not pingable at %s" % (
server['name'], addr))
+ def check_volumes(self):
+ """Check that the volumes are still there and attached."""
+ for volume in self.res['volumes']:
+ client = client_for_user(volume['owner'])
+ found = _get_volume_by_name(client, volume['name'])
+ self.assertIsNotNone(
+ found,
+ "Couldn't find expected volume %s" % volume['name'])
+
+ # Verify that a volume's attachment retrieved
+ server_id = _get_server_by_name(client, volume['server'])['id']
+ attachment = self.client.get_attachment_from_volume(volume)
+ self.assertEqual(volume['id'], attachment['volume_id'])
+ self.assertEqual(server_id, attachment['server_id'])
+
#######################
#
@@ -339,6 +357,40 @@
#######################
#
+# VOLUMES
+#
+#######################
+
+def _get_volume_by_name(client, name):
+ r, body = client.volumes.list_volumes()
+ for volume in body['volumes']:
+ if name == volume['name']:
+ return volume
+ return None
+
+
+def create_volumes(volumes):
+ for volume in volumes:
+ client = client_for_user(volume['owner'])
+
+ # only create a volume if the name isn't here
+ r, body = client.volumes.list_volumes()
+ if any(item['name'] == volume['name'] for item in body):
+ continue
+
+ client.volumes.create_volume(volume['name'], volume['size'])
+
+
+def attach_volumes(volumes):
+ for volume in volumes:
+ client = client_for_user(volume['owner'])
+
+ server_id = _get_server_by_name(client, volume['server'])['id']
+ client.volumes.attach_volume(volume['name'], server_id)
+
+
+#######################
+#
# MAIN LOGIC
#
#######################
@@ -355,6 +407,8 @@
create_objects(RES['objects'])
create_images(RES['images'])
create_servers(RES['servers'])
+ create_volumes(RES['volumes'])
+ attach_volumes(RES['volumes'])
def get_options():
diff --git a/tempest/cmd/resources.yaml b/tempest/cmd/resources.yaml
index f7cb8a9..a1f567b 100644
--- a/tempest/cmd/resources.yaml
+++ b/tempest/cmd/resources.yaml
@@ -32,9 +32,14 @@
aki: cirros-0.3.2-x86_64-vmlinuz
ari: cirros-0.3.2-x86_64-initrd
volumes:
- - assegai:
- - owner: javelin
- - gb: 1
+ - name: assegai
+ server: peltast
+ owner: javelin
+ size: 1
+ - name: pifpouf
+ server: hoplite
+ owner: javelin
+ size: 2
servers:
- name: peltast
owner: javelin
diff --git a/tempest/common/isolated_creds.py b/tempest/common/isolated_creds.py
index b4618ed..8166de5 100644
--- a/tempest/common/isolated_creds.py
+++ b/tempest/common/isolated_creds.py
@@ -106,12 +106,23 @@
roles = self.identity_admin_client.roles.list()
return roles
- def _assign_user_role(self, tenant, user, role):
+ def _assign_user_role(self, tenant, user, role_name):
+ role = None
+ try:
+ roles = self._list_roles()
+ if self.tempest_client:
+ role = next(r for r in roles if r['name'] == role_name)
+ else:
+ role = next(r for r in roles if r.name == role_name)
+ except StopIteration:
+ msg = 'No "%s" role found' % role_name
+ raise exceptions.NotFound(msg)
if self.tempest_client:
- self.identity_admin_client.assign_user_role(tenant, user, role)
+ self.identity_admin_client.assign_user_role(tenant['id'],
+ user['id'], role['id'])
else:
- self.identity_admin_client.roles.add_user_role(user,
- role, tenant=tenant)
+ self.identity_admin_client.roles.add_user_role(user.id, role.id,
+ tenant.id)
def _delete_user(self, user):
if self.tempest_client:
@@ -149,22 +160,11 @@
email = data_utils.rand_name(root) + suffix + "@example.com"
user = self._create_user(username, self.password,
tenant, email)
+ # NOTE(andrey-mp): user needs this role to create containers in swift
+ swift_operator_role = CONF.object_storage.operator_role
+ self._assign_user_role(tenant, user, swift_operator_role)
if admin:
- role = None
- try:
- roles = self._list_roles()
- admin_role = CONF.identity.admin_role
- if self.tempest_client:
- role = next(r for r in roles if r['name'] == admin_role)
- else:
- role = next(r for r in roles if r.name == admin_role)
- except StopIteration:
- msg = "No admin role found"
- raise exceptions.NotFound(msg)
- if self.tempest_client:
- self._assign_user_role(tenant['id'], user['id'], role['id'])
- else:
- self._assign_user_role(tenant.id, user.id, role.id)
+ self._assign_user_role(tenant, user, CONF.identity.admin_role)
return self._get_credentials(user, tenant)
def _get_credentials(self, user, tenant):
diff --git a/tempest/config.py b/tempest/config.py
index 7d871cb..6475844 100644
--- a/tempest/config.py
+++ b/tempest/config.py
@@ -442,6 +442,12 @@
default=['all'],
help='A list of enabled network extensions with a special '
'entry all which indicates every extension is enabled'),
+ cfg.BoolOpt('ipv6_subnet_attributes',
+ default=False,
+ help="Allow the execution of IPv6 subnet tests that use "
+ "the extended IPv6 attributes ipv6_ra_mode "
+ "and ipv6_address_mode"
+ )
]
queuing_group = cfg.OptGroup(name='queuing',
@@ -455,6 +461,28 @@
default=20,
help='The maximum number of queue records per page when '
'listing queues'),
+ cfg.IntOpt('max_queue_metadata',
+ default=65536,
+ help='The maximum metadata size for a queue'),
+ cfg.IntOpt('max_messages_per_page',
+ default=20,
+ help='The maximum number of queue message per page when '
+ 'listing (or) posting messages'),
+ cfg.IntOpt('max_message_size',
+ default=262144,
+ help='The maximum size of a message body'),
+ cfg.IntOpt('max_messages_per_claim',
+ default=20,
+ help='The maximum number of messages per claim'),
+ cfg.IntOpt('max_message_ttl',
+ default=1209600,
+ help='The maximum ttl for a message'),
+ cfg.IntOpt('max_claim_ttl',
+ default=43200,
+ help='The maximum ttl for a claim'),
+ cfg.IntOpt('max_claim_grace',
+ default=43200,
+ help='The maximum grace period for a claim'),
]
volume_group = cfg.OptGroup(name='volume',
diff --git a/tempest/openstack/common/config/generator.py b/tempest/openstack/common/config/generator.py
index 8156cc5..664200e 100644
--- a/tempest/openstack/common/config/generator.py
+++ b/tempest/openstack/common/config/generator.py
@@ -150,7 +150,7 @@
def _is_in_group(opt, group):
- "Check if opt is in group."
+ """Check if opt is in group."""
for value in group._opts.values():
# NOTE(llu): Temporary workaround for bug #1262148, wait until
# newly released oslo.config support '==' operator.
@@ -159,7 +159,7 @@
return False
-def _guess_groups(opt, mod_obj):
+def _guess_groups(opt):
# is it in the DEFAULT group?
if _is_in_group(opt, cfg.CONF):
return 'DEFAULT'
@@ -193,7 +193,7 @@
ret = {}
for opt in opts:
- ret.setdefault(_guess_groups(opt, obj), []).append(opt)
+ ret.setdefault(_guess_groups(opt), []).append(opt)
return ret.items()
@@ -223,6 +223,8 @@
def _sanitize_default(name, value):
"""Set up a reasonably sensible default for pybasedir, my_ip and host."""
+ hostname = socket.gethostname()
+ fqdn = socket.getfqdn()
if value.startswith(sys.prefix):
# NOTE(jd) Don't use os.path.join, because it is likely to think the
# second part is an absolute pathname and therefore drop the first
@@ -234,8 +236,13 @@
return value.replace(BASEDIR, '')
elif value == _get_my_ip():
return '10.0.0.1'
- elif value in (socket.gethostname(), socket.getfqdn()) and 'host' in name:
- return 'tempest'
+ elif value in (hostname, fqdn):
+ if 'host' in name:
+ return 'tempest'
+ elif value.endswith(hostname):
+ return value.replace(hostname, 'tempest')
+ elif value.endswith(fqdn):
+ return value.replace(fqdn, 'tempest')
elif value.strip() != value:
return '"%s"' % value
return value
@@ -246,7 +253,6 @@
if not opt_help:
sys.stderr.write('WARNING: "%s" is missing help string.\n' % opt_name)
opt_help = ""
- opt_type = None
try:
opt_type = OPTION_REGEX.search(str(type(opt))).group(0)
except (ValueError, AttributeError) as err:
diff --git a/tempest/openstack/common/gettextutils.py b/tempest/openstack/common/gettextutils.py
index 17f66f7..6102e67 100644
--- a/tempest/openstack/common/gettextutils.py
+++ b/tempest/openstack/common/gettextutils.py
@@ -32,24 +32,113 @@
from babel import localedata
import six
-_localedir = os.environ.get('tempest'.upper() + '_LOCALEDIR')
-_t = gettext.translation('tempest', localedir=_localedir, fallback=True)
-
-# We use separate translation catalogs for each log level, so set up a
-# mapping between the log level name and the translator. The domain
-# for the log level is project_name + "-log-" + log_level so messages
-# for each level end up in their own catalog.
-_t_log_levels = dict(
- (level, gettext.translation('tempest' + '-log-' + level,
- localedir=_localedir,
- fallback=True))
- for level in ['info', 'warning', 'error', 'critical']
-)
-
_AVAILABLE_LANGUAGES = {}
+
+# FIXME(dhellmann): Remove this when moving to oslo.i18n.
USE_LAZY = False
+class TranslatorFactory(object):
+ """Create translator functions
+ """
+
+ def __init__(self, domain, lazy=False, localedir=None):
+ """Establish a set of translation functions for the domain.
+
+ :param domain: Name of translation domain,
+ specifying a message catalog.
+ :type domain: str
+ :param lazy: Delays translation until a message is emitted.
+ Defaults to False.
+ :type lazy: Boolean
+ :param localedir: Directory with translation catalogs.
+ :type localedir: str
+ """
+ self.domain = domain
+ self.lazy = lazy
+ if localedir is None:
+ localedir = os.environ.get(domain.upper() + '_LOCALEDIR')
+ self.localedir = localedir
+
+ def _make_translation_func(self, domain=None):
+ """Return a new translation function ready for use.
+
+ Takes into account whether or not lazy translation is being
+ done.
+
+ The domain can be specified to override the default from the
+ factory, but the localedir from the factory is always used
+ because we assume the log-level translation catalogs are
+ installed in the same directory as the main application
+ catalog.
+
+ """
+ if domain is None:
+ domain = self.domain
+ if self.lazy:
+ return functools.partial(Message, domain=domain)
+ t = gettext.translation(
+ domain,
+ localedir=self.localedir,
+ fallback=True,
+ )
+ if six.PY3:
+ return t.gettext
+ return t.ugettext
+
+ @property
+ def primary(self):
+ "The default translation function."
+ return self._make_translation_func()
+
+ def _make_log_translation_func(self, level):
+ return self._make_translation_func(self.domain + '-log-' + level)
+
+ @property
+ def log_info(self):
+ "Translate info-level log messages."
+ return self._make_log_translation_func('info')
+
+ @property
+ def log_warning(self):
+ "Translate warning-level log messages."
+ return self._make_log_translation_func('warning')
+
+ @property
+ def log_error(self):
+ "Translate error-level log messages."
+ return self._make_log_translation_func('error')
+
+ @property
+ def log_critical(self):
+ "Translate critical-level log messages."
+ return self._make_log_translation_func('critical')
+
+
+# NOTE(dhellmann): When this module moves out of the incubator into
+# oslo.i18n, these global variables can be moved to an integration
+# module within each application.
+
+# Create the global translation functions.
+_translators = TranslatorFactory('tempest')
+
+# The primary translation function using the well-known name "_"
+_ = _translators.primary
+
+# Translators for log levels.
+#
+# The abbreviated names are meant to reflect the usual use of a short
+# name like '_'. The "L" is for "log" and the other letter comes from
+# the level.
+_LI = _translators.log_info
+_LW = _translators.log_warning
+_LE = _translators.log_error
+_LC = _translators.log_critical
+
+# NOTE(dhellmann): End of globals that will move to the application's
+# integration module.
+
+
def enable_lazy():
"""Convenience function for configuring _() to use lazy gettext
@@ -58,41 +147,18 @@
your project is importing _ directly instead of using the
gettextutils.install() way of importing the _ function.
"""
- global USE_LAZY
+ # FIXME(dhellmann): This function will be removed in oslo.i18n,
+ # because the TranslatorFactory makes it superfluous.
+ global _, _LI, _LW, _LE, _LC, USE_LAZY
+ tf = TranslatorFactory('tempest', lazy=True)
+ _ = tf.primary
+ _LI = tf.log_info
+ _LW = tf.log_warning
+ _LE = tf.log_error
+ _LC = tf.log_critical
USE_LAZY = True
-def _(msg):
- if USE_LAZY:
- return Message(msg, domain='tempest')
- else:
- if six.PY3:
- return _t.gettext(msg)
- return _t.ugettext(msg)
-
-
-def _log_translation(msg, level):
- """Build a single translation of a log message
- """
- if USE_LAZY:
- return Message(msg, domain='tempest' + '-log-' + level)
- else:
- translator = _t_log_levels[level]
- if six.PY3:
- return translator.gettext(msg)
- return translator.ugettext(msg)
-
-# Translators for log levels.
-#
-# The abbreviated names are meant to reflect the usual use of a short
-# name like '_'. The "L" is for "log" and the other letter comes from
-# the level.
-_LI = functools.partial(_log_translation, level='info')
-_LW = functools.partial(_log_translation, level='warning')
-_LE = functools.partial(_log_translation, level='error')
-_LC = functools.partial(_log_translation, level='critical')
-
-
def install(domain, lazy=False):
"""Install a _() function using the given translation domain.
@@ -112,26 +178,9 @@
any available locale.
"""
if lazy:
- # NOTE(mrodden): Lazy gettext functionality.
- #
- # The following introduces a deferred way to do translations on
- # messages in OpenStack. We override the standard _() function
- # and % (format string) operation to build Message objects that can
- # later be translated when we have more information.
- def _lazy_gettext(msg):
- """Create and return a Message object.
-
- Lazy gettext function for a given domain, it is a factory method
- for a project/module to get a lazy gettext function for its own
- translation domain (i.e. nova, glance, cinder, etc.)
-
- Message encapsulates a string so that we can translate
- it later when needed.
- """
- return Message(msg, domain=domain)
-
from six import moves
- moves.builtins.__dict__['_'] = _lazy_gettext
+ tf = TranslatorFactory(domain, lazy=True)
+ moves.builtins.__dict__['_'] = tf.primary
else:
localedir = '%s_LOCALEDIR' % domain.upper()
if six.PY3:
@@ -274,13 +323,14 @@
def __radd__(self, other):
return self.__add__(other)
- def __str__(self):
- # NOTE(luisg): Logging in python 2.6 tries to str() log records,
- # and it expects specifically a UnicodeError in order to proceed.
- msg = _('Message objects do not support str() because they may '
- 'contain non-ascii characters. '
- 'Please use unicode() or translate() instead.')
- raise UnicodeError(msg)
+ if six.PY2:
+ def __str__(self):
+ # NOTE(luisg): Logging in python 2.6 tries to str() log records,
+ # and it expects specifically a UnicodeError in order to proceed.
+ msg = _('Message objects do not support str() because they may '
+ 'contain non-ascii characters. '
+ 'Please use unicode() or translate() instead.')
+ raise UnicodeError(msg)
def get_available_languages(domain):
diff --git a/tempest/openstack/common/importutils.py b/tempest/openstack/common/importutils.py
index 6c0d3b2..d5dd22f 100644
--- a/tempest/openstack/common/importutils.py
+++ b/tempest/openstack/common/importutils.py
@@ -24,10 +24,10 @@
def import_class(import_str):
"""Returns a class from a string including module and class."""
mod_str, _sep, class_str = import_str.rpartition('.')
+ __import__(mod_str)
try:
- __import__(mod_str)
return getattr(sys.modules[mod_str], class_str)
- except (ValueError, AttributeError):
+ except AttributeError:
raise ImportError('Class %s cannot be found (%s)' %
(class_str,
traceback.format_exception(*sys.exc_info())))
diff --git a/tempest/services/compute/json/servers_client.py b/tempest/services/compute/json/servers_client.py
index 23c1e64..e1661c0 100644
--- a/tempest/services/compute/json/servers_client.py
+++ b/tempest/services/compute/json/servers_client.py
@@ -432,8 +432,9 @@
return self.action(server_id, 'shelveOffload', None, **kwargs)
def get_console_output(self, server_id, length):
+ kwargs = {'length': length} if length else {}
return self.action(server_id, 'os-getConsoleOutput', 'output',
- common_schema.get_console_output, length=length)
+ common_schema.get_console_output, **kwargs)
def list_virtual_interfaces(self, server_id):
"""
diff --git a/tempest/services/compute/xml/servers_client.py b/tempest/services/compute/xml/servers_client.py
index 626e655..156d889 100644
--- a/tempest/services/compute/xml/servers_client.py
+++ b/tempest/services/compute/xml/servers_client.py
@@ -594,8 +594,9 @@
return resp, body
def get_console_output(self, server_id, length):
+ kwargs = {'length': length} if length else {}
return self.action(server_id, 'os-getConsoleOutput', 'output',
- length=length)
+ **kwargs)
def list_virtual_interfaces(self, server_id):
"""
diff --git a/tempest/services/image/v1/json/image_client.py b/tempest/services/image/v1/json/image_client.py
index 2353908..4a7c163 100644
--- a/tempest/services/image/v1/json/image_client.py
+++ b/tempest/services/image/v1/json/image_client.py
@@ -22,6 +22,7 @@
from tempest.common import glance_http
from tempest.common import rest_client
+from tempest.common.utils import misc as misc_utils
from tempest import config
from tempest import exceptions
from tempest.openstack.common import log as logging
@@ -297,6 +298,9 @@
'while waiting for %s, '
'but we got %s.' %
(self.build_timeout, status, value))
+ caller = misc_utils.find_test_caller()
+ if caller:
+ message = '(%s) %s' % (caller, message)
raise exceptions.TimeoutException(message)
time.sleep(self.build_interval)
old_value = value
diff --git a/tempest/services/object_storage/object_client.py b/tempest/services/object_storage/object_client.py
index f3f4eb6..b2f8205 100644
--- a/tempest/services/object_storage/object_client.py
+++ b/tempest/services/object_storage/object_client.py
@@ -13,7 +13,9 @@
# License for the specific language governing permissions and limitations
# under the License.
+import httplib
import urllib
+import urlparse
from tempest.common import http
from tempest.common import rest_client
@@ -143,6 +145,31 @@
resp, body = self.put(url, data)
return resp, body
+ def put_object_with_chunk(self, container, name, contents, chunk_size):
+ """
+ Put an object with Transfer-Encoding header
+ """
+ if self.base_url is None:
+ self._set_auth()
+
+ headers = {'Transfer-Encoding': 'chunked'}
+ if self.token:
+ headers['X-Auth-Token'] = self.token
+
+ conn = put_object_connection(self.base_url, container, name, contents,
+ chunk_size, headers)
+
+ resp = conn.getresponse()
+ body = resp.read()
+
+ resp_headers = {}
+ for header, value in resp.getheaders():
+ resp_headers[header.lower()] = value
+
+ self._error_checker('PUT', None, headers, contents, resp, body)
+
+ return resp.status, resp.reason, resp_headers
+
class ObjectClientCustomizedHeader(rest_client.RestClient):
@@ -220,3 +247,89 @@
url = "%s/%s" % (str(container), str(object_name))
resp, body = self.delete(url, headers=headers)
return resp, body
+
+ def create_object_continue(self, container, object_name,
+ data, metadata=None):
+ """Create storage object."""
+ headers = {}
+ if metadata:
+ for key in metadata:
+ headers[str(key)] = metadata[key]
+
+ if not data:
+ headers['content-length'] = '0'
+
+ if self.base_url is None:
+ self._set_auth()
+ headers['X-Auth-Token'] = self.token
+
+ conn = put_object_connection(self.base_url, str(container),
+ str(object_name), data, None, headers)
+
+ response = conn.response_class(conn.sock,
+ strict=conn.strict,
+ method=conn._method)
+ version, status, reason = response._read_status()
+ resp = {'version': version,
+ 'status': str(status),
+ 'reason': reason}
+
+ return resp
+
+
+def put_object_connection(base_url, container, name, contents=None,
+ chunk_size=65536, headers=None, query_string=None):
+ """
+ Helper function to make connection to put object with httplib
+ :param base_url: base_url of an object client
+ :param container: container name that the object is in
+ :param name: object name to put
+ :param contents: a string or a file like object to read object data
+ from; if None, a zero-byte put will be done
+ :param chunk_size: chunk size of data to write; it defaults to 65536;
+ used only if the the contents object has a 'read'
+ method, eg. file-like objects, ignored otherwise
+ :param headers: additional headers to include in the request, if any
+ :param query_string: if set will be appended with '?' to generated path
+ """
+ parsed = urlparse.urlparse(base_url)
+ if parsed.scheme == 'https':
+ conn = httplib.HTTPSConnection(parsed.netloc)
+ else:
+ conn = httplib.HTTPConnection(parsed.netloc)
+ path = str(parsed.path) + "/"
+ path += "%s/%s" % (str(container), str(name))
+
+ if query_string:
+ path += '?' + query_string
+ if headers:
+ headers = dict(headers)
+ else:
+ headers = {}
+ if hasattr(contents, 'read'):
+ conn.putrequest('PUT', path)
+ for header, value in headers.iteritems():
+ conn.putheader(header, value)
+ if 'Content-Length' not in headers:
+ if 'Transfer-Encoding' not in headers:
+ conn.putheader('Transfer-Encoding', 'chunked')
+ conn.endheaders()
+ chunk = contents.read(chunk_size)
+ while chunk:
+ conn.send('%x\r\n%s\r\n' % (len(chunk), chunk))
+ chunk = contents.read(chunk_size)
+ conn.send('0\r\n\r\n')
+ else:
+ conn.endheaders()
+ left = headers['Content-Length']
+ while left > 0:
+ size = chunk_size
+ if size > left:
+ size = left
+ chunk = contents.read(size)
+ conn.send(chunk)
+ left -= len(chunk)
+ else:
+ conn.request('PUT', path, contents, headers)
+
+ return conn
diff --git a/tempest/services/queuing/json/queuing_client.py b/tempest/services/queuing/json/queuing_client.py
index e5978f5..031c9c6 100644
--- a/tempest/services/queuing/json/queuing_client.py
+++ b/tempest/services/queuing/json/queuing_client.py
@@ -14,11 +14,14 @@
# limitations under the License.
import json
+import urllib
from tempest.api_schema.queuing.v1 import queues as queues_schema
from tempest.common import rest_client
+from tempest.common.utils import data_utils
from tempest import config
+
CONF = config.CONF
@@ -30,11 +33,16 @@
self.version = '1'
self.uri_prefix = 'v{0}'.format(self.version)
+ client_id = data_utils.rand_uuid_hex()
+ self.headers = {'Client-ID': client_id}
+
def list_queues(self):
uri = '{0}/queues'.format(self.uri_prefix)
resp, body = self.get(uri)
- body = json.loads(body)
- self.validate_response(queues_schema.list_queues, resp, body)
+
+ if resp['status'] != '204':
+ body = json.loads(body)
+ self.validate_response(queues_schema.list_queues, resp, body)
return resp, body
def create_queue(self, queue_name):
@@ -74,3 +82,80 @@
uri = '{0}/queues/{1}/metadata'.format(self.uri_prefix, queue_name)
resp, body = self.put(uri, body=json.dumps(rbody))
return resp, body
+
+ def post_messages(self, queue_name, rbody):
+ uri = '{0}/queues/{1}/messages'.format(self.uri_prefix, queue_name)
+ resp, body = self.post(uri, body=json.dumps(rbody),
+ extra_headers=True,
+ headers=self.headers)
+
+ body = json.loads(body)
+ return resp, body
+
+ def list_messages(self, queue_name):
+ uri = '{0}/queues/{1}/messages?echo=True'.format(self.uri_prefix,
+ queue_name)
+ resp, body = self.get(uri, extra_headers=True, headers=self.headers)
+
+ if resp['status'] != '204':
+ body = json.loads(body)
+ self.validate_response(queues_schema.list_messages, resp, body)
+
+ return resp, body
+
+ def get_single_message(self, message_uri):
+ resp, body = self.get(message_uri, extra_headers=True,
+ headers=self.headers)
+ if resp['status'] != '204':
+ body = json.loads(body)
+ self.validate_response(queues_schema.get_single_message, resp,
+ body)
+ return resp, body
+
+ def get_multiple_messages(self, message_uri):
+ resp, body = self.get(message_uri, extra_headers=True,
+ headers=self.headers)
+
+ if resp['status'] != '204':
+ body = json.loads(body)
+ self.validate_response(queues_schema.get_multiple_messages,
+ resp,
+ body)
+
+ return resp, body
+
+ def delete_messages(self, message_uri):
+ resp, body = self.delete(message_uri)
+ assert(resp['status'] == '204')
+ return resp, body
+
+ def post_claims(self, queue_name, rbody, url_params=False):
+ uri = '{0}/queues/{1}/claims'.format(self.uri_prefix, queue_name)
+ if url_params:
+ uri += '?%s' % urllib.urlencode(url_params)
+
+ resp, body = self.post(uri, body=json.dumps(rbody),
+ extra_headers=True,
+ headers=self.headers)
+
+ body = json.loads(body)
+ self.validate_response(queues_schema.claim_messages, resp, body)
+ return resp, body
+
+ def query_claim(self, claim_uri):
+ resp, body = self.get(claim_uri)
+
+ if resp['status'] != '204':
+ body = json.loads(body)
+ self.validate_response(queues_schema.query_claim, resp, body)
+ return resp, body
+
+ def update_claim(self, claim_uri, rbody):
+ resp, body = self.patch(claim_uri, body=json.dumps(rbody))
+ assert(resp['status'] == '204')
+ return resp, body
+
+ def release_claim(self, claim_uri):
+ resp, body = self.delete(claim_uri)
+ assert(resp['status'] == '204')
+ return resp, body
diff --git a/tempest/test.py b/tempest/test.py
index 650fad7..afe7a96 100644
--- a/tempest/test.py
+++ b/tempest/test.py
@@ -108,6 +108,7 @@
'object_storage': CONF.service_available.swift,
'dashboard': CONF.service_available.horizon,
'ceilometer': CONF.service_available.ceilometer,
+ 'data_processing': CONF.service_available.sahara
}
def decorator(f):
diff --git a/tempest/tests/test_tenant_isolation.py b/tempest/tests/test_tenant_isolation.py
index 485beff..bbc3d15 100644
--- a/tempest/tests/test_tenant_isolation.py
+++ b/tempest/tests/test_tenant_isolation.py
@@ -42,6 +42,8 @@
self.fake_http = fake_http.fake_httplib2(return_type=200)
self.stubs.Set(http.ClosingHttp, 'request',
fake_identity._fake_v2_response)
+ cfg.CONF.set_default('operator_role', 'FakeRole',
+ group='object-storage')
def test_tempest_client(self):
iso_creds = isolated_creds.IsolatedCreds('test class')
@@ -92,6 +94,31 @@
{'id': id, 'name': name})))
return tenant_fix
+ def _mock_list_roles(self, id, name):
+ roles_fix = self.useFixture(mockpatch.PatchObject(
+ json_iden_client.IdentityClientJSON,
+ 'list_roles',
+ return_value=({'status': 200},
+ [{'id': id, 'name': name},
+ {'id': '1', 'name': 'FakeRole'}])))
+ return roles_fix
+
+ def _mock_assign_user_role(self):
+ tenant_fix = self.useFixture(mockpatch.PatchObject(
+ json_iden_client.IdentityClientJSON,
+ 'assign_user_role',
+ return_value=({'status': 200},
+ {})))
+ return tenant_fix
+
+ def _mock_list_role(self):
+ roles_fix = self.useFixture(mockpatch.PatchObject(
+ json_iden_client.IdentityClientJSON,
+ 'list_roles',
+ return_value=({'status': 200},
+ [{'id': '1', 'name': 'FakeRole'}])))
+ return roles_fix
+
def _mock_network_create(self, iso_creds, id, name):
net_fix = self.useFixture(mockpatch.PatchObject(
iso_creds.network_admin_client,
@@ -121,6 +148,8 @@
cfg.CONF.set_default('neutron', False, 'service_available')
iso_creds = isolated_creds.IsolatedCreds('test class',
password='fake_password')
+ self._mock_assign_user_role()
+ self._mock_list_role()
self._mock_tenant_create('1234', 'fake_prim_tenant')
self._mock_user_create('1234', 'fake_prim_user')
primary_creds = iso_creds.get_primary_creds()
@@ -135,13 +164,9 @@
cfg.CONF.set_default('neutron', False, 'service_available')
iso_creds = isolated_creds.IsolatedCreds('test class',
password='fake_password')
+ self._mock_list_roles('1234', 'admin')
self._mock_user_create('1234', 'fake_admin_user')
self._mock_tenant_create('1234', 'fake_admin_tenant')
- self.useFixture(mockpatch.PatchObject(
- json_iden_client.IdentityClientJSON,
- 'list_roles',
- return_value=({'status': 200},
- [{'id': '1234', 'name': 'admin'}])))
user_mock = mock.patch.object(json_iden_client.IdentityClientJSON,
'assign_user_role')
@@ -150,7 +175,9 @@
with mock.patch.object(json_iden_client.IdentityClientJSON,
'assign_user_role') as user_mock:
admin_creds = iso_creds.get_admin_creds()
- user_mock.assert_called_once_with('1234', '1234', '1234')
+ user_mock.assert_has_calls([
+ mock.call('1234', '1234', '1'),
+ mock.call('1234', '1234', '1234')])
self.assertEqual(admin_creds.username, 'fake_admin_user')
self.assertEqual(admin_creds.tenant_name, 'fake_admin_tenant')
# Verify IDs
@@ -162,6 +189,8 @@
cfg.CONF.set_default('neutron', False, 'service_available')
iso_creds = isolated_creds.IsolatedCreds('test class',
password='fake_password')
+ self._mock_assign_user_role()
+ roles_fix = self._mock_list_role()
tenant_fix = self._mock_tenant_create('1234', 'fake_prim_tenant')
user_fix = self._mock_user_create('1234', 'fake_prim_user')
iso_creds.get_primary_creds()
@@ -172,16 +201,11 @@
iso_creds.get_alt_creds()
tenant_fix.cleanUp()
user_fix.cleanUp()
+ roles_fix.cleanUp()
tenant_fix = self._mock_tenant_create('123456', 'fake_admin_tenant')
user_fix = self._mock_user_create('123456', 'fake_admin_user')
- self.useFixture(mockpatch.PatchObject(
- json_iden_client.IdentityClientJSON,
- 'list_roles',
- return_value=({'status': 200},
- [{'id': '123456', 'name': 'admin'}])))
- with mock.patch.object(json_iden_client.IdentityClientJSON,
- 'assign_user_role'):
- iso_creds.get_admin_creds()
+ self._mock_list_roles('123456', 'admin')
+ iso_creds.get_admin_creds()
user_mock = self.patch(
'tempest.services.identity.json.identity_client.'
'IdentityClientJSON.delete_user')
@@ -209,6 +233,8 @@
cfg.CONF.set_default('neutron', False, 'service_available')
iso_creds = isolated_creds.IsolatedCreds('test class',
password='fake_password')
+ self._mock_assign_user_role()
+ self._mock_list_role()
self._mock_user_create('1234', 'fake_alt_user')
self._mock_tenant_create('1234', 'fake_alt_tenant')
alt_creds = iso_creds.get_alt_creds()
@@ -222,6 +248,8 @@
def test_network_creation(self, MockRestClient):
iso_creds = isolated_creds.IsolatedCreds('test class',
password='fake_password')
+ self._mock_assign_user_role()
+ self._mock_list_role()
self._mock_user_create('1234', 'fake_prim_user')
self._mock_tenant_create('1234', 'fake_prim_tenant')
self._mock_network_create(iso_creds, '1234', 'fake_net')
@@ -247,6 +275,8 @@
iso_creds = isolated_creds.IsolatedCreds('test class',
password='fake_password')
# Create primary tenant and network
+ self._mock_assign_user_role()
+ roles_fix = self._mock_list_role()
user_fix = self._mock_user_create('1234', 'fake_prim_user')
tenant_fix = self._mock_tenant_create('1234', 'fake_prim_tenant')
net_fix = self._mock_network_create(iso_creds, '1234', 'fake_net')
@@ -278,6 +308,7 @@
net_fix.cleanUp()
subnet_fix.cleanUp()
router_fix.cleanUp()
+ roles_fix.cleanUp()
# Create admin tenant and networks
user_fix = self._mock_user_create('123456', 'fake_admin_user')
tenant_fix = self._mock_tenant_create('123456', 'fake_admin_tenant')
@@ -286,14 +317,8 @@
subnet_fix = self._mock_subnet_create(iso_creds, '123456',
'fake_admin_subnet')
router_fix = self._mock_router_create('123456', 'fake_admin_router')
- self.useFixture(mockpatch.PatchObject(
- json_iden_client.IdentityClientJSON,
- 'list_roles',
- return_value=({'status': 200},
- [{'id': '123456', 'name': 'admin'}])))
- with mock.patch.object(json_iden_client.IdentityClientJSON,
- 'assign_user_role'):
- iso_creds.get_admin_creds()
+ self._mock_list_roles('123456', 'admin')
+ iso_creds.get_admin_creds()
self.patch('tempest.services.identity.json.identity_client.'
'IdentityClientJSON.delete_user')
self.patch('tempest.services.identity.json.identity_client.'
@@ -348,6 +373,8 @@
def test_network_alt_creation(self, MockRestClient):
iso_creds = isolated_creds.IsolatedCreds('test class',
password='fake_password')
+ self._mock_assign_user_role()
+ self._mock_list_role()
self._mock_user_create('1234', 'fake_alt_user')
self._mock_tenant_create('1234', 'fake_alt_tenant')
self._mock_network_create(iso_creds, '1234', 'fake_alt_net')
@@ -372,6 +399,7 @@
def test_network_admin_creation(self, MockRestClient):
iso_creds = isolated_creds.IsolatedCreds('test class',
password='fake_password')
+ self._mock_assign_user_role()
self._mock_user_create('1234', 'fake_admin_user')
self._mock_tenant_create('1234', 'fake_admin_tenant')
self._mock_network_create(iso_creds, '1234', 'fake_admin_net')
@@ -380,14 +408,8 @@
router_interface_mock = self.patch(
'tempest.services.network.json.network_client.NetworkClientJSON.'
'add_router_interface_with_subnet_id')
- self.useFixture(mockpatch.PatchObject(
- json_iden_client.IdentityClientJSON,
- 'list_roles',
- return_value=({'status': 200},
- [{'id': '123456', 'name': 'admin'}])))
- with mock.patch.object(json_iden_client.IdentityClientJSON,
- 'assign_user_role'):
- iso_creds.get_admin_creds()
+ self._mock_list_roles('123456', 'admin')
+ iso_creds.get_admin_creds()
router_interface_mock.called_once_with('1234', '1234')
network = iso_creds.get_admin_network()
subnet = iso_creds.get_admin_subnet()
@@ -410,6 +432,8 @@
iso_creds = isolated_creds.IsolatedCreds('test class',
password='fake_password',
network_resources=net_dict)
+ self._mock_assign_user_role()
+ self._mock_list_role()
self._mock_user_create('1234', 'fake_prim_user')
self._mock_tenant_create('1234', 'fake_prim_tenant')
net = mock.patch.object(iso_creds.network_admin_client,
@@ -444,6 +468,8 @@
iso_creds = isolated_creds.IsolatedCreds('test class',
password='fake_password',
network_resources=net_dict)
+ self._mock_assign_user_role()
+ self._mock_list_role()
self._mock_user_create('1234', 'fake_prim_user')
self._mock_tenant_create('1234', 'fake_prim_tenant')
self.assertRaises(exceptions.InvalidConfiguration,
@@ -460,6 +486,8 @@
iso_creds = isolated_creds.IsolatedCreds('test class',
password='fake_password',
network_resources=net_dict)
+ self._mock_assign_user_role()
+ self._mock_list_role()
self._mock_user_create('1234', 'fake_prim_user')
self._mock_tenant_create('1234', 'fake_prim_tenant')
self.assertRaises(exceptions.InvalidConfiguration,
@@ -476,6 +504,8 @@
iso_creds = isolated_creds.IsolatedCreds('test class',
password='fake_password',
network_resources=net_dict)
+ self._mock_assign_user_role()
+ self._mock_list_role()
self._mock_user_create('1234', 'fake_prim_user')
self._mock_tenant_create('1234', 'fake_prim_tenant')
self.assertRaises(exceptions.InvalidConfiguration,
diff --git a/test-requirements.txt b/test-requirements.txt
index 215f28b..13ef291 100644
--- a/test-requirements.txt
+++ b/test-requirements.txt
@@ -8,3 +8,4 @@
mock>=1.0
coverage>=3.6
oslotest
+stevedore>=0.14
diff --git a/tools/config/check_uptodate.sh b/tools/config/check_uptodate.sh
index 528bd5b..0f0d77e 100755
--- a/tools/config/check_uptodate.sh
+++ b/tools/config/check_uptodate.sh
@@ -16,6 +16,10 @@
trap "rm -rf $TEMPDIR" EXIT
tools/config/generate_sample.sh -b ./ -p ${PROJECT_NAME} -o ${TEMPDIR}
+if [ $? != 0 ]
+then
+ exit 1
+fi
if ! diff -u ${TEMPDIR}/${CFGFILE_NAME} ${CFGFILE}
then
diff --git a/tools/config/generate_sample.sh b/tools/config/generate_sample.sh
index 20ddfbb..d22b2f0 100755
--- a/tools/config/generate_sample.sh
+++ b/tools/config/generate_sample.sh
@@ -1,5 +1,15 @@
#!/usr/bin/env bash
+# Generate sample configuration for your project.
+#
+# Aside from the command line flags, it also respects a config file which
+# should be named oslo.config.generator.rc and be placed in the same directory.
+#
+# You can then export the following variables:
+# TEMPEST_CONFIG_GENERATOR_EXTRA_MODULES: list of modules to interrogate for options.
+# TEMPEST_CONFIG_GENERATOR_EXTRA_LIBRARIES: list of libraries to discover.
+# TEMPEST_CONFIG_GENERATOR_EXCLUDED_FILES: list of files to remove from automatic listing.
+
print_hint() {
echo "Try \`${0##*/} --help' for more information." >&2
}
@@ -95,6 +105,10 @@
source "$RC_FILE"
fi
+for filename in ${TEMPEST_CONFIG_GENERATOR_EXCLUDED_FILES}; do
+ FILES="${FILES[@]/$filename/}"
+done
+
for mod in ${TEMPEST_CONFIG_GENERATOR_EXTRA_MODULES}; do
MODULES="$MODULES -m $mod"
done
@@ -111,6 +125,11 @@
MODULEPATH=${MODULEPATH:-$DEFAULT_MODULEPATH}
OUTPUTFILE=$OUTPUTDIR/$PACKAGENAME.conf.sample
python -m $MODULEPATH $MODULES $LIBRARIES $FILES > $OUTPUTFILE
+if [ $? != 0 ]
+then
+ echo "Can not generate $OUTPUTFILE"
+ exit 1
+fi
# Hook to allow projects to append custom config file snippets
CONCAT_FILES=$(ls $BASEDIR/tools/config/*.conf.sample 2>/dev/null)
diff --git a/tools/config/oslo.config.generator.rc b/tools/config/oslo.config.generator.rc
new file mode 100644
index 0000000..303e156
--- /dev/null
+++ b/tools/config/oslo.config.generator.rc
@@ -0,0 +1 @@
+MODULEPATH=tempest.common.generate_sample_tempest
diff --git a/tools/generate_sample.sh b/tools/generate_sample.sh
deleted file mode 100755
index 9b312c9..0000000
--- a/tools/generate_sample.sh
+++ /dev/null
@@ -1,2 +0,0 @@
-#!/bin/sh
-MODULEPATH=tempest.common.generate_sample_tempest tools/config/generate_sample.sh $@
diff --git a/tox.ini b/tox.ini
index 514ad23..7f69fad 100644
--- a/tox.ini
+++ b/tox.ini
@@ -85,7 +85,6 @@
-r{toxinidir}/test-requirements.txt
[testenv:pep8]
-setenv = MODULEPATH=tempest.common.generate_sample_tempest
commands =
flake8 {posargs}
{toxinidir}/tools/config/check_uptodate.sh