New minimum guaranteed bandwidth tests
New tests for MinBwAllocationPlacementTest: server resize and migration.
Blueprint: support-move-ops-with-qos-ports-ussuri
Change-Id: I8c573c4a11a76932057d8415d76114a03269b9f3
diff --git a/tempest/scenario/test_minbw_allocation_placement.py b/tempest/scenario/test_minbw_allocation_placement.py
index 74d4ed9..86639c5 100644
--- a/tempest/scenario/test_minbw_allocation_placement.py
+++ b/tempest/scenario/test_minbw_allocation_placement.py
@@ -12,7 +12,7 @@
# License for the specific language governing permissions and limitations
# under the License.
-from oslo_log import log as logging
+import testtools
from tempest.common import utils
from tempest.common import waiters
@@ -23,7 +23,6 @@
from tempest.scenario import manager
-LOG = logging.getLogger(__name__)
CONF = config.CONF
@@ -65,6 +64,8 @@
cls.routers_client = cls.os_adm.routers_client
cls.qos_client = cls.os_admin.qos_client
cls.qos_min_bw_client = cls.os_admin.qos_min_bw_client
+ cls.flavors_client = cls.os_adm.flavors_client
+ cls.servers_client = cls.os_adm.servers_client
@classmethod
def skip_checks(cls):
@@ -74,6 +75,11 @@
"placement based QoS allocation."
raise cls.skipException(msg)
+ def setUp(self):
+ super(MinBwAllocationPlacementTest, self).setUp()
+ self._check_if_allocation_is_possible()
+ self._create_network_and_qos_policies()
+
def _create_policy_and_min_bw_rule(self, name_prefix, min_kbps):
policy = self.qos_client.create_qos_policy(
name=data_utils.rand_name(name_prefix),
@@ -139,6 +145,33 @@
self.fail('For %s:%s there should be no available candidate!' %
(self.INGRESS_RESOURCE_CLASS, self.PLACEMENT_MAX_INT))
+ def _boot_vm_with_min_bw(self, qos_policy_id, status='ACTIVE'):
+ wait_until = (None if status == 'ERROR' else status)
+ port = self.create_port(
+ self.prov_network['id'], qos_policy_id=qos_policy_id)
+
+ server = self.create_server(networks=[{'port': port['id']}],
+ wait_until=wait_until)
+ waiters.wait_for_server_status(
+ client=self.os_primary.servers_client, server_id=server['id'],
+ status=status, ready_wait=False, raise_on_error=False)
+ return server, port
+
+ def _assert_allocation_is_as_expected(self, allocations, port_id):
+ self.assertGreater(len(allocations['allocations']), 0)
+ bw_resource_in_alloc = False
+ for rp, resources in allocations['allocations'].items():
+ if self.INGRESS_RESOURCE_CLASS in resources['resources']:
+ bw_resource_in_alloc = True
+ allocation_rp = rp
+ self.assertTrue(bw_resource_in_alloc)
+
+ # Check binding_profile of the port is not empty and equals with the
+ # rp uuid
+ port = self.os_admin.ports_client.show_port(port_id)
+ self.assertEqual(allocation_rp,
+ port['port']['binding:profile']['allocation'])
+
@decorators.idempotent_id('78625d92-212c-400e-8695-dd51706858b8')
@decorators.attr(type='slow')
@utils.services('compute', 'network')
@@ -162,40 +195,13 @@
it should fail.
"""
- self._check_if_allocation_is_possible()
-
- self._create_network_and_qos_policies()
-
- valid_port = self.create_port(
- self.prov_network['id'], qos_policy_id=self.qos_policy_valid['id'])
-
- server1 = self.create_server(
- networks=[{'port': valid_port['id']}])
+ server1, valid_port = self._boot_vm_with_min_bw(
+ qos_policy_id=self.qos_policy_valid['id'])
allocations = self.placement_client.list_allocations(server1['id'])
+ self._assert_allocation_is_as_expected(allocations, valid_port['id'])
- self.assertGreater(len(allocations['allocations']), 0)
- bw_resource_in_alloc = False
- for rp, resources in allocations['allocations'].items():
- if self.INGRESS_RESOURCE_CLASS in resources['resources']:
- bw_resource_in_alloc = True
- allocation_rp = rp
- self.assertTrue(bw_resource_in_alloc)
- # Check that binding_profile of the port is not empty and equals with
- # the rp uuid
- port = self.os_admin.ports_client.show_port(valid_port['id'])
- self.assertEqual(allocation_rp,
- port['port']['binding:profile']['allocation'])
-
- # boot another vm with max int bandwidth
- not_valid_port = self.create_port(
- self.prov_network['id'],
- qos_policy_id=self.qos_policy_not_valid['id'])
- server2 = self.create_server(
- wait_until=None,
- networks=[{'port': not_valid_port['id']}])
- waiters.wait_for_server_status(
- client=self.os_primary.servers_client, server_id=server2['id'],
- status='ERROR', ready_wait=False, raise_on_error=False)
+ server2, not_valid_port = self._boot_vm_with_min_bw(
+ self.qos_policy_not_valid['id'], status='ERROR')
allocations = self.placement_client.list_allocations(server2['id'])
self.assertEqual(0, len(allocations['allocations']))
@@ -205,3 +211,90 @@
# Check that binding_profile of the port is empty
port = self.os_admin.ports_client.show_port(not_valid_port['id'])
self.assertEqual(0, len(port['port']['binding:profile']))
+
+ @decorators.idempotent_id('8a98150c-a506-49a5-96c6-73a5e7b04ada')
+ @testtools.skipUnless(CONF.compute_feature_enabled.cold_migration,
+ 'Cold migration is not available.')
+ @testtools.skipUnless(CONF.compute.min_compute_nodes > 1,
+ 'Less than 2 compute nodes, skipping multinode '
+ 'tests.')
+ @utils.services('compute', 'network')
+ def test_migrate_with_qos_min_bw_allocation(self):
+ """Scenario to migrate VM with QoS min bw allocation in placement
+
+ Boot a VM like in test_qos_min_bw_allocation_basic, do the same
+ checks, and
+ * migrate the server
+ * confirm the resize, if the VM state is VERIFY_RESIZE
+ * If the VM goes to ACTIVE state check that allocations are as
+ expected.
+ """
+ server, valid_port = self._boot_vm_with_min_bw(
+ qos_policy_id=self.qos_policy_valid['id'])
+ allocations = self.placement_client.list_allocations(server['id'])
+ self._assert_allocation_is_as_expected(allocations, valid_port['id'])
+
+ self.servers_client.migrate_server(server_id=server['id'])
+ waiters.wait_for_server_status(
+ client=self.os_primary.servers_client, server_id=server['id'],
+ status='VERIFY_RESIZE', ready_wait=False, raise_on_error=False)
+ allocations = self.placement_client.list_allocations(server['id'])
+
+ # TODO(lajoskatona): Check that the allocations are ok for the
+ # migration?
+ self._assert_allocation_is_as_expected(allocations, valid_port['id'])
+
+ self.servers_client.confirm_resize_server(server_id=server['id'])
+ waiters.wait_for_server_status(
+ client=self.os_primary.servers_client, server_id=server['id'],
+ status='ACTIVE', ready_wait=False, raise_on_error=True)
+ allocations = self.placement_client.list_allocations(server['id'])
+ self._assert_allocation_is_as_expected(allocations, valid_port['id'])
+
+ @decorators.idempotent_id('c29e7fd3-035d-4993-880f-70819847683f')
+ @testtools.skipUnless(CONF.compute_feature_enabled.resize,
+ 'Resize not available.')
+ @utils.services('compute', 'network')
+ def test_resize_with_qos_min_bw_allocation(self):
+ """Scenario to resize VM with QoS min bw allocation in placement.
+
+ Boot a VM like in test_qos_min_bw_allocation_basic, do the same
+ checks, and
+ * resize the server with new flavor
+ * confirm the resize, if the VM state is VERIFY_RESIZE
+ * If the VM goes to ACTIVE state check that allocations are as
+ expected.
+ """
+ server, valid_port = self._boot_vm_with_min_bw(
+ qos_policy_id=self.qos_policy_valid['id'])
+ allocations = self.placement_client.list_allocations(server['id'])
+ self._assert_allocation_is_as_expected(allocations, valid_port['id'])
+
+ old_flavor = self.flavors_client.show_flavor(
+ CONF.compute.flavor_ref)['flavor']
+ new_flavor = self.flavors_client.create_flavor(**{
+ 'ram': old_flavor['ram'],
+ 'vcpus': old_flavor['vcpus'],
+ 'name': old_flavor['name'] + 'extra',
+ 'disk': old_flavor['disk'] + 1
+ })['flavor']
+ self.addCleanup(test_utils.call_and_ignore_notfound_exc,
+ self.flavors_client.delete_flavor, new_flavor['id'])
+
+ self.servers_client.resize_server(
+ server_id=server['id'], flavor_ref=new_flavor['id'])
+ waiters.wait_for_server_status(
+ client=self.os_primary.servers_client, server_id=server['id'],
+ status='VERIFY_RESIZE', ready_wait=False, raise_on_error=False)
+ allocations = self.placement_client.list_allocations(server['id'])
+
+ # TODO(lajoskatona): Check that the allocations are ok for the
+ # migration?
+ self._assert_allocation_is_as_expected(allocations, valid_port['id'])
+
+ self.servers_client.confirm_resize_server(server_id=server['id'])
+ waiters.wait_for_server_status(
+ client=self.os_primary.servers_client, server_id=server['id'],
+ status='ACTIVE', ready_wait=False, raise_on_error=True)
+ allocations = self.placement_client.list_allocations(server['id'])
+ self._assert_allocation_is_as_expected(allocations, valid_port['id'])
diff --git a/zuul.d/integrated-gate.yaml b/zuul.d/integrated-gate.yaml
index 1a1d523..bf689c5 100644
--- a/zuul.d/integrated-gate.yaml
+++ b/zuul.d/integrated-gate.yaml
@@ -76,6 +76,18 @@
FORCE_CONFIG_DRIVE: true
ENABLE_VOLUME_MULTIATTACH: true
GLANCE_USE_IMPORT_WORKFLOW: True
+ devstack_plugins:
+ neutron: https://opendev.org/openstack/neutron
+ devstack_local_conf:
+ post-config:
+ "/$NEUTRON_CORE_PLUGIN_CONF":
+ ovs:
+ bridge_mappings: public:br-ex
+ resource_provider_bandwidths: br-ex:1000000:1000000
+ test-config:
+ $TEMPEST_CONFIG:
+ network-feature-enabled:
+ qos_placement_physnet: public
devstack_services:
s-account: false
s-container: false
@@ -89,6 +101,8 @@
# especially because the tests fail at a high rate (see bugs 1483434,
# 1813217, 1745168)
c-bak: false
+ neutron-placement: true
+ neutron-qos: true
- job:
name: tempest-integrated-networking
@@ -244,6 +258,21 @@
vars:
devstack_localrc:
USE_PYTHON3: true
+ devstack_plugins:
+ neutron: https://opendev.org/openstack/neutron
+ devstack_local_conf:
+ post-config:
+ "/$NEUTRON_CORE_PLUGIN_CONF":
+ ovs:
+ bridge_mappings: public:br-ex
+ resource_provider_bandwidths: br-ex:1000000:1000000
+ test-config:
+ $TEMPEST_CONFIG:
+ network-feature-enabled:
+ qos_placement_physnet: public
+ devstack_services:
+ neutron-placement: true
+ neutron-qos: true
group-vars:
subnode:
devstack_localrc: