|  | Tempest Coding Guide | 
|  | ==================== | 
|  |  | 
|  | - Step 1: Read the OpenStack Style Commandments | 
|  | https://docs.openstack.org/hacking/latest/ | 
|  | - Step 2: Read on | 
|  |  | 
|  | Tempest Specific Commandments | 
|  | ----------------------------- | 
|  |  | 
|  | - [T102] Cannot import OpenStack python clients in tempest/api & | 
|  | tempest/scenario tests | 
|  | - [T104] Scenario tests require a services decorator | 
|  | - [T105] Tests cannot use setUpClass/tearDownClass | 
|  | - [T107] Check that a service tag isn't in the module path | 
|  | - [T108] Check no hyphen at the end of rand_name() argument | 
|  | - [T109] Cannot use testtools.skip decorator; instead use | 
|  | decorators.skip_because from tempest.lib | 
|  | - [T110] Check that service client names of GET should be consistent | 
|  | - [T111] Check that service client names of DELETE should be consistent | 
|  | - [T112] Check that tempest.lib should not import local tempest code | 
|  | - [T113] Check that tests use data_utils.rand_uuid() instead of uuid.uuid4() | 
|  | - [T114] Check that tempest.lib does not use tempest config | 
|  | - [T115] Check that admin tests should exist under admin path | 
|  | - [N322] Method's default argument shouldn't be mutable | 
|  | - [T116] Unsupported 'message' Exception attribute in PY3 | 
|  | - [T117] Check negative tests have ``@decorators.attr(type=['negative'])`` | 
|  | applied. | 
|  | - [T118] LOG.warn is deprecated. Enforce use of LOG.warning. | 
|  |  | 
|  | It is recommended to use ``tox -eautopep8`` before submitting a patch. | 
|  |  | 
|  | Test Data/Configuration | 
|  | ----------------------- | 
|  | - Assume nothing about existing test data | 
|  | - Tests should be self contained (provide their own data) | 
|  | - Clean up test data at the completion of each test | 
|  | - Use configuration files for values that will vary by environment | 
|  |  | 
|  | Supported OpenStack Components | 
|  | ------------------------------ | 
|  |  | 
|  | Tempest's :ref:`library` and :ref:`plugin interface <tempest_plugin>` can be | 
|  | leveraged to support integration testing for virtually any OpenStack component. | 
|  |  | 
|  | However, Tempest only offers **in-tree** integration testing coverage for the | 
|  | following components: | 
|  |  | 
|  | * Cinder | 
|  | * Glance | 
|  | * Keystone | 
|  | * Neutron | 
|  | * Nova | 
|  | * Swift | 
|  |  | 
|  | Historically, Tempest offered in-tree testing for other components as well, but | 
|  | since the introduction of the `External Plugin Interface`_, Tempest's in-tree | 
|  | testing scope has been limited to the projects above. Integration tests for | 
|  | projects not included above should go into one of the | 
|  | `relevant plugin projects`_. | 
|  |  | 
|  | .. _External Plugin Interface: https://specs.openstack.org/openstack/qa-specs/specs/tempest/implemented/tempest-external-plugin-interface.html | 
|  | .. _relevant plugin projects: https://docs.openstack.org/tempest/latest/plugins/plugin-registry.html#detected-plugins | 
|  |  | 
|  | Exception Handling | 
|  | ------------------ | 
|  | According to the ``The Zen of Python`` the | 
|  | ``Errors should never pass silently.`` | 
|  | Tempest usually runs in special environment (jenkins gate jobs), in every | 
|  | error or failure situation we should provide as much error related | 
|  | information as possible, because we usually do not have the chance to | 
|  | investigate the situation after the issue happened. | 
|  |  | 
|  | In every test case the abnormal situations must be very verbosely explained, | 
|  | by the exception and the log. | 
|  |  | 
|  | In most cases the very first issue is the most important information. | 
|  |  | 
|  | Try to avoid using ``try`` blocks in the test cases, as both the ``except`` | 
|  | and ``finally`` blocks could replace the original exception, | 
|  | when the additional operations leads to another exception. | 
|  |  | 
|  | Just letting an exception to propagate, is not a bad idea in a test case, | 
|  | at all. | 
|  |  | 
|  | Try to avoid using any exception handling construct which can hide the errors | 
|  | origin. | 
|  |  | 
|  | If you really need to use a ``try`` block, please ensure the original | 
|  | exception at least logged.  When the exception is logged you usually need | 
|  | to ``raise`` the same or a different exception anyway. | 
|  |  | 
|  | Use of ``self.addCleanup`` is often a good way to avoid having to catch | 
|  | exceptions and still ensure resources are correctly cleaned up if the | 
|  | test fails part way through. | 
|  |  | 
|  | Use the ``self.assert*`` methods provided by the unit test framework. | 
|  | This signals the failures early on. | 
|  |  | 
|  | Avoid using the ``self.fail`` alone, its stack trace will signal | 
|  | the ``self.fail`` line as the origin of the error. | 
|  |  | 
|  | Avoid constructing complex boolean expressions for assertion. | 
|  | The ``self.assertTrue`` or ``self.assertFalse`` without a ``msg`` argument, | 
|  | will just tell you the single boolean value, and you will not know anything | 
|  | about the values used in the formula, the ``msg`` argument might be good enough | 
|  | for providing more information. | 
|  |  | 
|  | Most other assert method can include more information by default. | 
|  | For example ``self.assertIn`` can include the whole set. | 
|  |  | 
|  | It is recommended to use testtools `matcher`_ for the more tricky assertions. | 
|  | You can implement your own specific `matcher`_ as well. | 
|  |  | 
|  | .. _matcher: https://testtools.readthedocs.org/en/latest/for-test-authors.html#matchers | 
|  |  | 
|  | If the test case fails you can see the related logs and the information | 
|  | carried by the exception (exception class, backtrack and exception info). | 
|  | This and the service logs are your only guide to finding the root cause of flaky | 
|  | issues. | 
|  |  | 
|  | Test cases are independent | 
|  | -------------------------- | 
|  | Every ``test_method`` must be callable individually and MUST NOT depends on, | 
|  | any other ``test_method`` or ``test_method`` ordering. | 
|  |  | 
|  | Test cases MAY depend on commonly initialized resources/facilities, like | 
|  | credentials management, testresources and so on. These facilities, MUST be able | 
|  | to work even if just one ``test_method`` is selected for execution. | 
|  |  | 
|  | Service Tagging | 
|  | --------------- | 
|  | Service tagging is used to specify which services are exercised by a particular | 
|  | test method. You specify the services with the ``tempest.common.utils.services`` | 
|  | decorator. For example: | 
|  |  | 
|  | ``@utils.services('compute', 'image')`` | 
|  |  | 
|  | Valid service tag names are the same as the list of directories in tempest.api | 
|  | that have tests. | 
|  |  | 
|  | For scenario tests having a service tag is required. For the API tests service | 
|  | tags are only needed if the test method makes an API call (either directly or | 
|  | indirectly through another service) that differs from the parent directory | 
|  | name. For example, any test that make an API call to a service other than Nova | 
|  | in ``tempest.api.compute`` would require a service tag for those services, | 
|  | however they do not need to be tagged as ``compute``. | 
|  |  | 
|  | Test Attributes | 
|  | --------------- | 
|  | Tempest leverages `test attributes`_ which are a simple but effective way of | 
|  | distinguishing between different "types" of API tests. A test can be "tagged" | 
|  | with such attributes using the ``decorators.attr`` decorator, for example:: | 
|  |  | 
|  | @decorators.attr(type=['negative']) | 
|  | def test_aggregate_create_aggregate_name_length_less_than_1(self): | 
|  | [...] | 
|  |  | 
|  | These test attributes can be used for test selection via regular expressions. | 
|  | For example, ``(?!.*\[.*\bslow\b.*\])(^tempest\.scenario)`` runs all the tests | 
|  | in the ``scenario`` test module, *except* for those tagged with the ``slow`` | 
|  | attribute (via a negative lookahead in the regular expression). These | 
|  | attributes are used in Tempest's ``tox.ini`` as well as Tempest's Zuul job | 
|  | definitions for specifying particular batches of Tempest test suites to run. | 
|  |  | 
|  | .. _test attributes: https://testtools.readthedocs.io/en/latest/for-test-authors.html?highlight=attr#test-attributes | 
|  |  | 
|  | Negative Attribute | 
|  | ^^^^^^^^^^^^^^^^^^ | 
|  | The ``type='negative'`` attribute is used to signify that a test is a negative | 
|  | test, which is a test that handles invalid input gracefully. This attribute | 
|  | should be applied to all negative test scenarios. | 
|  |  | 
|  | This attribute must be applied to each test that belongs to a negative test | 
|  | class, i.e. a test class name ending with "Negative.*" substring. | 
|  |  | 
|  | Slow Attribute | 
|  | ^^^^^^^^^^^^^^ | 
|  | The ``type='slow'`` attribute is used to signify that a test takes a long time | 
|  | to run, relatively speaking. This attribute is usually applied to | 
|  | :ref:`scenario tests <scenario_field_guide>`, which involve a complicated | 
|  | series of API operations, the total runtime of which can be relatively long. | 
|  | This long runtime has performance implications on `Zuul`_ jobs, which is why | 
|  | the ``slow`` attribute is leveraged to run slow tests on a selective basis, | 
|  | to keep total `Zuul`_ job runtime down to a reasonable time frame. | 
|  |  | 
|  | .. _Zuul: https://zuul-ci.org/docs/zuul/latest/ | 
|  |  | 
|  | Smoke Attribute | 
|  | ^^^^^^^^^^^^^^^ | 
|  | The ``type='smoke'`` attribute is used to signify that a test is a so-called | 
|  | smoke test, which is a type of test that tests the most vital OpenStack | 
|  | functionality, like listing servers or flavors or creating volumes. The | 
|  | attribute should be sparingly applied to only the tests that sanity-check the | 
|  | most essential functionality of an OpenStack cloud. | 
|  |  | 
|  | Multinode Attribute | 
|  | ^^^^^^^^^^^^^^^^^^^ | 
|  | The ``type='multinode'`` attribute is used to signify that a test is desired | 
|  | to be executed in a multinode environment. By marking the tests with this | 
|  | attribute we can avoid running tests which aren't that beneficial for the | 
|  | multinode setup and thus reduce the consumption of resources. | 
|  |  | 
|  | Test fixtures and resources | 
|  | --------------------------- | 
|  | Test level resources should be cleaned-up after the test execution. Clean-up | 
|  | is best scheduled using ``addCleanup`` which ensures that the resource cleanup | 
|  | code is always invoked, and in reverse order with respect to the creation | 
|  | order. | 
|  |  | 
|  | Test class level resources should be defined in the ``resource_setup`` method | 
|  | of the test class, except for any credential obtained from the credentials | 
|  | provider, which should be set-up in the ``setup_credentials`` method. | 
|  | Cleanup is best scheduled using ``addClassResourceCleanup`` which ensures that | 
|  | the cleanup code is always invoked, and in reverse order with respect to the | 
|  | creation order. | 
|  |  | 
|  | In both cases - test level and class level cleanups - a wait loop should be | 
|  | scheduled before the actual delete of resources with an asynchronous delete. | 
|  |  | 
|  | The test base class ``BaseTestCase`` defines Tempest framework for class level | 
|  | fixtures. ``setUpClass`` and ``tearDownClass`` are defined here and cannot be | 
|  | overwritten by subclasses (enforced via hacking rule T105). | 
|  |  | 
|  | Set-up is split in a series of steps (setup stages), which can be overwritten | 
|  | by test classes. Set-up stages are: | 
|  |  | 
|  | - ``skip_checks`` | 
|  | - ``setup_credentials`` | 
|  | - ``setup_clients`` | 
|  | - ``resource_setup`` | 
|  |  | 
|  | Tear-down is also split in a series of steps (teardown stages), which are | 
|  | stacked for execution only if the corresponding setup stage had been | 
|  | reached during the setup phase. Tear-down stages are: | 
|  |  | 
|  | - ``clear_credentials`` (defined in the base test class) | 
|  | - ``resource_cleanup`` | 
|  |  | 
|  | Skipping Tests | 
|  | -------------- | 
|  | Skipping tests should be based on configuration only. If that is not possible, | 
|  | it is likely that either a configuration flag is missing, or the test should | 
|  | fail rather than be skipped. | 
|  | Using discovery for skipping tests is generally discouraged. | 
|  |  | 
|  | When running a test that requires a certain "feature" in the target | 
|  | cloud, if that feature is missing we should fail, because either the test | 
|  | configuration is invalid, or the cloud is broken and the expected "feature" is | 
|  | not there even if the cloud was configured with it. | 
|  |  | 
|  | Negative Tests | 
|  | -------------- | 
|  | Error handling is an important aspect of API design and usage. Negative | 
|  | tests are a way to ensure that an application can gracefully handle | 
|  | invalid or unexpected input. However, as a black box integration test | 
|  | suite, Tempest is not suitable for handling all negative test cases, as | 
|  | the wide variety and complexity of negative tests can lead to long test | 
|  | runs and knowledge of internal implementation details. The bulk of | 
|  | negative testing should be handled with project function tests. | 
|  | All negative tests should be based on `API-WG guideline`_ . Such negative | 
|  | tests can block any changes from accurate failure code to invalid one. | 
|  |  | 
|  | .. _API-WG guideline: https://specs.openstack.org/openstack/api-wg/guidelines/http.html#failure-code-clarifications | 
|  |  | 
|  | If facing some gray area which is not clarified on the above guideline, propose | 
|  | a new guideline to the API-WG. With a proposal to the API-WG we will be able to | 
|  | build a consensus across all OpenStack projects and improve the quality and | 
|  | consistency of all the APIs. | 
|  |  | 
|  | In addition, we have some guidelines for additional negative tests. | 
|  |  | 
|  | - About BadRequest(HTTP400) case: We can add a single negative tests of | 
|  | BadRequest for each resource and method(POST, PUT). | 
|  | Please don't implement more negative tests on the same combination of | 
|  | resource and method even if API request parameters are different from | 
|  | the existing test. | 
|  | - About NotFound(HTTP404) case: We can add a single negative tests of | 
|  | NotFound for each resource and method(GET, PUT, DELETE, HEAD). | 
|  | Please don't implement more negative tests on the same combination | 
|  | of resource and method. | 
|  |  | 
|  | The above guidelines don't cover all cases and we will grow these guidelines | 
|  | organically over time. Patches outside of the above guidelines are left up to | 
|  | the reviewers' discretion and if we face some conflicts between reviewers, we | 
|  | will expand the guideline based on our discussion and experience. | 
|  |  | 
|  | Test skips because of Known Bugs | 
|  | -------------------------------- | 
|  | If a test is broken because of a bug it is appropriate to skip the test until | 
|  | bug has been fixed. You should use the ``skip_because`` decorator so that | 
|  | Tempest's skip tracking tool can watch the bug status. | 
|  |  | 
|  | Example:: | 
|  |  | 
|  | @skip_because(bug="980688") | 
|  | def test_this_and_that(self): | 
|  | ... | 
|  |  | 
|  | Guidelines | 
|  | ---------- | 
|  | - Do not submit changesets with only testcases which are skipped as | 
|  | they will not be merged. | 
|  | - Consistently check the status code of responses in testcases. The | 
|  | earlier a problem is detected the easier it is to debug, especially | 
|  | where there is complicated setup required. | 
|  |  | 
|  | Parallel Test Execution | 
|  | ----------------------- | 
|  | Tempest by default runs its tests in parallel this creates the possibility for | 
|  | interesting interactions between tests which can cause unexpected failures. | 
|  | Dynamic credentials provides protection from most of the potential race | 
|  | conditions between tests outside the same class. But there are still a few of | 
|  | things to watch out for to try to avoid issues when running your tests in | 
|  | parallel. | 
|  |  | 
|  | - Resources outside of a project scope still have the potential to conflict. This | 
|  | is a larger concern for the admin tests since most resources and actions that | 
|  | require admin privileges are outside of projects. | 
|  |  | 
|  | - Races between methods in the same class are not a problem because | 
|  | parallelization in Tempest is at the test class level, but if there is a json | 
|  | and xml version of the same test class there could still be a race between | 
|  | methods. | 
|  |  | 
|  | - The rand_name() function from tempest.lib.common.utils.data_utils should be | 
|  | used anywhere a resource is created with a name. Static naming should be | 
|  | avoided to prevent resource conflicts. | 
|  |  | 
|  | - If the execution of a set of tests is required to be serialized then locking | 
|  | can be used to perform this. See usage of ``LockFixture`` for examples of | 
|  | using locking. However, LockFixture only helps if you want to separate the | 
|  | execution of two small sets of test cases. On the other hand, if you need to | 
|  | run a set of tests separately from potentially all other tests then | 
|  | ``LockFixture`` does not scale as you would need to take the lock in all the | 
|  | other tests too. In this case, you can use the ``@serial`` decorator on top | 
|  | of the test class holding the tests that need to run separately from the | 
|  | potentially parallel test set. See more in :ref:`tempest_test_writing`. | 
|  |  | 
|  |  | 
|  | Sample Configuration File | 
|  | ------------------------- | 
|  | The sample config file is autogenerated using a script. If any changes are made | 
|  | to the config variables in tempest/config.py then the sample config file must be | 
|  | regenerated. This can be done running:: | 
|  |  | 
|  | tox -e genconfig | 
|  |  | 
|  | Unit Tests | 
|  | ---------- | 
|  | Unit tests are a separate class of tests in Tempest. They verify Tempest | 
|  | itself, and thus have a different set of guidelines around them: | 
|  |  | 
|  | 1. They can not require anything running externally. All you should need to | 
|  | run the unit tests is the git tree, python and the dependencies installed. | 
|  | This includes running services, a config file, etc. | 
|  |  | 
|  | 2. The unit tests cannot use setUpClass, instead fixtures and testresources | 
|  | should be used for shared state between tests. | 
|  |  | 
|  |  | 
|  | .. _TestDocumentation: | 
|  |  | 
|  | Test Documentation | 
|  | ------------------ | 
|  | For tests being added we need to require inline documentation in the form of | 
|  | docstrings to explain what is being tested. In API tests for a new API a class | 
|  | level docstring should be added to an API reference doc. If one doesn't exist | 
|  | a TODO comment should be put indicating that the reference needs to be added. | 
|  | For individual API test cases a method level docstring should be used to | 
|  | explain the functionality being tested if the test name isn't descriptive | 
|  | enough. For example:: | 
|  |  | 
|  | def test_get_role_by_id(self): | 
|  | """Get a role by its id.""" | 
|  |  | 
|  | the docstring there is superfluous and shouldn't be added. but for a method | 
|  | like:: | 
|  |  | 
|  | def test_volume_backup_create_get_detailed_list_restore_delete(self): | 
|  | pass | 
|  |  | 
|  | a docstring would be useful because while the test title is fairly descriptive | 
|  | the operations being performed are complex enough that a bit more explanation | 
|  | will help people figure out the intent of the test. | 
|  |  | 
|  | For scenario tests a class level docstring describing the steps in the scenario | 
|  | is required. If there is more than one test case in the class individual | 
|  | docstrings for the workflow in each test methods can be used instead. A good | 
|  | example of this would be:: | 
|  |  | 
|  | class TestServerBasicOps(manager.ScenarioTest): | 
|  |  | 
|  | """The test suite for server basic operations | 
|  |  | 
|  | This smoke test case follows this basic set of operations: | 
|  | * Create a keypair for use in launching an instance | 
|  | * Create a security group to control network access in instance | 
|  | * Add simple permissive rules to the security group | 
|  | * Launch an instance | 
|  | * Perform ssh to instance | 
|  | * Verify metadata service | 
|  | * Verify metadata on config_drive | 
|  | * Terminate the instance | 
|  | """ | 
|  |  | 
|  | Test Identification with Idempotent ID | 
|  | -------------------------------------- | 
|  |  | 
|  | Every function that provides a test must have an ``idempotent_id`` decorator | 
|  | that is a unique ``uuid-4`` instance. This ID is used to complement the fully | 
|  | qualified test name and track test functionality through refactoring. The | 
|  | format of the metadata looks like:: | 
|  |  | 
|  | @decorators.idempotent_id('585e934c-448e-43c4-acbf-d06a9b899997') | 
|  | def test_list_servers_with_detail(self): | 
|  | # The created server should be in the detailed list of all servers | 
|  | ... | 
|  |  | 
|  | Tempest.lib includes a ``check-uuid`` tool that will test for the existence | 
|  | and uniqueness of idempotent_id metadata for every test. If you have | 
|  | Tempest installed you run the tool against Tempest by calling from the | 
|  | Tempest repo:: | 
|  |  | 
|  | check-uuid | 
|  |  | 
|  | It can be invoked against any test suite by passing a package name:: | 
|  |  | 
|  | check-uuid --package <package_name> | 
|  |  | 
|  | Tests without an ``idempotent_id`` can be automatically fixed by running | 
|  | the command with the ``--fix`` flag, which will modify the source package | 
|  | by inserting randomly generated uuids for every test that does not have | 
|  | one:: | 
|  |  | 
|  | check-uuid --fix | 
|  |  | 
|  | The ``check-uuid`` tool is used as part of the Tempest gate job | 
|  | to ensure that all tests have an ``idempotent_id`` decorator. | 
|  |  | 
|  | Branchless Tempest Considerations | 
|  | --------------------------------- | 
|  |  | 
|  | Starting with the OpenStack Icehouse release Tempest no longer has any stable | 
|  | branches. This is to better ensure API consistency between releases because | 
|  | the API behavior should not change between releases. This means that the stable | 
|  | branches are also gated by the Tempest master branch, which also means that | 
|  | proposed commits to Tempest must work against both the master and all the | 
|  | currently supported stable branches of the projects. As such there are a few | 
|  | special considerations that have to be accounted for when pushing new changes | 
|  | to Tempest. | 
|  |  | 
|  | 1. New Tests for new features | 
|  | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ | 
|  |  | 
|  | When adding tests for new features that were not in previous releases of the | 
|  | projects the new test has to be properly skipped with a feature flag. This can | 
|  | be just as simple as using the ``@utils.requires_ext()`` or | 
|  | ``testtools.skipUnless`` decorators to check if the required extension (or | 
|  | discoverable optional API) or feature is enabled or can be as difficult as | 
|  | adding a new config option to the appropriate section. If there isn't a method | 
|  | of selecting the new **feature** from the config file then there won't be a | 
|  | mechanism to disable the test with older stable releases and the new test | 
|  | won't be able to merge. | 
|  |  | 
|  | Introduction of a new feature flag requires specifying a default value for | 
|  | the corresponding config option that is appropriate in the latest OpenStack | 
|  | release. Because Tempest is branchless, the feature flag's default value will | 
|  | need to be overridden to a value that is appropriate in earlier releases | 
|  | in which the feature isn't available. In DevStack, this can be accomplished | 
|  | by modifying Tempest's `lib installation script`_ for previous branches | 
|  | (because DevStack is branched). | 
|  |  | 
|  | .. _lib installation script: https://opendev.org/openstack/devstack/src/branch/master/lib/tempest | 
|  |  | 
|  | 2. Bug fix on core project needing Tempest changes | 
|  | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ | 
|  |  | 
|  | When trying to land a bug fix which changes a tested API you'll have to use the | 
|  | following procedure:: | 
|  |  | 
|  | 1. Propose change to the project, get a +2 on the change even with failing | 
|  | 2. Propose skip on Tempest which will only be approved after the | 
|  | corresponding change in the project has a +2 on change | 
|  | 3. Land project change in master and all open stable branches (if required) | 
|  | 4. Land changed test in Tempest | 
|  |  | 
|  | Otherwise the bug fix won't be able to land in the project. | 
|  |  | 
|  | Handily, `Zuul's cross-repository dependencies | 
|  | <https://zuul-ci.org/docs/zuul/latest/gating.html#cross-project-dependencies>`_. | 
|  | can be leveraged to do without step 2 and to have steps 3 and 4 happen | 
|  | "atomically". To do that, make the patch written in step 1 to depend (refer to | 
|  | Zuul's documentation above) on the patch written in step 4. The commit message | 
|  | for the Tempest change should have a link to the Gerrit review that justifies | 
|  | that change. | 
|  |  | 
|  | 3. New Tests for existing features | 
|  | ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ | 
|  |  | 
|  | If a test is being added for a feature that exists in all the current releases | 
|  | of the projects then the only concern is that the API behavior is the same | 
|  | across all the versions of the project being tested. If the behavior is not | 
|  | consistent the test will not be able to merge. | 
|  |  | 
|  | API Stability | 
|  | ------------- | 
|  |  | 
|  | For new tests being added to Tempest the assumption is that the API being | 
|  | tested is considered stable and adheres to the OpenStack API stability | 
|  | guidelines. If an API is still considered experimental or in development then | 
|  | it should not be tested by Tempest until it is considered stable. |