blob: 5b9c0f18367b2ccbe3d23a8350d1a1d7190f450c [file] [log] [blame]
Attila Fazekas23fdf1d2013-06-09 16:35:23 +02001Tempest Coding Guide
2====================
3
Joe Gordon1374f882013-07-12 17:00:34 +01004- Step 1: Read the OpenStack Style Commandments
chenxinge98720a2017-07-19 03:42:23 +00005 https://docs.openstack.org/hacking/latest/
Joe Gordon1374f882013-07-12 17:00:34 +01006- Step 2: Read on
7
8Tempest Specific Commandments
9------------------------------
10
ghanshyam50f19472014-11-26 17:04:37 +090011- [T102] Cannot import OpenStack python clients in tempest/api &
Masayuki Igawab78b9232017-11-17 16:12:37 +090012 tempest/scenario tests
Matthew Treinish5e4c0f22013-09-10 18:38:28 +000013- [T104] Scenario tests require a services decorator
Andrea Frittolia5ddd552014-08-19 18:30:00 +010014- [T105] Tests cannot use setUpClass/tearDownClass
Masayuki Igawafcacf962014-02-19 14:00:01 +090015- [T106] vim configuration should not be kept in source files.
Ken'ichi Ohmichi7581bcd2015-02-16 04:09:58 +000016- [T107] Check that a service tag isn't in the module path
Ken'ichi Ohmichi80369a92015-04-06 23:41:14 +000017- [T108] Check no hyphen at the end of rand_name() argument
John Warren3059a092015-08-31 15:34:49 -040018- [T109] Cannot use testtools.skip decorator; instead use
Masayuki Igawab78b9232017-11-17 16:12:37 +090019 decorators.skip_because from tempest.lib
Ken'ichi Ohmichic0d96be2015-11-11 12:33:48 +000020- [T110] Check that service client names of GET should be consistent
Ken'ichi Ohmichi4f525f72016-03-25 15:20:01 -070021- [T111] Check that service client names of DELETE should be consistent
Ken'ichi Ohmichi0dc97472016-03-25 15:10:08 -070022- [T112] Check that tempest.lib should not import local tempest code
Ken'ichi Ohmichid079c892016-04-19 11:23:36 -070023- [T113] Check that tests use data_utils.rand_uuid() instead of uuid.uuid4()
Matthew Treinish59d9eaa2016-05-31 23:42:55 -040024- [T114] Check that tempest.lib does not use tempest config
Ken'ichi Ohmichif741d0b2017-05-01 16:56:14 -070025- [T115] Check that admin tests should exist under admin path
Ghanshyam2a180b82014-06-16 13:54:22 +090026- [N322] Method's default argument shouldn't be mutable
junbolibc2ae862017-07-29 15:46:48 +080027- [T116] Unsupported 'message' Exception attribute in PY3
Felipe Monteiro4d011af2018-07-18 00:11:48 -040028- [T117] Check negative tests have ``@decorators.attr(type=['negative'])``
29 applied.
Attila Fazekas23fdf1d2013-06-09 16:35:23 +020030
Matthew Treinish8b372892012-12-07 17:13:16 -050031Test Data/Configuration
32-----------------------
33- Assume nothing about existing test data
34- Tests should be self contained (provide their own data)
35- Clean up test data at the completion of each test
36- Use configuration files for values that will vary by environment
37
38
Attila Fazekas10fd63d2013-07-04 18:38:21 +020039Exception Handling
40------------------
41According to the ``The Zen of Python`` the
Attila Fazekas58d23302013-07-24 10:25:02 +020042``Errors should never pass silently.``
Attila Fazekas10fd63d2013-07-04 18:38:21 +020043Tempest usually runs in special environment (jenkins gate jobs), in every
44error or failure situation we should provide as much error related
45information as possible, because we usually do not have the chance to
46investigate the situation after the issue happened.
47
48In every test case the abnormal situations must be very verbosely explained,
49by the exception and the log.
50
51In most cases the very first issue is the most important information.
52
Mithil Arunbe067ec2014-11-05 15:58:50 +053053Try to avoid using ``try`` blocks in the test cases, as both the ``except``
54and ``finally`` blocks could replace the original exception,
Attila Fazekas10fd63d2013-07-04 18:38:21 +020055when the additional operations leads to another exception.
56
Mithil Arunbe067ec2014-11-05 15:58:50 +053057Just letting an exception to propagate, is not a bad idea in a test case,
Bruce R. Montague44a6a192013-12-17 09:06:04 -080058at all.
Attila Fazekas10fd63d2013-07-04 18:38:21 +020059
60Try to avoid using any exception handling construct which can hide the errors
61origin.
62
63If you really need to use a ``try`` block, please ensure the original
64exception at least logged. When the exception is logged you usually need
65to ``raise`` the same or a different exception anyway.
66
Chris Yeohc2ff7272013-07-22 22:25:25 +093067Use of ``self.addCleanup`` is often a good way to avoid having to catch
68exceptions and still ensure resources are correctly cleaned up if the
69test fails part way through.
70
Mithil Arunbe067ec2014-11-05 15:58:50 +053071Use the ``self.assert*`` methods provided by the unit test framework.
72This signals the failures early on.
Attila Fazekas10fd63d2013-07-04 18:38:21 +020073
Mithil Arunbe067ec2014-11-05 15:58:50 +053074Avoid using the ``self.fail`` alone, its stack trace will signal
Bruce R. Montague44a6a192013-12-17 09:06:04 -080075the ``self.fail`` line as the origin of the error.
Attila Fazekas10fd63d2013-07-04 18:38:21 +020076
77Avoid constructing complex boolean expressions for assertion.
Attila Fazekas7899d312013-08-16 09:18:17 +020078The ``self.assertTrue`` or ``self.assertFalse`` without a ``msg`` argument,
79will just tell you the single boolean value, and you will not know anything
80about the values used in the formula, the ``msg`` argument might be good enough
81for providing more information.
82
83Most other assert method can include more information by default.
Attila Fazekas10fd63d2013-07-04 18:38:21 +020084For example ``self.assertIn`` can include the whole set.
85
Matthew Treinishf45ba2e2015-08-24 15:05:01 -040086It is recommended to use testtools `matcher`_ for the more tricky assertions.
87You can implement your own specific `matcher`_ as well.
Attila Fazekas7899d312013-08-16 09:18:17 +020088
davyyyac670dc2017-11-16 21:27:03 +080089.. _matcher: https://testtools.readthedocs.org/en/latest/for-test-authors.html#matchers
Attila Fazekas7899d312013-08-16 09:18:17 +020090
Attila Fazekas10fd63d2013-07-04 18:38:21 +020091If the test case fails you can see the related logs and the information
92carried by the exception (exception class, backtrack and exception info).
Mithil Arunbe067ec2014-11-05 15:58:50 +053093This and the service logs are your only guide to finding the root cause of flaky
94issues.
Attila Fazekas10fd63d2013-07-04 18:38:21 +020095
Attila Fazekas7899d312013-08-16 09:18:17 +020096Test cases are independent
97--------------------------
98Every ``test_method`` must be callable individually and MUST NOT depends on,
99any other ``test_method`` or ``test_method`` ordering.
100
101Test cases MAY depend on commonly initialized resources/facilities, like
102credentials management, testresources and so on. These facilities, MUST be able
Mithil Arunbe067ec2014-11-05 15:58:50 +0530103to work even if just one ``test_method`` is selected for execution.
Attila Fazekas7899d312013-08-16 09:18:17 +0200104
Matthew Treinish5e4c0f22013-09-10 18:38:28 +0000105Service Tagging
106---------------
107Service tagging is used to specify which services are exercised by a particular
mmkmmk57ce3bb9b2017-09-20 13:41:41 +0900108test method. You specify the services with the ``tempest.common.utils.services``
Jordan Pittier74a56ab2017-04-26 16:46:20 +0200109decorator. For example:
Matthew Treinish5e4c0f22013-09-10 18:38:28 +0000110
Felipe Monteiro46920b82018-07-09 23:58:20 -0400111``@utils.services('compute', 'image')``
Matthew Treinish5e4c0f22013-09-10 18:38:28 +0000112
113Valid service tag names are the same as the list of directories in tempest.api
114that have tests.
115
Jordan Pittier74a56ab2017-04-26 16:46:20 +0200116For scenario tests having a service tag is required. For the API tests service
117tags are only needed if the test method makes an API call (either directly or
Matthew Treinish5e4c0f22013-09-10 18:38:28 +0000118indirectly through another service) that differs from the parent directory
Jordan Pittier74a56ab2017-04-26 16:46:20 +0200119name. For example, any test that make an API call to a service other than Nova
120in ``tempest.api.compute`` would require a service tag for those services,
121however they do not need to be tagged as ``compute``.
Matthew Treinish5e4c0f22013-09-10 18:38:28 +0000122
Felipe Monteiro46920b82018-07-09 23:58:20 -0400123Test Attributes
124---------------
125Tempest leverages `test attributes`_ which are a simple but effective way of
126distinguishing between different "types" of API tests. A test can be "tagged"
127with such attributes using the ``decorators.attr`` decorator, for example::
128
129 @decorators.attr(type=['negative'])
130 def test_aggregate_create_aggregate_name_length_less_than_1(self):
131 [...]
132
133These test attributes can be used for test selection via regular expressions.
134For example, ``(?!.*\[.*\bslow\b.*\])(^tempest\.scenario)`` runs all the tests
135in the ``scenario`` test module, *except* for those tagged with the ``slow``
136attribute (via a negative lookahead in the regular expression). These
137attributes are used in Tempest's ``tox.ini`` as well as Tempest's Zuul job
138definitions for specifying particular batches of Tempest test suites to run.
139
140.. _test attributes: https://testtools.readthedocs.io/en/latest/for-test-authors.html?highlight=attr#test-attributes
141
142Negative Attribute
143^^^^^^^^^^^^^^^^^^
144The ``type='negative'`` attribute is used to signify that a test is a negative
145test, which is a test that handles invalid input gracefully. This attribute
146should be applied to all negative test scenarios.
147
148This attribute must be applied to each test that belongs to a negative test
149class, i.e. a test class name ending with "Negative.*" substring.
150
Felipe Monteiro46920b82018-07-09 23:58:20 -0400151Slow Attribute
152^^^^^^^^^^^^^^
153The ``type='slow'`` attribute is used to signify that a test takes a long time
154to run, relatively speaking. This attribute is usually applied to
155:ref:`scenario tests <scenario_field_guide>`, which involve a complicated
156series of API operations, the total runtime of which can be relatively long.
157This long runtime has performance implications on `Zuul`_ jobs, which is why
158the ``slow`` attribute is leveraged to run slow tests on a selective basis,
159to keep total `Zuul`_ job runtime down to a reasonable time frame.
160
161.. _Zuul: https://docs.openstack.org/infra/zuul/
162
163Smoke Attribute
164^^^^^^^^^^^^^^^
165The ``type='smoke'`` attribute is used to signify that a test is a so-called
166smoke test, which is a type of test that tests the most vital OpenStack
167functionality, like listing servers or flavors or creating volumes. The
168attribute should be sparingly applied to only the tests that sanity-check the
169most essential functionality of an OpenStack cloud.
170
Andrea Frittolia5ddd552014-08-19 18:30:00 +0100171Test fixtures and resources
172---------------------------
173Test level resources should be cleaned-up after the test execution. Clean-up
Masayuki Igawabbbaad62017-11-21 16:04:03 +0900174is best scheduled using ``addCleanup`` which ensures that the resource cleanup
Andrea Frittolia5ddd552014-08-19 18:30:00 +0100175code is always invoked, and in reverse order with respect to the creation
176order.
177
Masayuki Igawabbbaad62017-11-21 16:04:03 +0900178Test class level resources should be defined in the ``resource_setup`` method
179of the test class, except for any credential obtained from the credentials
180provider, which should be set-up in the ``setup_credentials`` method.
181Cleanup is best scheduled using ``addClassResourceCleanup`` which ensures that
Andrea Frittoli3be57482017-08-25 22:41:26 +0100182the cleanup code is always invoked, and in reverse order with respect to the
183creation order.
184
185In both cases - test level and class level cleanups - a wait loop should be
186scheduled before the actual delete of resources with an asynchronous delete.
Andrea Frittolia5ddd552014-08-19 18:30:00 +0100187
Masayuki Igawabbbaad62017-11-21 16:04:03 +0900188The test base class ``BaseTestCase`` defines Tempest framework for class level
189fixtures. ``setUpClass`` and ``tearDownClass`` are defined here and cannot be
Andrea Frittolia5ddd552014-08-19 18:30:00 +0100190overwritten by subclasses (enforced via hacking rule T105).
191
192Set-up is split in a series of steps (setup stages), which can be overwritten
193by test classes. Set-up stages are:
Masayuki Igawae63cf0f2016-05-25 10:25:21 +0900194
Masayuki Igawabbbaad62017-11-21 16:04:03 +0900195- ``skip_checks``
196- ``setup_credentials``
197- ``setup_clients``
198- ``resource_setup``
Andrea Frittolia5ddd552014-08-19 18:30:00 +0100199
200Tear-down is also split in a series of steps (teardown stages), which are
201stacked for execution only if the corresponding setup stage had been
202reached during the setup phase. Tear-down stages are:
Masayuki Igawae63cf0f2016-05-25 10:25:21 +0900203
Masayuki Igawabbbaad62017-11-21 16:04:03 +0900204- ``clear_credentials`` (defined in the base test class)
205- ``resource_cleanup``
Andrea Frittolia5ddd552014-08-19 18:30:00 +0100206
207Skipping Tests
208--------------
209Skipping tests should be based on configuration only. If that is not possible,
210it is likely that either a configuration flag is missing, or the test should
211fail rather than be skipped.
212Using discovery for skipping tests is generally discouraged.
213
214When running a test that requires a certain "feature" in the target
215cloud, if that feature is missing we should fail, because either the test
216configuration is invalid, or the cloud is broken and the expected "feature" is
217not there even if the cloud was configured with it.
218
Matthew Treinish8b79bb32013-10-10 17:11:05 -0400219Negative Tests
220--------------
Chris Hoge2b478412016-06-23 16:03:28 -0700221Error handling is an important aspect of API design and usage. Negative
222tests are a way to ensure that an application can gracefully handle
223invalid or unexpected input. However, as a black box integration test
224suite, Tempest is not suitable for handling all negative test cases, as
225the wide variety and complexity of negative tests can lead to long test
226runs and knowledge of internal implementation details. The bulk of
Ken'ichi Ohmichi8db40752016-09-28 14:43:05 -0700227negative testing should be handled with project function tests.
228All negative tests should be based on `API-WG guideline`_ . Such negative
229tests can block any changes from accurate failure code to invalid one.
230
davyyyac670dc2017-11-16 21:27:03 +0800231.. _API-WG guideline: https://specs.openstack.org/openstack/api-wg/guidelines/http.html#failure-code-clarifications
Ken'ichi Ohmichi8db40752016-09-28 14:43:05 -0700232
233If facing some gray area which is not clarified on the above guideline, propose
234a new guideline to the API-WG. With a proposal to the API-WG we will be able to
235build a consensus across all OpenStack projects and improve the quality and
236consistency of all the APIs.
237
238In addition, we have some guidelines for additional negative tests.
239
240- About BadRequest(HTTP400) case: We can add a single negative tests of
241 BadRequest for each resource and method(POST, PUT).
242 Please don't implement more negative tests on the same combination of
243 resource and method even if API request parameters are different from
244 the existing test.
245- About NotFound(HTTP404) case: We can add a single negative tests of
246 NotFound for each resource and method(GET, PUT, DELETE, HEAD).
247 Please don't implement more negative tests on the same combination
248 of resource and method.
249
250The above guidelines don't cover all cases and we will grow these guidelines
251organically over time. Patches outside of the above guidelines are left up to
252the reviewers' discretion and if we face some conflicts between reviewers, we
253will expand the guideline based on our discussion and experience.
Matthew Treinish8b79bb32013-10-10 17:11:05 -0400254
Giulio Fidente83181a92013-10-01 06:02:24 +0200255Test skips because of Known Bugs
256--------------------------------
Giulio Fidente83181a92013-10-01 06:02:24 +0200257If a test is broken because of a bug it is appropriate to skip the test until
Jordan Pittier74a56ab2017-04-26 16:46:20 +0200258bug has been fixed. You should use the ``skip_because`` decorator so that
Giulio Fidente83181a92013-10-01 06:02:24 +0200259Tempest's skip tracking tool can watch the bug status.
260
261Example::
262
263 @skip_because(bug="980688")
264 def test_this_and_that(self):
265 ...
266
Chris Yeohc2ff7272013-07-22 22:25:25 +0930267Guidelines
268----------
269- Do not submit changesets with only testcases which are skipped as
270 they will not be merged.
271- Consistently check the status code of responses in testcases. The
272 earlier a problem is detected the easier it is to debug, especially
273 where there is complicated setup required.
Matthew Treinish96c28d12013-09-16 17:05:09 +0000274
DennyZhang900f02b2013-09-23 08:34:04 -0500275Parallel Test Execution
276-----------------------
Matthew Treinish96c28d12013-09-16 17:05:09 +0000277Tempest by default runs its tests in parallel this creates the possibility for
278interesting interactions between tests which can cause unexpected failures.
Andrea Frittoli (andreaf)17209bb2015-05-22 10:16:57 -0700279Dynamic credentials provides protection from most of the potential race
280conditions between tests outside the same class. But there are still a few of
281things to watch out for to try to avoid issues when running your tests in
282parallel.
Matthew Treinish96c28d12013-09-16 17:05:09 +0000283
Sean Dagueed6e5862016-04-04 10:49:13 -0400284- Resources outside of a project scope still have the potential to conflict. This
Matthew Treinish96c28d12013-09-16 17:05:09 +0000285 is a larger concern for the admin tests since most resources and actions that
Sean Dagueed6e5862016-04-04 10:49:13 -0400286 require admin privileges are outside of projects.
Matthew Treinish96c28d12013-09-16 17:05:09 +0000287
288- Races between methods in the same class are not a problem because
Jordan Pittier74a56ab2017-04-26 16:46:20 +0200289 parallelization in Tempest is at the test class level, but if there is a json
Matthew Treinish96c28d12013-09-16 17:05:09 +0000290 and xml version of the same test class there could still be a race between
291 methods.
292
jeremy.zhangc0f95562017-05-26 13:41:57 +0800293- The rand_name() function from tempest.lib.common.utils.data_utils should be
294 used anywhere a resource is created with a name. Static naming should be
295 avoided to prevent resource conflicts.
Matthew Treinish96c28d12013-09-16 17:05:09 +0000296
297- If the execution of a set of tests is required to be serialized then locking
Jordan Pittier74a56ab2017-04-26 16:46:20 +0200298 can be used to perform this. See usage of ``LockFixture`` for examples of
299 using locking.
Marc Koderer31fe4832013-11-06 17:02:03 +0100300
Matthew Treinish6eb05852013-11-26 15:28:12 +0000301Sample Configuration File
302-------------------------
303The sample config file is autogenerated using a script. If any changes are made
David Kranzfb0f51f2014-11-11 14:07:20 -0500304to the config variables in tempest/config.py then the sample config file must be
305regenerated. This can be done running::
306
Hai Shi6f52fc52017-04-03 21:17:37 +0800307 tox -e genconfig
Matthew Treinishecf212c2013-12-06 18:23:54 +0000308
309Unit Tests
310----------
Jordan Pittier74a56ab2017-04-26 16:46:20 +0200311Unit tests are a separate class of tests in Tempest. They verify Tempest
Matthew Treinishecf212c2013-12-06 18:23:54 +0000312itself, and thus have a different set of guidelines around them:
313
3141. They can not require anything running externally. All you should need to
315 run the unit tests is the git tree, python and the dependencies installed.
316 This includes running services, a config file, etc.
317
3182. The unit tests cannot use setUpClass, instead fixtures and testresources
319 should be used for shared state between tests.
Matthew Treinish55078882014-08-12 19:01:34 -0400320
321
322.. _TestDocumentation:
323
324Test Documentation
325------------------
326For tests being added we need to require inline documentation in the form of
Xicheng Chang6fb98ec2015-08-13 14:02:52 -0700327docstrings to explain what is being tested. In API tests for a new API a class
Matthew Treinish55078882014-08-12 19:01:34 -0400328level docstring should be added to an API reference doc. If one doesn't exist
329a TODO comment should be put indicating that the reference needs to be added.
330For individual API test cases a method level docstring should be used to
331explain the functionality being tested if the test name isn't descriptive
332enough. For example::
333
334 def test_get_role_by_id(self):
335 """Get a role by its id."""
336
337the docstring there is superfluous and shouldn't be added. but for a method
338like::
339
340 def test_volume_backup_create_get_detailed_list_restore_delete(self):
341 pass
342
343a docstring would be useful because while the test title is fairly descriptive
344the operations being performed are complex enough that a bit more explanation
345will help people figure out the intent of the test.
346
347For scenario tests a class level docstring describing the steps in the scenario
348is required. If there is more than one test case in the class individual
349docstrings for the workflow in each test methods can be used instead. A good
350example of this would be::
351
Masayuki Igawa93424e52014-10-06 13:54:26 +0900352 class TestVolumeBootPattern(manager.ScenarioTest):
Dougal Matthews4bebca02014-10-28 08:36:04 +0000353 """
354 This test case attempts to reproduce the following steps:
Matthew Treinish55078882014-08-12 19:01:34 -0400355
Dougal Matthews4bebca02014-10-28 08:36:04 +0000356 * Create in Cinder some bootable volume importing a Glance image
357 * Boot an instance from the bootable volume
358 * Write content to the volume
359 * Delete an instance and Boot a new instance from the volume
360 * Check written content in the instance
361 * Create a volume snapshot while the instance is running
362 * Boot an additional instance from the new snapshot based volume
363 * Check written content in the instance booted from snapshot
364 """
Matthew Treinisha970d652015-03-11 15:39:24 -0400365
Chris Hoge0e000ed2015-07-28 14:19:53 -0500366Test Identification with Idempotent ID
367--------------------------------------
368
369Every function that provides a test must have an ``idempotent_id`` decorator
370that is a unique ``uuid-4`` instance. This ID is used to complement the fully
Naomichi Wakuidbe9aab2015-08-26 03:36:02 +0000371qualified test name and track test functionality through refactoring. The
Chris Hoge0e000ed2015-07-28 14:19:53 -0500372format of the metadata looks like::
373
Ken'ichi Ohmichi8a082112017-03-06 16:03:17 -0800374 @decorators.idempotent_id('585e934c-448e-43c4-acbf-d06a9b899997')
Chris Hoge0e000ed2015-07-28 14:19:53 -0500375 def test_list_servers_with_detail(self):
376 # The created server should be in the detailed list of all servers
377 ...
378
Andrea Frittoli (andreaf)1370baf2016-04-29 14:26:22 -0500379Tempest.lib includes a ``check-uuid`` tool that will test for the existence
Matthew Treinishc1802bc2015-12-03 18:48:11 -0500380and uniqueness of idempotent_id metadata for every test. If you have
Jordan Pittier74a56ab2017-04-26 16:46:20 +0200381Tempest installed you run the tool against Tempest by calling from the
382Tempest repo::
Chris Hoge0e000ed2015-07-28 14:19:53 -0500383
Matthew Treinishc1802bc2015-12-03 18:48:11 -0500384 check-uuid
Chris Hoge0e000ed2015-07-28 14:19:53 -0500385
386It can be invoked against any test suite by passing a package name::
387
Matthew Treinishc1802bc2015-12-03 18:48:11 -0500388 check-uuid --package <package_name>
Chris Hoge0e000ed2015-07-28 14:19:53 -0500389
390Tests without an ``idempotent_id`` can be automatically fixed by running
391the command with the ``--fix`` flag, which will modify the source package
392by inserting randomly generated uuids for every test that does not have
393one::
394
Matthew Treinishc1802bc2015-12-03 18:48:11 -0500395 check-uuid --fix
Chris Hoge0e000ed2015-07-28 14:19:53 -0500396
Jordan Pittier74a56ab2017-04-26 16:46:20 +0200397The ``check-uuid`` tool is used as part of the Tempest gate job
Chris Hoge0e000ed2015-07-28 14:19:53 -0500398to ensure that all tests have an ``idempotent_id`` decorator.
399
Matthew Treinisha970d652015-03-11 15:39:24 -0400400Branchless Tempest Considerations
401---------------------------------
402
403Starting with the OpenStack Icehouse release Tempest no longer has any stable
404branches. This is to better ensure API consistency between releases because
405the API behavior should not change between releases. This means that the stable
406branches are also gated by the Tempest master branch, which also means that
407proposed commits to Tempest must work against both the master and all the
408currently supported stable branches of the projects. As such there are a few
409special considerations that have to be accounted for when pushing new changes
Jordan Pittier74a56ab2017-04-26 16:46:20 +0200410to Tempest.
Matthew Treinisha970d652015-03-11 15:39:24 -0400411
4121. New Tests for new features
413^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
414
415When adding tests for new features that were not in previous releases of the
Felipe Monteiro356f0592018-03-26 21:51:52 -0400416projects the new test has to be properly skipped with a feature flag. This can
417be just as simple as using the ``@utils.requires_ext()`` or
418``testtools.skipUnless`` decorators to check if the required extension (or
419discoverable optional API) or feature is enabled or can be as difficult as
Andrea Frittolicd368412017-08-14 21:37:56 +0100420adding a new config option to the appropriate section. If there isn't a method
421of selecting the new **feature** from the config file then there won't be a
Felipe Monteiro356f0592018-03-26 21:51:52 -0400422mechanism to disable the test with older stable releases and the new test
423won't be able to merge.
424
425Introduction of a new feature flag requires specifying a default value for
426the corresponding config option that is appropriate in the latest OpenStack
427release. Because Tempest is branchless, the feature flag's default value will
428need to be overridden to a value that is appropriate in earlier releases
429in which the feature isn't available. In DevStack, this can be accomplished
430by modifying Tempest's `lib installation script`_ for previous branches
431(because DevStack is branched).
432
433.. _lib installation script: http://git.openstack.org/cgit/openstack-dev/devstack/tree/lib/tempest
Matthew Treinisha970d652015-03-11 15:39:24 -0400434
4352. Bug fix on core project needing Tempest changes
436^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
437
438When trying to land a bug fix which changes a tested API you'll have to use the
439following procedure::
440
Jordan Pittier74a56ab2017-04-26 16:46:20 +0200441 1. Propose change to the project, get a +2 on the change even with failing
442 2. Propose skip on Tempest which will only be approved after the
Matthew Treinisha970d652015-03-11 15:39:24 -0400443 corresponding change in the project has a +2 on change
Jordan Pittier74a56ab2017-04-26 16:46:20 +0200444 3. Land project change in master and all open stable branches (if required)
445 4. Land changed test in Tempest
Matthew Treinisha970d652015-03-11 15:39:24 -0400446
447Otherwise the bug fix won't be able to land in the project.
448
gaofei6ec582f2018-01-24 14:08:36 +0800449Handily, `Zuul's cross-repository dependencies
junboli477fd022017-09-06 17:25:11 +0800450<https://docs.openstack.org/infra/zuul/user/gating.html#cross-project-dependencies>`_.
Jordan Pittier74a56ab2017-04-26 16:46:20 +0200451can be leveraged to do without step 2 and to have steps 3 and 4 happen
452"atomically". To do that, make the patch written in step 1 to depend (refer to
453Zuul's documentation above) on the patch written in step 4. The commit message
454for the Tempest change should have a link to the Gerrit review that justifies
455that change.
456
Matthew Treinisha970d652015-03-11 15:39:24 -04004573. New Tests for existing features
458^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
459
460If a test is being added for a feature that exists in all the current releases
461of the projects then the only concern is that the API behavior is the same
462across all the versions of the project being tested. If the behavior is not
463consistent the test will not be able to merge.
464
465API Stability
466-------------
467
468For new tests being added to Tempest the assumption is that the API being
469tested is considered stable and adheres to the OpenStack API stability
470guidelines. If an API is still considered experimental or in development then
471it should not be tested by Tempest until it is considered stable.