blob: ec7ff6a225ebeb3f364d1880e4ee5a46247cc224 [file] [log] [blame]
Attila Fazekas23fdf1d2013-06-09 16:35:23 +02001Tempest Coding Guide
2====================
3
Joe Gordon1374f882013-07-12 17:00:34 +01004- Step 1: Read the OpenStack Style Commandments
Matthew Treinish97072c82013-10-01 11:54:15 -04005 http://docs.openstack.org/developer/hacking/
Joe Gordon1374f882013-07-12 17:00:34 +01006- Step 2: Read on
7
8Tempest Specific Commandments
9------------------------------
10
ghanshyam50f19472014-11-26 17:04:37 +090011- [T102] Cannot import OpenStack python clients in tempest/api &
12 tempest/scenario tests
Matthew Treinish5e4c0f22013-09-10 18:38:28 +000013- [T104] Scenario tests require a services decorator
Andrea Frittolia5ddd552014-08-19 18:30:00 +010014- [T105] Tests cannot use setUpClass/tearDownClass
Masayuki Igawafcacf962014-02-19 14:00:01 +090015- [T106] vim configuration should not be kept in source files.
Ken'ichi Ohmichi7581bcd2015-02-16 04:09:58 +000016- [T107] Check that a service tag isn't in the module path
Ken'ichi Ohmichi80369a92015-04-06 23:41:14 +000017- [T108] Check no hyphen at the end of rand_name() argument
John Warren3059a092015-08-31 15:34:49 -040018- [T109] Cannot use testtools.skip decorator; instead use
Andrea Frittoli (andreaf)1370baf2016-04-29 14:26:22 -050019 decorators.skip_because from tempest.lib
Ken'ichi Ohmichic0d96be2015-11-11 12:33:48 +000020- [T110] Check that service client names of GET should be consistent
Ken'ichi Ohmichi4f525f72016-03-25 15:20:01 -070021- [T111] Check that service client names of DELETE should be consistent
Ken'ichi Ohmichi0dc97472016-03-25 15:10:08 -070022- [T112] Check that tempest.lib should not import local tempest code
Ken'ichi Ohmichid079c892016-04-19 11:23:36 -070023- [T113] Check that tests use data_utils.rand_uuid() instead of uuid.uuid4()
Ghanshyam2a180b82014-06-16 13:54:22 +090024- [N322] Method's default argument shouldn't be mutable
Attila Fazekas23fdf1d2013-06-09 16:35:23 +020025
Matthew Treinish8b372892012-12-07 17:13:16 -050026Test Data/Configuration
27-----------------------
28- Assume nothing about existing test data
29- Tests should be self contained (provide their own data)
30- Clean up test data at the completion of each test
31- Use configuration files for values that will vary by environment
32
33
Attila Fazekas10fd63d2013-07-04 18:38:21 +020034Exception Handling
35------------------
36According to the ``The Zen of Python`` the
Attila Fazekas58d23302013-07-24 10:25:02 +020037``Errors should never pass silently.``
Attila Fazekas10fd63d2013-07-04 18:38:21 +020038Tempest usually runs in special environment (jenkins gate jobs), in every
39error or failure situation we should provide as much error related
40information as possible, because we usually do not have the chance to
41investigate the situation after the issue happened.
42
43In every test case the abnormal situations must be very verbosely explained,
44by the exception and the log.
45
46In most cases the very first issue is the most important information.
47
Mithil Arunbe067ec2014-11-05 15:58:50 +053048Try to avoid using ``try`` blocks in the test cases, as both the ``except``
49and ``finally`` blocks could replace the original exception,
Attila Fazekas10fd63d2013-07-04 18:38:21 +020050when the additional operations leads to another exception.
51
Mithil Arunbe067ec2014-11-05 15:58:50 +053052Just letting an exception to propagate, is not a bad idea in a test case,
Bruce R. Montague44a6a192013-12-17 09:06:04 -080053at all.
Attila Fazekas10fd63d2013-07-04 18:38:21 +020054
55Try to avoid using any exception handling construct which can hide the errors
56origin.
57
58If you really need to use a ``try`` block, please ensure the original
59exception at least logged. When the exception is logged you usually need
60to ``raise`` the same or a different exception anyway.
61
Chris Yeohc2ff7272013-07-22 22:25:25 +093062Use of ``self.addCleanup`` is often a good way to avoid having to catch
63exceptions and still ensure resources are correctly cleaned up if the
64test fails part way through.
65
Mithil Arunbe067ec2014-11-05 15:58:50 +053066Use the ``self.assert*`` methods provided by the unit test framework.
67This signals the failures early on.
Attila Fazekas10fd63d2013-07-04 18:38:21 +020068
Mithil Arunbe067ec2014-11-05 15:58:50 +053069Avoid using the ``self.fail`` alone, its stack trace will signal
Bruce R. Montague44a6a192013-12-17 09:06:04 -080070the ``self.fail`` line as the origin of the error.
Attila Fazekas10fd63d2013-07-04 18:38:21 +020071
72Avoid constructing complex boolean expressions for assertion.
Attila Fazekas7899d312013-08-16 09:18:17 +020073The ``self.assertTrue`` or ``self.assertFalse`` without a ``msg`` argument,
74will just tell you the single boolean value, and you will not know anything
75about the values used in the formula, the ``msg`` argument might be good enough
76for providing more information.
77
78Most other assert method can include more information by default.
Attila Fazekas10fd63d2013-07-04 18:38:21 +020079For example ``self.assertIn`` can include the whole set.
80
Matthew Treinishf45ba2e2015-08-24 15:05:01 -040081It is recommended to use testtools `matcher`_ for the more tricky assertions.
82You can implement your own specific `matcher`_ as well.
Attila Fazekas7899d312013-08-16 09:18:17 +020083
Matthew Treinishf45ba2e2015-08-24 15:05:01 -040084.. _matcher: http://testtools.readthedocs.org/en/latest/for-test-authors.html#matchers
Attila Fazekas7899d312013-08-16 09:18:17 +020085
Attila Fazekas10fd63d2013-07-04 18:38:21 +020086If the test case fails you can see the related logs and the information
87carried by the exception (exception class, backtrack and exception info).
Mithil Arunbe067ec2014-11-05 15:58:50 +053088This and the service logs are your only guide to finding the root cause of flaky
89issues.
Attila Fazekas10fd63d2013-07-04 18:38:21 +020090
Attila Fazekas7899d312013-08-16 09:18:17 +020091Test cases are independent
92--------------------------
93Every ``test_method`` must be callable individually and MUST NOT depends on,
94any other ``test_method`` or ``test_method`` ordering.
95
96Test cases MAY depend on commonly initialized resources/facilities, like
97credentials management, testresources and so on. These facilities, MUST be able
Mithil Arunbe067ec2014-11-05 15:58:50 +053098to work even if just one ``test_method`` is selected for execution.
Attila Fazekas7899d312013-08-16 09:18:17 +020099
Matthew Treinish5e4c0f22013-09-10 18:38:28 +0000100Service Tagging
101---------------
102Service tagging is used to specify which services are exercised by a particular
103test method. You specify the services with the tempest.test.services decorator.
104For example:
105
106@services('compute', 'image')
107
108Valid service tag names are the same as the list of directories in tempest.api
109that have tests.
110
111For scenario tests having a service tag is required. For the api tests service
112tags are only needed if the test method makes an api call (either directly or
113indirectly through another service) that differs from the parent directory
114name. For example, any test that make an api call to a service other than nova
115in tempest.api.compute would require a service tag for those services, however
116they do not need to be tagged as compute.
117
Andrea Frittolia5ddd552014-08-19 18:30:00 +0100118Test fixtures and resources
119---------------------------
120Test level resources should be cleaned-up after the test execution. Clean-up
121is best scheduled using `addCleanup` which ensures that the resource cleanup
122code is always invoked, and in reverse order with respect to the creation
123order.
124
125Test class level resources should be defined in the `resource_setup` method of
126the test class, except for any credential obtained from the credentials
127provider, which should be set-up in the `setup_credentials` method.
128
129The test base class `BaseTestCase` defines Tempest framework for class level
130fixtures. `setUpClass` and `tearDownClass` are defined here and cannot be
131overwritten by subclasses (enforced via hacking rule T105).
132
133Set-up is split in a series of steps (setup stages), which can be overwritten
134by test classes. Set-up stages are:
Masayuki Igawae63cf0f2016-05-25 10:25:21 +0900135
136- `skip_checks`
137- `setup_credentials`
138- `setup_clients`
139- `resource_setup`
Andrea Frittolia5ddd552014-08-19 18:30:00 +0100140
141Tear-down is also split in a series of steps (teardown stages), which are
142stacked for execution only if the corresponding setup stage had been
143reached during the setup phase. Tear-down stages are:
Masayuki Igawae63cf0f2016-05-25 10:25:21 +0900144
145- `clear_credentials` (defined in the base test class)
146- `resource_cleanup`
Andrea Frittolia5ddd552014-08-19 18:30:00 +0100147
148Skipping Tests
149--------------
150Skipping tests should be based on configuration only. If that is not possible,
151it is likely that either a configuration flag is missing, or the test should
152fail rather than be skipped.
153Using discovery for skipping tests is generally discouraged.
154
155When running a test that requires a certain "feature" in the target
156cloud, if that feature is missing we should fail, because either the test
157configuration is invalid, or the cloud is broken and the expected "feature" is
158not there even if the cloud was configured with it.
159
Matthew Treinish8b79bb32013-10-10 17:11:05 -0400160Negative Tests
161--------------
Luz Cazarese28c18f2016-04-29 08:53:04 -0700162TODO: Write the guideline related to negative tests.
Matthew Treinish8b79bb32013-10-10 17:11:05 -0400163
Giulio Fidente83181a92013-10-01 06:02:24 +0200164Test skips because of Known Bugs
165--------------------------------
166
167If a test is broken because of a bug it is appropriate to skip the test until
168bug has been fixed. You should use the skip_because decorator so that
169Tempest's skip tracking tool can watch the bug status.
170
171Example::
172
173 @skip_because(bug="980688")
174 def test_this_and_that(self):
175 ...
176
Chris Yeohc2ff7272013-07-22 22:25:25 +0930177Guidelines
178----------
179- Do not submit changesets with only testcases which are skipped as
180 they will not be merged.
181- Consistently check the status code of responses in testcases. The
182 earlier a problem is detected the easier it is to debug, especially
183 where there is complicated setup required.
Matthew Treinish96c28d12013-09-16 17:05:09 +0000184
DennyZhang900f02b2013-09-23 08:34:04 -0500185Parallel Test Execution
186-----------------------
Matthew Treinish96c28d12013-09-16 17:05:09 +0000187Tempest by default runs its tests in parallel this creates the possibility for
188interesting interactions between tests which can cause unexpected failures.
Andrea Frittoli (andreaf)17209bb2015-05-22 10:16:57 -0700189Dynamic credentials provides protection from most of the potential race
190conditions between tests outside the same class. But there are still a few of
191things to watch out for to try to avoid issues when running your tests in
192parallel.
Matthew Treinish96c28d12013-09-16 17:05:09 +0000193
Sean Dagueed6e5862016-04-04 10:49:13 -0400194- Resources outside of a project scope still have the potential to conflict. This
Matthew Treinish96c28d12013-09-16 17:05:09 +0000195 is a larger concern for the admin tests since most resources and actions that
Sean Dagueed6e5862016-04-04 10:49:13 -0400196 require admin privileges are outside of projects.
Matthew Treinish96c28d12013-09-16 17:05:09 +0000197
198- Races between methods in the same class are not a problem because
199 parallelization in tempest is at the test class level, but if there is a json
200 and xml version of the same test class there could still be a race between
201 methods.
202
203- The rand_name() function from tempest.common.utils.data_utils should be used
204 anywhere a resource is created with a name. Static naming should be avoided
205 to prevent resource conflicts.
206
207- If the execution of a set of tests is required to be serialized then locking
208 can be used to perform this. See AggregatesAdminTest in
209 tempest.api.compute.admin for an example of using locking.
Marc Koderer31fe4832013-11-06 17:02:03 +0100210
211Stress Tests in Tempest
212-----------------------
213Any tempest test case can be flagged as a stress test. With this flag it will
214be automatically discovery and used in the stress test runs. The stress test
215framework itself is a facility to spawn and control worker processes in order
216to find race conditions (see ``tempest/stress/`` for more information). Please
217note that these stress tests can't be used for benchmarking purposes since they
218don't measure any performance characteristics.
219
220Example::
221
222 @stresstest(class_setup_per='process')
223 def test_this_and_that(self):
224 ...
225
226This will flag the test ``test_this_and_that`` as a stress test. The parameter
227``class_setup_per`` gives control when the setUpClass function should be called.
228
229Good candidates for stress tests are:
230
231- Scenario tests
232- API tests that have a wide focus
Matthew Treinish6eb05852013-11-26 15:28:12 +0000233
234Sample Configuration File
235-------------------------
236The sample config file is autogenerated using a script. If any changes are made
David Kranzfb0f51f2014-11-11 14:07:20 -0500237to the config variables in tempest/config.py then the sample config file must be
238regenerated. This can be done running::
239
240 tox -egenconfig
Matthew Treinishecf212c2013-12-06 18:23:54 +0000241
242Unit Tests
243----------
244Unit tests are a separate class of tests in tempest. They verify tempest
245itself, and thus have a different set of guidelines around them:
246
2471. They can not require anything running externally. All you should need to
248 run the unit tests is the git tree, python and the dependencies installed.
249 This includes running services, a config file, etc.
250
2512. The unit tests cannot use setUpClass, instead fixtures and testresources
252 should be used for shared state between tests.
Matthew Treinish55078882014-08-12 19:01:34 -0400253
254
255.. _TestDocumentation:
256
257Test Documentation
258------------------
259For tests being added we need to require inline documentation in the form of
Xicheng Chang6fb98ec2015-08-13 14:02:52 -0700260docstrings to explain what is being tested. In API tests for a new API a class
Matthew Treinish55078882014-08-12 19:01:34 -0400261level docstring should be added to an API reference doc. If one doesn't exist
262a TODO comment should be put indicating that the reference needs to be added.
263For individual API test cases a method level docstring should be used to
264explain the functionality being tested if the test name isn't descriptive
265enough. For example::
266
267 def test_get_role_by_id(self):
268 """Get a role by its id."""
269
270the docstring there is superfluous and shouldn't be added. but for a method
271like::
272
273 def test_volume_backup_create_get_detailed_list_restore_delete(self):
274 pass
275
276a docstring would be useful because while the test title is fairly descriptive
277the operations being performed are complex enough that a bit more explanation
278will help people figure out the intent of the test.
279
280For scenario tests a class level docstring describing the steps in the scenario
281is required. If there is more than one test case in the class individual
282docstrings for the workflow in each test methods can be used instead. A good
283example of this would be::
284
Masayuki Igawa93424e52014-10-06 13:54:26 +0900285 class TestVolumeBootPattern(manager.ScenarioTest):
Dougal Matthews4bebca02014-10-28 08:36:04 +0000286 """
287 This test case attempts to reproduce the following steps:
Matthew Treinish55078882014-08-12 19:01:34 -0400288
Dougal Matthews4bebca02014-10-28 08:36:04 +0000289 * Create in Cinder some bootable volume importing a Glance image
290 * Boot an instance from the bootable volume
291 * Write content to the volume
292 * Delete an instance and Boot a new instance from the volume
293 * Check written content in the instance
294 * Create a volume snapshot while the instance is running
295 * Boot an additional instance from the new snapshot based volume
296 * Check written content in the instance booted from snapshot
297 """
Matthew Treinisha970d652015-03-11 15:39:24 -0400298
Chris Hoge0e000ed2015-07-28 14:19:53 -0500299Test Identification with Idempotent ID
300--------------------------------------
301
302Every function that provides a test must have an ``idempotent_id`` decorator
303that is a unique ``uuid-4`` instance. This ID is used to complement the fully
Naomichi Wakuidbe9aab2015-08-26 03:36:02 +0000304qualified test name and track test functionality through refactoring. The
Chris Hoge0e000ed2015-07-28 14:19:53 -0500305format of the metadata looks like::
306
307 @test.idempotent_id('585e934c-448e-43c4-acbf-d06a9b899997')
308 def test_list_servers_with_detail(self):
309 # The created server should be in the detailed list of all servers
310 ...
311
Andrea Frittoli (andreaf)1370baf2016-04-29 14:26:22 -0500312Tempest.lib includes a ``check-uuid`` tool that will test for the existence
Matthew Treinishc1802bc2015-12-03 18:48:11 -0500313and uniqueness of idempotent_id metadata for every test. If you have
Andrea Frittoli (andreaf)1370baf2016-04-29 14:26:22 -0500314tempest installed you run the tool against Tempest by calling from the
Matthew Treinishc1802bc2015-12-03 18:48:11 -0500315tempest repo::
Chris Hoge0e000ed2015-07-28 14:19:53 -0500316
Matthew Treinishc1802bc2015-12-03 18:48:11 -0500317 check-uuid
Chris Hoge0e000ed2015-07-28 14:19:53 -0500318
319It can be invoked against any test suite by passing a package name::
320
Matthew Treinishc1802bc2015-12-03 18:48:11 -0500321 check-uuid --package <package_name>
Chris Hoge0e000ed2015-07-28 14:19:53 -0500322
323Tests without an ``idempotent_id`` can be automatically fixed by running
324the command with the ``--fix`` flag, which will modify the source package
325by inserting randomly generated uuids for every test that does not have
326one::
327
Matthew Treinishc1802bc2015-12-03 18:48:11 -0500328 check-uuid --fix
Chris Hoge0e000ed2015-07-28 14:19:53 -0500329
Matthew Treinishc1802bc2015-12-03 18:48:11 -0500330The ``check-uuid`` tool is used as part of the tempest gate job
Chris Hoge0e000ed2015-07-28 14:19:53 -0500331to ensure that all tests have an ``idempotent_id`` decorator.
332
Matthew Treinisha970d652015-03-11 15:39:24 -0400333Branchless Tempest Considerations
334---------------------------------
335
336Starting with the OpenStack Icehouse release Tempest no longer has any stable
337branches. This is to better ensure API consistency between releases because
338the API behavior should not change between releases. This means that the stable
339branches are also gated by the Tempest master branch, which also means that
340proposed commits to Tempest must work against both the master and all the
341currently supported stable branches of the projects. As such there are a few
342special considerations that have to be accounted for when pushing new changes
343to tempest.
344
3451. New Tests for new features
346^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
347
348When adding tests for new features that were not in previous releases of the
349projects the new test has to be properly skipped with a feature flag. Whether
350this is just as simple as using the @test.requires_ext() decorator to check
351if the required extension (or discoverable optional API) is enabled or adding
352a new config option to the appropriate section. If there isn't a method of
353selecting the new **feature** from the config file then there won't be a
354mechanism to disable the test with older stable releases and the new test won't
355be able to merge.
356
3572. Bug fix on core project needing Tempest changes
358^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
359
360When trying to land a bug fix which changes a tested API you'll have to use the
361following procedure::
362
363 - Propose change to the project, get a +2 on the change even with failing
364 - Propose skip on Tempest which will only be approved after the
365 corresponding change in the project has a +2 on change
366 - Land project change in master and all open stable branches (if required)
367 - Land changed test in Tempest
368
369Otherwise the bug fix won't be able to land in the project.
370
3713. New Tests for existing features
372^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
373
374If a test is being added for a feature that exists in all the current releases
375of the projects then the only concern is that the API behavior is the same
376across all the versions of the project being tested. If the behavior is not
377consistent the test will not be able to merge.
378
379API Stability
380-------------
381
382For new tests being added to Tempest the assumption is that the API being
383tested is considered stable and adheres to the OpenStack API stability
384guidelines. If an API is still considered experimental or in development then
385it should not be tested by Tempest until it is considered stable.