blob: e15e213d3d24f350bcc6af3ce2eb8145268dba12 [file] [log] [blame]
Attila Fazekas23fdf1d2013-06-09 16:35:23 +02001Tempest Coding Guide
2====================
3
Joe Gordon1374f882013-07-12 17:00:34 +01004- Step 1: Read the OpenStack Style Commandments
Matthew Treinish97072c82013-10-01 11:54:15 -04005 http://docs.openstack.org/developer/hacking/
Joe Gordon1374f882013-07-12 17:00:34 +01006- Step 2: Read on
7
8Tempest Specific Commandments
9------------------------------
10
ghanshyam50f19472014-11-26 17:04:37 +090011- [T102] Cannot import OpenStack python clients in tempest/api &
12 tempest/scenario tests
Matthew Treinish5e4c0f22013-09-10 18:38:28 +000013- [T104] Scenario tests require a services decorator
Andrea Frittolia5ddd552014-08-19 18:30:00 +010014- [T105] Tests cannot use setUpClass/tearDownClass
Masayuki Igawafcacf962014-02-19 14:00:01 +090015- [T106] vim configuration should not be kept in source files.
Ken'ichi Ohmichi7581bcd2015-02-16 04:09:58 +000016- [T107] Check that a service tag isn't in the module path
Ken'ichi Ohmichi80369a92015-04-06 23:41:14 +000017- [T108] Check no hyphen at the end of rand_name() argument
John Warren3059a092015-08-31 15:34:49 -040018- [T109] Cannot use testtools.skip decorator; instead use
19 decorators.skip_because from tempest-lib
Ghanshyam2a180b82014-06-16 13:54:22 +090020- [N322] Method's default argument shouldn't be mutable
Attila Fazekas23fdf1d2013-06-09 16:35:23 +020021
Matthew Treinish8b372892012-12-07 17:13:16 -050022Test Data/Configuration
23-----------------------
24- Assume nothing about existing test data
25- Tests should be self contained (provide their own data)
26- Clean up test data at the completion of each test
27- Use configuration files for values that will vary by environment
28
29
Attila Fazekas10fd63d2013-07-04 18:38:21 +020030Exception Handling
31------------------
32According to the ``The Zen of Python`` the
Attila Fazekas58d23302013-07-24 10:25:02 +020033``Errors should never pass silently.``
Attila Fazekas10fd63d2013-07-04 18:38:21 +020034Tempest usually runs in special environment (jenkins gate jobs), in every
35error or failure situation we should provide as much error related
36information as possible, because we usually do not have the chance to
37investigate the situation after the issue happened.
38
39In every test case the abnormal situations must be very verbosely explained,
40by the exception and the log.
41
42In most cases the very first issue is the most important information.
43
Mithil Arunbe067ec2014-11-05 15:58:50 +053044Try to avoid using ``try`` blocks in the test cases, as both the ``except``
45and ``finally`` blocks could replace the original exception,
Attila Fazekas10fd63d2013-07-04 18:38:21 +020046when the additional operations leads to another exception.
47
Mithil Arunbe067ec2014-11-05 15:58:50 +053048Just letting an exception to propagate, is not a bad idea in a test case,
Bruce R. Montague44a6a192013-12-17 09:06:04 -080049at all.
Attila Fazekas10fd63d2013-07-04 18:38:21 +020050
51Try to avoid using any exception handling construct which can hide the errors
52origin.
53
54If you really need to use a ``try`` block, please ensure the original
55exception at least logged. When the exception is logged you usually need
56to ``raise`` the same or a different exception anyway.
57
Chris Yeohc2ff7272013-07-22 22:25:25 +093058Use of ``self.addCleanup`` is often a good way to avoid having to catch
59exceptions and still ensure resources are correctly cleaned up if the
60test fails part way through.
61
Mithil Arunbe067ec2014-11-05 15:58:50 +053062Use the ``self.assert*`` methods provided by the unit test framework.
63This signals the failures early on.
Attila Fazekas10fd63d2013-07-04 18:38:21 +020064
Mithil Arunbe067ec2014-11-05 15:58:50 +053065Avoid using the ``self.fail`` alone, its stack trace will signal
Bruce R. Montague44a6a192013-12-17 09:06:04 -080066the ``self.fail`` line as the origin of the error.
Attila Fazekas10fd63d2013-07-04 18:38:21 +020067
68Avoid constructing complex boolean expressions for assertion.
Attila Fazekas7899d312013-08-16 09:18:17 +020069The ``self.assertTrue`` or ``self.assertFalse`` without a ``msg`` argument,
70will just tell you the single boolean value, and you will not know anything
71about the values used in the formula, the ``msg`` argument might be good enough
72for providing more information.
73
74Most other assert method can include more information by default.
Attila Fazekas10fd63d2013-07-04 18:38:21 +020075For example ``self.assertIn`` can include the whole set.
76
Matthew Treinishf45ba2e2015-08-24 15:05:01 -040077It is recommended to use testtools `matcher`_ for the more tricky assertions.
78You can implement your own specific `matcher`_ as well.
Attila Fazekas7899d312013-08-16 09:18:17 +020079
Matthew Treinishf45ba2e2015-08-24 15:05:01 -040080.. _matcher: http://testtools.readthedocs.org/en/latest/for-test-authors.html#matchers
Attila Fazekas7899d312013-08-16 09:18:17 +020081
Attila Fazekas10fd63d2013-07-04 18:38:21 +020082If the test case fails you can see the related logs and the information
83carried by the exception (exception class, backtrack and exception info).
Mithil Arunbe067ec2014-11-05 15:58:50 +053084This and the service logs are your only guide to finding the root cause of flaky
85issues.
Attila Fazekas10fd63d2013-07-04 18:38:21 +020086
Attila Fazekas7899d312013-08-16 09:18:17 +020087Test cases are independent
88--------------------------
89Every ``test_method`` must be callable individually and MUST NOT depends on,
90any other ``test_method`` or ``test_method`` ordering.
91
92Test cases MAY depend on commonly initialized resources/facilities, like
93credentials management, testresources and so on. These facilities, MUST be able
Mithil Arunbe067ec2014-11-05 15:58:50 +053094to work even if just one ``test_method`` is selected for execution.
Attila Fazekas7899d312013-08-16 09:18:17 +020095
Matthew Treinish5e4c0f22013-09-10 18:38:28 +000096Service Tagging
97---------------
98Service tagging is used to specify which services are exercised by a particular
99test method. You specify the services with the tempest.test.services decorator.
100For example:
101
102@services('compute', 'image')
103
104Valid service tag names are the same as the list of directories in tempest.api
105that have tests.
106
107For scenario tests having a service tag is required. For the api tests service
108tags are only needed if the test method makes an api call (either directly or
109indirectly through another service) that differs from the parent directory
110name. For example, any test that make an api call to a service other than nova
111in tempest.api.compute would require a service tag for those services, however
112they do not need to be tagged as compute.
113
Andrea Frittolia5ddd552014-08-19 18:30:00 +0100114Test fixtures and resources
115---------------------------
116Test level resources should be cleaned-up after the test execution. Clean-up
117is best scheduled using `addCleanup` which ensures that the resource cleanup
118code is always invoked, and in reverse order with respect to the creation
119order.
120
121Test class level resources should be defined in the `resource_setup` method of
122the test class, except for any credential obtained from the credentials
123provider, which should be set-up in the `setup_credentials` method.
124
125The test base class `BaseTestCase` defines Tempest framework for class level
126fixtures. `setUpClass` and `tearDownClass` are defined here and cannot be
127overwritten by subclasses (enforced via hacking rule T105).
128
129Set-up is split in a series of steps (setup stages), which can be overwritten
130by test classes. Set-up stages are:
131- `skip_checks`
132- `setup_credentials`
133- `setup_clients`
134- `resource_setup`
135
136Tear-down is also split in a series of steps (teardown stages), which are
137stacked for execution only if the corresponding setup stage had been
138reached during the setup phase. Tear-down stages are:
139- `clear_isolated_creds` (defined in the base test class)
140- `resource_cleanup`
141
142Skipping Tests
143--------------
144Skipping tests should be based on configuration only. If that is not possible,
145it is likely that either a configuration flag is missing, or the test should
146fail rather than be skipped.
147Using discovery for skipping tests is generally discouraged.
148
149When running a test that requires a certain "feature" in the target
150cloud, if that feature is missing we should fail, because either the test
151configuration is invalid, or the cloud is broken and the expected "feature" is
152not there even if the cloud was configured with it.
153
Matthew Treinish8b79bb32013-10-10 17:11:05 -0400154Negative Tests
155--------------
Marc Koderera5afb4f2014-02-11 15:38:15 +0100156Newly added negative tests should use the negative test framework. First step
Marc Kodererb3875b02014-11-27 09:52:50 +0100157is to create an interface description in a python file under
158`tempest/api_schema/request/`. These descriptions consists of two important
159sections for the test (one of those is mandatory):
Matthew Treinish8b79bb32013-10-10 17:11:05 -0400160
Marc Koderera5afb4f2014-02-11 15:38:15 +0100161 - A resource (part of the URL of the request): Resources needed for a test
Matthew Treinishf45ba2e2015-08-24 15:05:01 -0400162 must be created in `setUpClass` and registered with `set_resource` e.g.:
163 `cls.set_resource("server", server['id'])`
Matthew Treinish8b79bb32013-10-10 17:11:05 -0400164
Marc Koderera5afb4f2014-02-11 15:38:15 +0100165 - A json schema: defines properties for a request.
166
167After that a test class must be added to automatically generate test scenarios
Marc Koderer313cbd52014-03-26 08:56:59 +0100168out of the given interface description::
169
170 load_tests = test.NegativeAutoTest.load_tests
Marc Koderera5afb4f2014-02-11 15:38:15 +0100171
Marc Kodererb3875b02014-11-27 09:52:50 +0100172 @test.SimpleNegativeAutoTest
173 class SampleTestNegativeTestJSON(<your base class>, test.NegativeAutoTest):
Marc Koderera5afb4f2014-02-11 15:38:15 +0100174 _service = 'compute'
Marc Kodererb3875b02014-11-27 09:52:50 +0100175 _schema = <your schema file>
Marc Koderera5afb4f2014-02-11 15:38:15 +0100176
Marc Kodererb3875b02014-11-27 09:52:50 +0100177The class decorator `SimpleNegativeAutoTest` will automatically generate test
178cases out of the given schema in the attribute `_schema`.
Marc Koderera5afb4f2014-02-11 15:38:15 +0100179
180All negative tests should be added into a separate negative test file.
181If such a file doesn't exist for the particular resource being tested a new
Marc Kodererb3875b02014-11-27 09:52:50 +0100182test file should be added.
Matthew Treinish8b79bb32013-10-10 17:11:05 -0400183
Giulio Fidente83181a92013-10-01 06:02:24 +0200184Test skips because of Known Bugs
185--------------------------------
186
187If a test is broken because of a bug it is appropriate to skip the test until
188bug has been fixed. You should use the skip_because decorator so that
189Tempest's skip tracking tool can watch the bug status.
190
191Example::
192
193 @skip_because(bug="980688")
194 def test_this_and_that(self):
195 ...
196
Chris Yeohc2ff7272013-07-22 22:25:25 +0930197Guidelines
198----------
199- Do not submit changesets with only testcases which are skipped as
200 they will not be merged.
201- Consistently check the status code of responses in testcases. The
202 earlier a problem is detected the easier it is to debug, especially
203 where there is complicated setup required.
Matthew Treinish96c28d12013-09-16 17:05:09 +0000204
DennyZhang900f02b2013-09-23 08:34:04 -0500205Parallel Test Execution
206-----------------------
Matthew Treinish96c28d12013-09-16 17:05:09 +0000207Tempest by default runs its tests in parallel this creates the possibility for
208interesting interactions between tests which can cause unexpected failures.
209Tenant isolation provides protection from most of the potential race conditions
210between tests outside the same class. But there are still a few of things to
211watch out for to try to avoid issues when running your tests in parallel.
212
213- Resources outside of a tenant scope still have the potential to conflict. This
214 is a larger concern for the admin tests since most resources and actions that
DennyZhang900f02b2013-09-23 08:34:04 -0500215 require admin privileges are outside of tenants.
Matthew Treinish96c28d12013-09-16 17:05:09 +0000216
217- Races between methods in the same class are not a problem because
218 parallelization in tempest is at the test class level, but if there is a json
219 and xml version of the same test class there could still be a race between
220 methods.
221
222- The rand_name() function from tempest.common.utils.data_utils should be used
223 anywhere a resource is created with a name. Static naming should be avoided
224 to prevent resource conflicts.
225
226- If the execution of a set of tests is required to be serialized then locking
227 can be used to perform this. See AggregatesAdminTest in
228 tempest.api.compute.admin for an example of using locking.
Marc Koderer31fe4832013-11-06 17:02:03 +0100229
230Stress Tests in Tempest
231-----------------------
232Any tempest test case can be flagged as a stress test. With this flag it will
233be automatically discovery and used in the stress test runs. The stress test
234framework itself is a facility to spawn and control worker processes in order
235to find race conditions (see ``tempest/stress/`` for more information). Please
236note that these stress tests can't be used for benchmarking purposes since they
237don't measure any performance characteristics.
238
239Example::
240
241 @stresstest(class_setup_per='process')
242 def test_this_and_that(self):
243 ...
244
245This will flag the test ``test_this_and_that`` as a stress test. The parameter
246``class_setup_per`` gives control when the setUpClass function should be called.
247
248Good candidates for stress tests are:
249
250- Scenario tests
251- API tests that have a wide focus
Matthew Treinish6eb05852013-11-26 15:28:12 +0000252
253Sample Configuration File
254-------------------------
255The sample config file is autogenerated using a script. If any changes are made
David Kranzfb0f51f2014-11-11 14:07:20 -0500256to the config variables in tempest/config.py then the sample config file must be
257regenerated. This can be done running::
258
259 tox -egenconfig
Matthew Treinishecf212c2013-12-06 18:23:54 +0000260
261Unit Tests
262----------
263Unit tests are a separate class of tests in tempest. They verify tempest
264itself, and thus have a different set of guidelines around them:
265
2661. They can not require anything running externally. All you should need to
267 run the unit tests is the git tree, python and the dependencies installed.
268 This includes running services, a config file, etc.
269
2702. The unit tests cannot use setUpClass, instead fixtures and testresources
271 should be used for shared state between tests.
Matthew Treinish55078882014-08-12 19:01:34 -0400272
273
274.. _TestDocumentation:
275
276Test Documentation
277------------------
278For tests being added we need to require inline documentation in the form of
Xicheng Chang6fb98ec2015-08-13 14:02:52 -0700279docstrings to explain what is being tested. In API tests for a new API a class
Matthew Treinish55078882014-08-12 19:01:34 -0400280level docstring should be added to an API reference doc. If one doesn't exist
281a TODO comment should be put indicating that the reference needs to be added.
282For individual API test cases a method level docstring should be used to
283explain the functionality being tested if the test name isn't descriptive
284enough. For example::
285
286 def test_get_role_by_id(self):
287 """Get a role by its id."""
288
289the docstring there is superfluous and shouldn't be added. but for a method
290like::
291
292 def test_volume_backup_create_get_detailed_list_restore_delete(self):
293 pass
294
295a docstring would be useful because while the test title is fairly descriptive
296the operations being performed are complex enough that a bit more explanation
297will help people figure out the intent of the test.
298
299For scenario tests a class level docstring describing the steps in the scenario
300is required. If there is more than one test case in the class individual
301docstrings for the workflow in each test methods can be used instead. A good
302example of this would be::
303
Masayuki Igawa93424e52014-10-06 13:54:26 +0900304 class TestVolumeBootPattern(manager.ScenarioTest):
Dougal Matthews4bebca02014-10-28 08:36:04 +0000305 """
306 This test case attempts to reproduce the following steps:
Matthew Treinish55078882014-08-12 19:01:34 -0400307
Dougal Matthews4bebca02014-10-28 08:36:04 +0000308 * Create in Cinder some bootable volume importing a Glance image
309 * Boot an instance from the bootable volume
310 * Write content to the volume
311 * Delete an instance and Boot a new instance from the volume
312 * Check written content in the instance
313 * Create a volume snapshot while the instance is running
314 * Boot an additional instance from the new snapshot based volume
315 * Check written content in the instance booted from snapshot
316 """
Matthew Treinisha970d652015-03-11 15:39:24 -0400317
Chris Hoge0e000ed2015-07-28 14:19:53 -0500318Test Identification with Idempotent ID
319--------------------------------------
320
321Every function that provides a test must have an ``idempotent_id`` decorator
322that is a unique ``uuid-4`` instance. This ID is used to complement the fully
Naomichi Wakuidbe9aab2015-08-26 03:36:02 +0000323qualified test name and track test functionality through refactoring. The
Chris Hoge0e000ed2015-07-28 14:19:53 -0500324format of the metadata looks like::
325
326 @test.idempotent_id('585e934c-448e-43c4-acbf-d06a9b899997')
327 def test_list_servers_with_detail(self):
328 # The created server should be in the detailed list of all servers
329 ...
330
331Tempest includes a ``check_uuid.py`` tool that will test for the existence
332and uniqueness of idempotent_id metadata for every test. By default the
333tool runs against the Tempest package by calling::
334
335 python check_uuid.py
336
337It can be invoked against any test suite by passing a package name::
338
339 python check_uuid.py --package <package_name>
340
341Tests without an ``idempotent_id`` can be automatically fixed by running
342the command with the ``--fix`` flag, which will modify the source package
343by inserting randomly generated uuids for every test that does not have
344one::
345
346 python check_uuid.py --fix
347
348The ``check_uuid.py`` tool is used as part of the tempest gate job
349to ensure that all tests have an ``idempotent_id`` decorator.
350
Matthew Treinisha970d652015-03-11 15:39:24 -0400351Branchless Tempest Considerations
352---------------------------------
353
354Starting with the OpenStack Icehouse release Tempest no longer has any stable
355branches. This is to better ensure API consistency between releases because
356the API behavior should not change between releases. This means that the stable
357branches are also gated by the Tempest master branch, which also means that
358proposed commits to Tempest must work against both the master and all the
359currently supported stable branches of the projects. As such there are a few
360special considerations that have to be accounted for when pushing new changes
361to tempest.
362
3631. New Tests for new features
364^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
365
366When adding tests for new features that were not in previous releases of the
367projects the new test has to be properly skipped with a feature flag. Whether
368this is just as simple as using the @test.requires_ext() decorator to check
369if the required extension (or discoverable optional API) is enabled or adding
370a new config option to the appropriate section. If there isn't a method of
371selecting the new **feature** from the config file then there won't be a
372mechanism to disable the test with older stable releases and the new test won't
373be able to merge.
374
3752. Bug fix on core project needing Tempest changes
376^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
377
378When trying to land a bug fix which changes a tested API you'll have to use the
379following procedure::
380
381 - Propose change to the project, get a +2 on the change even with failing
382 - Propose skip on Tempest which will only be approved after the
383 corresponding change in the project has a +2 on change
384 - Land project change in master and all open stable branches (if required)
385 - Land changed test in Tempest
386
387Otherwise the bug fix won't be able to land in the project.
388
3893. New Tests for existing features
390^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
391
392If a test is being added for a feature that exists in all the current releases
393of the projects then the only concern is that the API behavior is the same
394across all the versions of the project being tested. If the behavior is not
395consistent the test will not be able to merge.
396
397API Stability
398-------------
399
400For new tests being added to Tempest the assumption is that the API being
401tested is considered stable and adheres to the OpenStack API stability
402guidelines. If an API is still considered experimental or in development then
403it should not be tested by Tempest until it is considered stable.