Reasons for requirements¶
Preconditions¶
Development release fixed first¶
If a stable release is fixed but the development release is not, then after the development release becomes a stable release, users upgrading would face an unexpected regression. Therefore, it is a general requirement that the development release is fixed before fixes are backported to the stable releases. Equivalently for new upstream releases, this (or a newer) release must be in the development release.
It is also, in general, not appropriate to release updates for stable systems without first testing them in the current development branch.
One exception to this general rule is the case where the development release is not yet open. There can sometimes be a delay between the release of the most recent version of Ubuntu and the opening for development of the next version. Provided they are important enough, stable release updates should not and do not need to wait for the development release to open, as long as the development release upload is prepared and ready.
Newer Releases¶
If a bug is being fixed in a particular stable release, we would like for all subsequent releases that are still supported to also be fixed at the same time. This is to prevent a user from facing a regression when they upgrade to a newer release.
Exceptions:
The following two exceptions apply to bugfixes, but do not apply to hardware enablement or new features.
When there are two subsequent interim releases: if there are two subsequent interim releases that are both current, then, as a compromise, additionally fixing only the most recent one is acceptable. Rationale: a user facing this class of regression will at least have an upgrade path available to them that fixes it.
When you don’t want to fix a subsequent interim release at all: we recognise that making it a hard requirement to fix all subsequent interim releases would mandate more work, and that a team may not have the resources available to fix and verify (say) an LTS as well as a subsequent interim release that has fewer users. We wouldn’t want to block a fix from landing at all, so we are not making it a hard requirement that subsequent interim releases be fixed. However, we strongly recommend that subsequent interim releases be fixed, and it is our expectation that normally uploaders will ensure this. If you are unable to do this, then please: 1) create and mark bug tasks against the subsequent affected releases “Won’t Fix”; and 2) explicitly state in the bug that you are deliberately seeking to fix a release without fixing the subsequent releases. An SRU team member may then accept your upload at their discretion and on a case-by-case basis. If this is not done, then uploaders should expect an SRU review round trip while your intentions are clarified.
See also: Reference → Requirements → General requirements for all SRUs
Documentation¶
[Insert specifics here: SRU template, what is expected in each section, etc]
Documentation must be provided in order to meet our public documentation requirement as well to assist the SRU team to review your upload. This is usually done individually in the description area of the bug, for each bug being fixed by the SRU, and should follow the SRU bug template.
Explicit is better than implicit: if there’s anything a reviewer might find unexpected, calling it out will help us tell the difference between an inadvertent error or omission and a deliberate choice. The former is likely to result in a further review iteration. The latter gives us confidence that you, as the subject matter expert, have considered the matter and we are much more likely to accept your suggestion on how to deal with it. In any case we will save at least one review iteration in determining whether the matter is real or has been missed.
Detail in Test Plans¶
Consider if another person were to follow the Test Plan as written. Are they likely to perform the same actual steps in practice? If not, then the Test Plan is not sufficiently detailed and unambiguous. Otherwise we might end up performing insufficient testing because the Test Plan that was approved is not the Test Plan that was actually executed. Ambiguity would result in unnecessary regression risk, violating our principles.
If a regression is found, then we will need to consider what went wrong to see if there are any opportunities for improvement next time. To be able to analyse this effectively, our QA must be reproducible. Otherwise, if original tester is unavailable, developers could waste time due to the ambiguity.
It follows that the Test Plan must be unambiguously reproducible.
There are multiple roles interacting here: the developer who writes the Test Plan, the SRU team member who approves it, the person who performs it, and the SRU team member who reads the tester’s report before releasing the update. It is essential that everyone involved agrees on the actual steps to perform, without any misunderstanding.
How do we avoid needless detail? Here’s a suggestion: if another person were to follow the Test Plan and it’s likely that they will interpret it to use the same steps as you intend, then further detail is not needed.
Example¶
“Check that the service still works” is an ambiguous instruction. Should
the tester run systemctl status
and verify that systemd reports the
service as active and running, or actually verify that the service works
by running a query against the service? If a query, then what query
exactly?
If, later, a regression is found, then we will want to know what the tester actually did. Perhaps the regression occurred because they didn’t check that the service actually works by running a query against the service, even though the person who wrote the Test Plan intended it:
The developer would say: “Yes obviously you needed to check the service actually works; that’s what I meant when I wrote the Test Plan.
The SRU reviewer would say: “On review, given the nature of the bug being fixed and the changes being made, I thought it was important to check the service responds correctly to a query, but that’s what the Test Plan included so I approved it.
The tester would say: “I carried out the testing exactly as instructed” and then report in the bug for SRU verification “I have carried out the Test Plan specified against version X and it passed”.
…but this would then have regressed users solely because of the ambiguity.