From Fedora Project Wiki

Template documentation [edit]

This documentation is transcluded from Template:Release validation instructions/doc. It will not be transcluded on pages that use this template.
Warning.png
Wikitcms core page
This page is a core component of the Wikitcms system for storing release validation testing results in the Fedora wiki. This system relies on specific expectations regarding page naming, contents, and categorization. Please take special care when editing this page, especially if you change any element of wiki syntax, especially templating elements. If you are at all unsure about the consequences of your planned change, please contact the QA group before applying it.
This template is transcluded by the main template for release validation test pages, which is substituted into the test pages for each compose at the time of creation. For example, at the time Fedora 21 Beta TC1 was released, it was substituted to create Test_Results:Fedora_21_Beta_TC1_Base. The mess of #ifeq's at the top displays the 'no longer current' admon box if appropriate - it has to cover the possibility that the 'current' compose may be either a nightly or a TC/RC, so it has to use a couple of nested #ifeq statements.
Important.png
No longer current
The compose for which this page contains results is no longer the current one. This page contains the results for the current compose.

This page records $testtype validation testing test results for the Fedora $release $milestone $compose compose.

Which tests to run

Note.png
Test coverage page
This page provides information about test coverage for the tests on this page across all the composes for the current release: it can help you see which test cases most need to be run.

Tests with a Milestone of Basic, Beta or Final are the most important. Optional tests are less so, but still useful to run. The milestone indicates that for that milestone release or a later one to be approved for release, the test must have been run against the release candidate build (so at Beta, all Basic and Beta tests must have been run, for instance). However, it is important to run the tests for all milestones as early and often as possible. Please refer to the test coverage page linked above and try to find and run tests which have not been run at all, or not run recently, for the current release.

How to test

1. Download one or more media for testing: Template:Fedora Download


2. Perform one or more of the test cases and add your results to the table below

  • You can submit results by editing the page directly, or by using relval, with the relval report-results command. It provides a simple text interface for reporting test results.

3. If a test fails, file a bug report. You may propose the bug as a release blocker or freeze exception bug for the appropriate release - see blocker bug process and freeze exception bug process.

Some tests must be run against particular Products or images - for example, the #Default boot and install tests. If no particular product or image is specified either in this page or the test case page, you can use any appropriate image. For example, you can run most of the #General Tests with the Workstation live image, or either of the Server install images.

If you notice a problem during a test which does not constitute a complete failure of the test, you should still file a bug report, but it may not be appropriate to propose it as a release blocker bug. Use your judgment in deciding this, with reference to the Fedora_Release_Criteria, which these tests are intended to verify. If you are unsure, err on the side of proposing the bug as a blocker.

Important.png
Don't install updates
Don't install updates before performing any of the tests, as when you are testing pre-releases, available updates are not part of the proposed released package set.
Note.png
Virtual machine testing
In most cases, testing in a virtual machine is OK. You can test UEFI (including SecureBoot) in VMs as well.

Results summary page

The Test Results:Fedora $release $milestone Summary page contains the results from this page and all the other validation pages for the same compose listed together to provide an overview.

Add, Modify or Remove a Test Case

  1. Please request review for your changes by publishing your test case for review to the test mailing list and/or the appropriate working group mailing list (e.g. server, cloud, or desktop).
  2. Once reviewed, make your changes to any current documents that use the template (e.g. Test_Results:Current_$testtype_Test).
  3. Lastly, update Template:$testtype_test_matrix with the same changes.

Key

See the table below for a sample format for test results. All test results are posted using the result template.


Test Result Explanation Code Entered
none
Untested - This test has not been run, and is available for anyone to contribute feedback. {{result|none}}
Pass pass robatino
Passed - The test has been run and the tester determine the test met the expected results {{result|pass|robatino}}
Inprogress inprogress adamwill
Inprogress - An inprogress result is often used for tests that take a long time to execute. Inprogress results should be temporary and change to pass, fail or warn. {{result|inprogress|adamwill}}
Fail fail jlaska [1] [2]
Failed - Indicates a failed test. A link to a bug must be provided. See Template:Result for details on providing bug information.
  1. RHBZ #XYZ
  2. RHBZ #ZXY
{{result|fail|jlaska|XYZ|ZXY}}
Warning warn rhe
[1]
Warning - This test completed and met the expected results of the test, but other issues were encountered during testing that warrant attention.
  1. Brief description about the warning status
{{result|warn|rhe}} <ref>Brief description about the warning status</ref>
Pass pass hongqing
Warning warn kparal
Multiple results - More people can easily provide results to a single test case. {{result|pass|hongqing}} {{result|warn|kparal}}
Fail fail pboy [1] [2]
Failed - Same issue with LVM again
  1. RHBZ #2246871
  2. RHBZ #'2244305
{{result|fail|pboy|2246871|2244305}}
Pass pass previous <build> run
Result from previous test run - This test result is directly moved from the test run of previous <build>. {{result|pass|previous <build> run}}
Unsupported - An unsupported test or configuration. No testing is required.