From Fedora Project Wiki

Line 35: Line 35:
*** {{result|pass| Rawhide Acceptance Test Plan [[https://fedoraproject.org/wiki/QA:Rawhide_Acceptance_Test_Plan]]}}
*** {{result|pass| Rawhide Acceptance Test Plan [[https://fedoraproject.org/wiki/QA:Rawhide_Acceptance_Test_Plan]]}}
*** {{result|inprogress| DVD.iso Installation}}
*** {{result|inprogress| DVD.iso Installation}}
*** {{result|inprogress|URL Installation}}
*** Boot.iso/Netinst.iso Installation
*** Boot.iso/Netinst.iso Installation
*** Live.iso Installation
*** Live.iso Installation
*** {{result|inprogress|URL Installation}}
** upgrade an exiting system
** upgrade an exiting system
** system with basic video driver
** system with basic video driver

Revision as of 10:23, 28 April 2011

QA.png


This page provides a high-level roadmap for implementing the Is_anaconda_broken_proposal project. More detailed tasks can be found in autoqa TRAC roadmap. We follow these steps to define the methods by which we initiate testing


First, in order to provide a consistent and documented test approach, the existing Fedora Install test plan [1] will be revisited. The test plan will be adjusted to ensure proper test coverage for the failure scenarios listed above. Existing test cases will be reviewed for accuracy. New test cases will be created using the Template:QA/Test_Case template. Finally, the test plan will be adjusted to match the improved Fedora Release Criteria [2]. This includes adjusting the test case priority to match milestone criteria.

Next, in order to reduce the setup/execution time, improve efficiency and to provide test results on a more consistent basis, a subset of test cases will be chosen for automation. Tests will be written in python and will be developed and executed on a system supporting KVM virtualization. Test scripts will be responsible for preparing a virtual install environment, initiating a kickstart install and validating the results. Once an initial batch of tests exist, they will be formally integrated into the AutoQA project.

Last, a method will be developed for collecting test results into a single test result matrix. Results may be posted to the wiki directly, or a custom turbogears application may be needed to display results [3]. The results will be easily accessible for testers and the installer development team.


The project will be divided into several phases.

Phase#1 - proof of concept

  • Pass pass Revise Fedora Install test plan to ensure adequate test coverage exists for failure scenarios listed above
  • Pass pass Select a small, but representative, subset of test cases from the install test plan to automate
    The following test cases are selected:
    • Install a new system
      • Rawhide Acceptance Test Plan [[1]]
      • DVD.iso Installation
      • Boot.iso/Netinst.iso Installation
      • Live.iso Installation
      • URL Installation
    • upgrade an exiting system
    • system with basic video driver
    • Rescue installed system
    • Memory test [[2]]
  • Pass pass Create python scripts to prepare KVM-based virtual environments for testing, initiate kickstart installs, and validate results
  • Pass pass Investigate methods for leveraging GUI automation to aid in automating applicable test cases

Phase#2 - implementation

  • Implement the selected test cases
    • install a new system
      • Pass pass Rawhide Acceptance Test Plan [[3]]
      • Inprogress inprogress DVD.iso Installation
      • Inprogress inprogress URL Installation
      • Boot.iso/Netinst.iso Installation
      • Live.iso Installation
    • upgrade an exiting system
    • system with basic video driver
    • Rescue installed system
    • Memory test [[4]]
  • Identify and automate remainder of test cases from the install test plan
    • Rawhide Acceptance Test Plan
      • repodata validity
      • comps.xml validity
      • Core package dependency closure
      • Core package existence
      • installer image existence
      • Kernel boot
      • Anaconda loader fetching stage2
      • Anaconda stage2 disk probe
      • Anaconda package install
      • Anaconda bootloader setup
      • X startup/basic display configuration
      • X basic input handing
      • basic network connectivity
      • yum update functionality
    • DVD installation
      • mediakit_ISO size
      • mediakit_ISO checksums
      • mediakit repoclosure
      • mediakit file conflicts
      • boot methods
      • install source
      • additional http repository
      • additional ftp repository
      • additional mirrorlist repository
      • additional nfs repository
    • Boot.iso/netinst.iso installation
      • mediakit ISO size
      • mediakit ISO checksums
      • boot methods
      • install source
      • http repository
    • Live.iso installation
      • media ISO size
      • mediakit ISO checksums
      • install source live image
    • URL installation
      • environment check
      • sanity check
      • installation
  • Implement the general test cases
    • Anaconda user interface graphical [[5]]
    • Anaconda user interface basic video driver[[6]]
    • Anaconda user interface text [[7]]
    • Anaconda user interface VNC [[8]]
    • Anaconda user interface cmdline [[9]]
    • parse different repository: cdrom, http,ftp(anonymout, non anonymus), nfs, nfsiso, hard drive
    • parse different kickstart delivery: http, file, hard drive, nfs
    • different packages selections: default, minimal
    • different partation: autopart, autopart encrypted, autopart shrink install, autopart use free space, ext4 on native device, ext3 on native device, no swap, software raid
    • rescue mode
      • update.img via url installation source local media
      • Anaconda save traceback to remote system/ bugzilla/ disk /debug mode
    • upgrade
      • new boot loader
      • skip boot loader
      • updte boot loader
      • encrypted root
      • skip boot loader text mode
      • update boot loader text mode
    • preupgrade
      • preupgrade
      • preupgrade from older release
  • Identify and integrate anaconda built-in unit tests into tests
  • Identify test event triggers which will be used to automatically initiate testing.
  • Identify test cases where GUI automation will be required

Phase#3 - integration

  • Create appropriate control files and test wrappers to allow for scheduling tests through AutoQA (see Writing_AutoQA_Tests)
  • Develop or update AutoQA test event hooks to accommodate new test events (see Writing_AutoQA_Hooks)
  • Implement initial test result dashboard intended to eventually replace the wiki test matrix. The dashboard will also support FAS user test result submission. This will likely rely on