MetricsBasedTesting

Differences between revisions 1 and 3 (spanning 2 versions)
Revision 1 as of 2009-05-17 04:36:56
Size: 2671
Editor: 208-151-246-43
Comment:
Revision 3 as of 2009-06-10 11:40:56
Size: 2859
Editor: cpc4-oxfd8-0-0-cust39
Comment:
Deletions are marked like this. Additions are marked like this.
Line 4: Line 4:
 * '''Created''':  * '''Created''': UDS Karmic
Line 6: Line 6:
 * '''Packages affected''':
Line 10: Line 9:
All our current testing is a binary pass/fail but we would also like to track the evolution of certain parameters such as boot speed and power usage that may not have a clear pass/fail threshold but for which historical data is desireable. We will extend out infrastructure to collect and analyse such data. All our current testing is a binary pass/fail but we would also like to track the evolution of certain parameters such as boot speed and power usage that may not have a clear pass/fail threshold but for which historical data is desireable. We will extend out infrastructure to collect and display such data.
Line 12: Line 11:
== Release Note ==

This section should include a paragraph describing the end-user impact of this change. It is meant to be included in the release notes of the first release in which it is implemented. (Not all of these will actually be included in the release notes, at the release manager's discretion; but writing them is a useful exercise.)

It is mandatory.
Line 20: Line 14:
This should cover the _why_: why is this change being proposed, what justifies it, where we see this justified. Work to improve performance requires access to rich performance data rather than binary pass/fail.
Line 24: Line 18:
== Assumptions ==
Line 28: Line 21:
You can have subsections that better describe specific parts of the issue. Gather non-binary data during various checkbox tests and feed them on to the developers who need this for their work.

==== Examples ====
 * Tracking bootspeed (using bootchart)
 * Power usage
 * I/O throughput
 * Amount of memory being used
  * polled during install, application testing, immediately after boot, etc.
 * Webcam framerate
Line 32: Line 33:
This section should describe a plan of action (the "how") to implement the changes discussed. Could include subsections like: Initial pass should be able to be implented quickly and provide value to groups who need this data.
Line 34: Line 35:
=== UI Changes === === Phase I (Karmic) ===
Line 36: Line 37:
Should cover changes required to the UI, or specific UI that is required to implement this  * Gather metric data as attachments (files/blobs)
 * Provide access to gathered metric data for developers
 * Do not parse it in the certification system or results tracker -- just gather and regurgitate
Line 38: Line 41:
=== Code Changes === Provide access to the data without doing any parsing
 * Have a directory containing the data
 * .../bootchart/$machinename/$timestamp.tgz or .../$machinename/metrics/bootchart/$timestamp.tgz
  * linked from the certification 'librarian' so there is no necessity for copying the files around
 * Have a file containing metadata about the machine from this view
  * this data is already available from the website and can be retrieved programmatically
  * information such as processor speed, ram, etc.
Line 40: Line 49:
Code changes should include an overview of what needs to change, and in some cases even the specific details. === Phase II (Later) ===
Line 42: Line 51:
=== Migration ===  * Store metric data in database as numeric value
  * Do not yet parse or analyze heuristically
 * Provide simple access for ad-hoc reporting
 * Generate alerts when metrics change by a sufficiently large amount or above a threshold
Line 44: Line 56:
Include:
 * data migration, if any
 * redirects from old URLs to new ones, if any
 * how users will be pointed to the new way of doing things, if necessary.

== Test/Demo Plan ==

It's important that we are able to test new features, and demonstrate them to users. Use this section to describe a short plan that anybody can follow that demonstrates the feature is working. This can then be used during testing, and to show off after release. Please add an entry to http://testcases.qa.ubuntu.com/Coverage/NewFeatures for tracking test coverage.

This need not be added or completed until the specification is nearing beta.
Line 57: Line 59:
This should highlight any issues that should be addressed in further specifications, and not problems with the specification itself; since any specification with problems cannot be approved.  * Should this data be public or private?
  * Will proprietary data be exposed if the data is public? (E.g. unreleased hardware)
  * We should be able to make a split the same way we do with the existing released data
   * this is done with a flag in the certification database marking systems as private
 * Should benchmarks be pigeonholed into the pass/fail/skip system?
  * should numeric data be an addition to the existing result type or a new one?
   * heno feels that this should be a new type -- our existing results ain't broke so don't fix 'em
   * also certain data doesn't really necessarily have a pass or fail... e.g. boot speed
Line 59: Line 68:
== BoF agenda and discussion ==

Use this section to take notes during the BoF; if you keep it in the approved spec, use it for summarising what was discussed and note any options that were rejected.

Summary

All our current testing is a binary pass/fail but we would also like to track the evolution of certain parameters such as boot speed and power usage that may not have a clear pass/fail threshold but for which historical data is desireable. We will extend out infrastructure to collect and display such data.

Rationale

Work to improve performance requires access to rich performance data rather than binary pass/fail.

User stories

Design

Gather non-binary data during various checkbox tests and feed them on to the developers who need this for their work.

Examples

  • Tracking bootspeed (using bootchart)
  • Power usage
  • I/O throughput
  • Amount of memory being used
    • polled during install, application testing, immediately after boot, etc.
  • Webcam framerate

Implementation

Initial pass should be able to be implented quickly and provide value to groups who need this data.

Phase I (Karmic)

  • Gather metric data as attachments (files/blobs)
  • Provide access to gathered metric data for developers
  • Do not parse it in the certification system or results tracker -- just gather and regurgitate

Provide access to the data without doing any parsing

  • Have a directory containing the data
  • .../bootchart/$machinename/$timestamp.tgz or .../$machinename/metrics/bootchart/$timestamp.tgz
    • linked from the certification 'librarian' so there is no necessity for copying the files around
  • Have a file containing metadata about the machine from this view
    • this data is already available from the website and can be retrieved programmatically
    • information such as processor speed, ram, etc.

Phase II (Later)

  • Store metric data in database as numeric value
    • Do not yet parse or analyze heuristically
  • Provide simple access for ad-hoc reporting
  • Generate alerts when metrics change by a sufficiently large amount or above a threshold

Unresolved issues

  • Should this data be public or private?
    • Will proprietary data be exposed if the data is public? (E.g. unreleased hardware)
    • We should be able to make a split the same way we do with the existing released data
      • this is done with a flag in the certification database marking systems as private
  • Should benchmarks be pigeonholed into the pass/fail/skip system?
    • should numeric data be an addition to the existing result type or a new one?
      • heno feels that this should be a new type -- our existing results ain't broke so don't fix 'em
      • also certain data doesn't really necessarily have a pass or fail... e.g. boot speed


CategorySpec

QATeam/Specs/MetricsBasedTesting (last edited 2009-08-21 15:26:54 by ua-178)