MetricsBasedTesting

Revision 2 as of 2009-05-27 08:39:15

Clear message

Summary

All our current testing is a binary pass/fail but we would also like to track the evolution of certain parameters such as boot speed and power usage that may not have a clear pass/fail threshold but for which historical data is desireable. We will extend out infrastructure to collect and analyse such data.

Release Note

This section should include a paragraph describing the end-user impact of this change. It is meant to be included in the release notes of the first release in which it is implemented. (Not all of these will actually be included in the release notes, at the release manager's discretion; but writing them is a useful exercise.)

It is mandatory.

Rationale

This should cover the _why_: why is this change being proposed, what justifies it, where we see this justified.

User stories

Assumptions

Design

You can have subsections that better describe specific parts of the issue.

Implementation

This section should describe a plan of action (the "how") to implement the changes discussed. Could include subsections like:

UI Changes

Should cover changes required to the UI, or specific UI that is required to implement this

Code Changes

Code changes should include an overview of what needs to change, and in some cases even the specific details.

Migration

Include:

  • data migration, if any
  • redirects from old URLs to new ones, if any
  • how users will be pointed to the new way of doing things, if necessary.

Test/Demo Plan

It's important that we are able to test new features, and demonstrate them to users. Use this section to describe a short plan that anybody can follow that demonstrates the feature is working. This can then be used during testing, and to show off after release. Please add an entry to http://testcases.qa.ubuntu.com/Coverage/NewFeatures for tracking test coverage.

This need not be added or completed until the specification is nearing beta.

Unresolved issues

This should highlight any issues that should be addressed in further specifications, and not problems with the specification itself; since any specification with problems cannot be approved.

BoF agenda and discussion

Metrics: tracking test results over time for trending analysis

Examples

  • tracking bootspeed (using bootchart)
  • power usage
  • I/O throughput
  • amount of memory being used
    • polled during install, application testing, immediately after boot, etc.
  • Webcam framerate

Questions

  • How do we get the metrics from the test systems to the people who care about them?
    • currently this data is just stored on the filesystem of the certification server
    • we would like to be able to parse out this data and present it in a usable form to people who are interested
    • however, we don't want to be the bottleneck for this data and we need to make sure users have access to the raw data
  • Should this data be public or private?
    • Will proprietary data be exposed if the data is public? (E.g. unreleased hardware)
    • We should be able to make a split the same way we do with the existing released data
      • this is done with a flag in the certification database marking systems as private
  • Should benchmarks be pigeonholed into the pass/fail/skip system?
    • should numeric data be an addition to the existing result type or a new one?
      • heno feels that this should be a new type -- our existing results ain't broke so don't fix 'em
      • also certain data doesn't really necessarily have a pass or fail... e.g. boot speed
  • Do we really need to see the analysis of these benchmarks online in our certification or results-tracking system?
    • this data might be better placed in a more specific offline analysis

Short-term Goals

  • Gather metric data as attachments (files/blobs)
  • Provide access to gathered metric data
  • Do not parse it in certification system or results tracker -- just gather and regurgitate

Mid-term Goals

  • Store metric data in database as numeric value
  • Do not yet parse or analyze heuristically
  • Provide simple access for ad-hoc reporting

Long-term Goals

  • Be able to track metrics heuristically
  • Generate alerts when metrics change by a sufficiently large amount
  • Automated quantitative analysis
  • Parse data from metrics (required for above goals)

Proposal

Initial pass should be able to be implented quickly and provide value to groups who need this data. Basically we want to provide access to the data without really doing any parsing ourselves.

  • Have a directory containing the data
  • .../bootchart/$machinename/$timestamp.tgz or .../$machinename/metrics/bootchart/$timestamp.tgz
    • linked from the certification 'librarian' so there is no necessity for copying the files around
  • possibly have a file containing metadata about the machine from this view
    • this data is already available from the website and can be retrieved programmatically
    • information such as processor speed, ram, etc.

Tracking Data Online Proposal

Track online only one number for each result that reports metric data. Continue to store any other information as an attachment, so if detailed analysis is desired we can provide access to the data.

The justification for this is that otherwise the server needs to know too much about each type of result... this just becomes untenable.


CategorySpec