- From: fantasai <fantasai.lists@inkedblade.net>
- Date: Mon, 21 Apr 2008 11:06:29 -0700
- To: "public-css-testsuite@w3.org" <public-css-testsuite@w3.org>, Dominique Hazael-Massieux <dom@w3.org>
One way to get more people involved in CSS conformance testing and to encourage implementors to send in implementation reports would be creating a system that makes it easy for anyone to submit pass/fail data. Lucky for us, the Mobile Web Initiative has already created such a system. :) There are some improvements I'd like to see before we start using it to generate real implementation reports, however, the first major one being to get the pass/fail buttons out of the test file so they don't interfere with the test. HP has volunteered to make improvements to the test harness, and they asked me for a wishlist. Here's what I sent: -------------------------------------------------------------- I promised you a test suite harness wishlist you could discuss. But first, here are the links I sent over the telecon: CSS Test Suite wiki: http://csswg.inkedblade.net/test/css2.1 Microsoft's tests: http://samples.msdn.microsoft.com/csstestpages/ Mobile Web Initiative harness: http://www.w3.org/2007/03/mth/harness Mobile Web Initiative harness source code: http://dev.w3.org/cvsweb/2007/mobile-test-harness/ One of the very neat things about the MWI test harness is that it associates the results with a user agent string. This means there's no need for users to log in or to select their UA. They just load a test and click Pass/Fail/Can't Tell. Very briefly, the improvements I'd like to see are (numbered approximately by priority): New Harnesses: 1. Harness using <iframe> to contain the test and pass/fail buttons on the containing page rather than inside the test file. (This format is good for desktop browsers.) 2. Harness using links targetted at a new window to open the test and pass/fail buttons on the page containing the link rather than inside the test file. (This format is necessary for print.) In both cases the test itself should be referenced as a link, not fed through the CGI script. (This avoids tampering with the HTTP headers that normally get served up with the tests.) 8. It would be nice if these harnesses could include some meta information about the test in addition to the buttons. E.g. the test ID (filename before extension), test title, any requirements documented in the test ("Warning: Must install Ahem font." etc). (I can extract this information for you into a flat-file database during the test suite build process so the harness doesn't have to do any analysis of the tests themselves.) Better Reporting: 3. Ability to consolidate results for various user agent strings under one category name. E.g. consolidate results for all UA strings that represent Opera 9.25 Beta 1 regardless of OS and localization. 4. Pass/fail scores for the whole test suite 5. Ability to report consolidated pass/fail scores for a named groups of tests. (To create e.g. a summary of what features are supported and to what level.) 6. Interface for generating reports based on various parameters. URLs to these reports should be short and clean so they can be passed around in blogs/IM/email etc 9. Prettier reports. :) Data cleanup: 7. Ability to delete all data for a given test (so that when a test is changed we can invalidate the results for that test). This process could be triggered manually on the command line rather than via CGI-- that avoids the need for a login system. ~fantasai
Received on Monday, 21 April 2008 18:07:11 UTC