- From: Mark Skall <mark.skall@nist.gov>
- Date: Mon, 05 Jan 2004 14:33:52 -0500
- To: www-qa-wg@w3.org
- Message-Id: <5.1.0.14.2.20040105142837.03c8fe70@mailserver.nist.gov>
QA Working Group Teleconference Monday, 05-January-2004 -- Scribe: Mark Skall Attendees: (DH) Dominique Hazaël-Massieux (W3C) (LH) Lofton Henderson (CGMO - WG co-chair) (LR) Lynne Rosenthal (NIST - IG co-chair) (MS) Mark Skall (NIST) Regrets: (KD) Karl Dubost (W3C, WG co-chair) (PC) Patrick Curran (Sun Microsystems) (SM) Sandra Martinez (NIST) (VV) Vanitha Venkatraman (Sun Microsystems) (dd) Dimitris Dimitriadis (Ontologicon) Absent: (MC) Martin Chamberlain (Microsoft) (AT) Andrew Thackrah (Open Group) Guests: (DM) David Marston (SH) Sandro Hawke Summary of New Action Items: No new action items Agenda: http://lists.w3.org/Archives/Public/www-qa-wg/2003Dec/0067.html Previous Telcon Minutes: http://lists.w3.org/Archives/Public/www-qa-wg/2003Dec/0072.html Minutes: 1.) roll call 11am EDT 2.) Any routine business? LH Next week will have Monday (Test Assertions) and Wednesday (TestGL draft) telcons MS Will we be addressing Jeremy Carroll’s comments? LH We will need to address this probably at the end of the month. I will send mail suggesting a particular day for this topic. 3.) Presentation by Sandro Hawke of the work done by the OWL WG and the RDF WG for the development of their test materials and the gathering of the tests results. A few references: - OWL Test repository: http://www.w3.org/2002/03owlt/ - OWL Test Cases document: http://www.w3.org/TR/owl-test/ - OWL Test Results: http://www.w3.org/2003/08/owl-systems/test-results-out - RDF Test Cases: http://www.w3.org/TR/rdf-testcases/ - RDF Test Results: http://www.w3.org/2003/11/results/rdf-core-tests The following are Sandro’s notes he e-mailed prior to the telcon: 1 Introduction This is Mostly a Story about a Web Page Prolog: A Test-Driven Implementation of OWL ("Surnia") Effects the page has had What's next? About Me [ At home today (a bit of a cold, you may hear kids) ] - W3C Team, Semantic Web Activity, DARPA/Research Funding - Joined WebOnt (OWL) in June 2 Surnia - Immediate motivate: demonstrating implementability of OWL Full - I Didn't Read The Spec! I just implemented based on my rough idea, then modified to pass more and more tests. - Results were decent - Also for RDF Core entailment & non-entailment tests 3 The Test Results Page 3.1 Test Results Ontology (for reporting results in RDF) - TestRun (PassingRun, FailingRun, IncompleteRun / UndecidedRun) + which test + where is the output (details of the run) + which system was tested (NOT tracking Project/Release/Platform/etc) + time test began + test duration 3.2 Test Results Page -- First Version SEE: http://www.w3.org/2003/08/owl-systems/res20030904 - big table of tests/systems - regen at bottom - self-explanitory 3.3 Feeds - give me the URL of your test data - results page let people try it themselves, to do some debugging - some people advertised their results telling people to do a custom version of the page! - various bugs in their systems and a few in mine 3.4 Test Results Refinements NOW: http://www.w3.org/2003/08/owl-systems/test-results-out - group tests - leave out systems with no results for a group - summary table at top - group by number-passed/failed 4 Effects? 4.1 Working Group - helped us decide which test to look at more closely, approve, move to extra-credit - let us push implementors on key tests 4.2 Implementors - could see who was doing what, get some publicity (not enough system data per test, though, IMHO -- output link) 4.3 W3C Members and Director - for advancement to PR -- Director liked it; I hope the members do 5 What's next? 5.1 Test Suite stable now, as we're at PR - non-normative tests from WG? - tests from other submitters? 5.2 Page usability - smaller bits! searchable! - better info/links about implementations 5.3 More info on page - benchmarking - change-over-time - nuances of test results (use of "output" link?) - facilitating *discussion* of tests Sandro’s notes end here. The following link is Sandro’s discretion of ontology test results: www.w3.org/2002/03owlt/resultsOntology SH I Work for W3C with DARPA funding and am a member of the web ontology Working Group. First thing I wanted to do was to implement the system. First implemented OWL Full without reading the spec and ran my implementation against the test suite (the test suite is normative). Then modified the code until a large number of the tests were passed. Also ran against RDF core tests and modified code to pass them as well. Other implementations were run against the tests. Put tests in RDF and created test results ontology for reporting results in RDF. For each test run, there is a url for each test, output and which system is being tested (no detail on versions). Also, asked for the time it takes to run a test. However, WG has not defined performance requirements. Turned output into html. Results of Surnia and old surnia were displayed. LH What does “incomplete” mean? SH It means that some form of the tests didn’t finish same as undecided. MS What does “undecided” mean with respect to conformance? SH It depends on the test. MS Is it documented what “undecided” means especially since tests are normative? SH For interoperability, we considered them as “fails.” MS What is “extra credit”? SH These tests were not expected to pass. MS Are these requirements? Are they “MUSTS”? SH What should pass depends on the type of system being built. LH Extra credit tests seem to be normative. What does extra credit tests mean with respect to conformance? SH We’re interested in conformance of documents, not systems. LH There are no conformance requirements on implementations? SH There are on some. SH Section 4 Effects. Looked at tests not being passed by a lot of systems. Is there something wrong with these tests? If important tests did not have enough passes, WG sent e-mail to try to get these tests passed. For the move to Proposed Rec (where you need to demonstrate interoperability), we used this to come up with statistics. LR Did you need 2 implementations that did everything? SH They had to be 2 out of a group need a certain fraction that passed tests for OWL and OWL Light. LH Test cases do not constitute a conforming test suite for OWL and interoperability is in terms of test cases. Aren’t there holes (no test cases for normative content in OWL)? SH You can’t test every possible combination of protocol messages. MS But this isn’t a combination. However, shouldn’t every requirement be tested at least once? SH These requirements missing were combinations. Everything is tested at a simple level. Froze test suite when OWL went to PR. LH Can you give a reference to a description of test results ontology? SH Will put it in IRC. DH - How does ontology relate to Earl? SH There’s a lot of overlap. Did it independently could have used EARL but it would have been awkward. LH In terms of how QAWG uses the term “normative”, in what sense is a test case itself “normative”? Normative prescribes required behavior. SH “Undecided” comes in to play. “Normativity” demonstrates what‘s entailed or not. MS Does “normativity” add additional requirements to what’s in the spec? SM In theory, just illustrates the requirements, doesn’t add to it. MS So what does it add to call tests “normative”? SH Failing a normative test means you don’t conform. MS So aren’t all test cases normative? SH Most WGs don’t put their test cases on the Rec track. DH You could have an informative test case on the Rec track. LH What would happen if there was a contradiction between 2 normative things (test cases vs. Rec)? SH Text at top says that test cases are subsidiary to Rec. However, conflict would be a cause for concern. LH Should test cases be written to define rules or after rules are written? SH Test cases are easier to understand (and figure out if they’re right) than semantic documents. DH Spec has a lot more semantics than test itself (e.g., conformance statement and prose). Test cases should be conformance example but spec should be first priority. Test cases are harder to understand. SH In these cases, test cases are easier to understand DH Because, in this case, you understand the technology. Adjourn at 1205. Note: The QA Working Group would like to express our thanks to Sandro for the informative and thought provoking presentation. **************************************************************** Mark Skall Chief, Software Diagnostics and Conformance Testing Division Information Technology Laboratory National Institute of Standards and Technology (NIST) 100 Bureau Drive, Stop 8970 Gaithersburg, MD 20899-8970 Voice: 301-975-3262 Fax: 301-590-9174 Email: skall@nist.gov ****************************************************************
Received on Monday, 5 January 2004 14:39:44 UTC