- From: Markus Krötzsch <mak@aifb.uni-karlsruhe.de>
- Date: Wed, 3 Jun 2009 12:15:02 +0200
- To: Ivan Herman <ivan@w3.org>
- Cc: Mike Smith <msmith@clarkparsia.com>, W3C OWL Working Group <public-owl-wg@w3.org>
- Message-Id: <200906031215.10741.mak@aifb.uni-karlsruhe.de>
Hi Ivan, I just answer directly to your remarks on multiple syntaxes, conversion, and "normativity" in this context. The test harness questions I con only leave to Mike. The reason why the field is called "normative syntax" is that this is a syntactic form that is normative *for the test*, i.e. one that tools can use to check if they pass the test (the test ontology also allows non-normative syntax forms that do not have an official status). The normative versions have been carefully checked before approving a test case in the working group, so there is some quality commitment that could not really be given for automatically generated translations. This is why these syntaxes are specifically marked, even for tests that do not involve syntax conversions. Maybe we should change "normative" to some other term in order to avoid confusion between "normative syntaxes for OWL" and "syntax of a normative ontology used in this test"? If the UMan conversion at some point is so reliable that we completely trust it, then we would still need the "alternative syntax" option, or otherwise the tests would rely on the current availability and correctness of the conversion service (which may change after tests were approved). The alternative syntaxes are also the ones that are included in the exported test metadata. The alternative syntaxes might become less relevant for users who browse the wiki and who may be happy with the online conversion -- this hoped-for situation is what motivated the current layout, where "alternative syntax" is not very prominently displayed. One could of course hide these links even further once the conversion is more reliable. Alternatively, one could re-design the UI to have a better labeling and placing of the "alternative syntax" links -- but if there is hope that the UMan conversion will be fixed soon, I would rather not change all this (in any case, the alternative syntaxes will never be as complete as the service's conversions). -- Markus On Mittwoch, 3. Juni 2009, Ivan Herman wrote: > Hi Mike, > > Thanks! Comments below... > > (Just a side remark: I tried to look at the tests from an OWL RL point > of view. Many remarks might very well be valid for other profiles or for > OWL Full, but I concentrated on this case only for now...) > > Mike Smith wrote: > > I've quoted and responded to those bits for which I have useful feedback. > > > > On Thu, May 28, 2009 at 06:15, Ivan Herman <ivan@w3.org> wrote: > >> - Markus, I did download the RL tests[1]. However, I must admit that, at > >> least for me, this has only a limited usability as is. To test my > >> implementation, I need the individual 'premise' ontologies independently > >> of one another, and all in RDF/XML. The file[1] includes all these as > >> string literals, so I'd have to make an extra script that extracts those > >> string literals, and stores the results in separate RDF/XML files > > > > Alternatively, people can do this by writing a small amount with the > > harness, even if the goal is to run tests with a non-Java tool. I > > added a bit the the Test_Running_Guide page to hint at this. > > O.k. I have not tried to run the tool, and the > > http://github.com/msmithcp/owlwg-test/tree/master > > does not hint at using this tool just to extract the specific tests (I > presume this is on your t.b.d. list) but that is a good way to do it > indeed. > > I (and testers) actually would be interested to know how the harness can > be run. What does it require if I have, say, a web service returning an > expanded RDF graph using the RL rules, to use this harness? > > My comments below are related to the case when the harness cannot be run... > > >> - I picked one test (DisjointClasses-001[3]). It is a bit discomforting > >> that the whole test is described in Functional Syntax that, as I said, I > >> do not understand > >> > >> - However, I find the link at the bottom which says 'Auxiliary syntax > >> documents' which does present the whole test in RDF/XML[4]. This is what > >> I really need! Great. > > > > Each test page shows the format the test was initially created in - > > for most this is RDF, for some it is functional syntax. Some tests > > (mostly those with fs) have multiple normative formats. If an > > auxiliary syntax link is available (as it was in this case), it is > > because the test was manually translated to have multiple normative > > formats. Both formats are included in the "download owl" link and the > > exports, and test test may be used as a syntax translation test. > > I know I am a pain in the back side here, my apologies:-( But, at the > moment, the syntax translators via the M'ter service do not work. When > do we plan to have that up and running? We have already contacted some > of our potential implementers/testers and the deadline we give them to > complete the tests (mid July) is fairly short. Ie, these translations to > other formats should be available very soon... > > A cosmetic issue: the page says 'Normative syntax: Functional'. I am not > sure what this means and I think we should be careful using the > 'normative' word in this case. It of course makes sense for tests that > convert one syntax to the other, but not for others... > > >> I wonder whether that link should not appear in a more prominent place > >> on[3] and not labelled as 'Auxiliary' but simply as 'RDF/XML version'. > >> Alternatively, we could have a complete alternative of [3], with all the > >> additional infos there, but in RDF/XML instead of FS. That could then be > >> linked from[2], ie, we can save the user some extra hops. > > > > That link is not just for RDF/XML. A test could be initially in > > RDF/XML and that link would provide a functional syntax version, or an > > OWL/XML version. > > So if the M'ter conversion service works for all the tests, I am not > really sure what the reason of having those links are. Aren't these just > a source of confusion then? > > >> - This particular test is labelled (on [3]) as 'applicable under both > >> direct and RDF-based semantics'. However, as far as I can see, this test > >> cannot be completed using the OWL RL Rule set. This may be an example > >> where the Direct semantics of RL and the RDF based semantics with the > >> rules diverge or, more exactly, where the Rule set is incomplete. This > >> is fine per se, as long as this is clearly stated on the test page > >> somewhere; otherwise implementers may not understand why they cannot > >> complete this test. > > > > The entailed ontology in this test does not satisfy the requirements > > of Theorem PR1. I believe, then, that the RL + RDF Semantics > > entailment checker could return unknown. > > I would rather say 'non applicable'. Maybe an extra class should be > added to the result ontology indicating this. At the moment I see > 'failing run', 'passing run', or 'incomplete run', and none of these > really describe this case... > > > The test cases indicate > > applicability of Direct Semantics and RDF-Based semantics. They do > > not have an indicator for the partial axiomization of the RDF-Based > > semantics provided by the RL rules. > > > > *** > > I believe this was discussed in the past but no action was taken. > > Would you like to propose enhancing the metadata for RL tests to > > indicate if PR1 is satisfied? > > *** > > I think this is certainly good to have. > > >> - Provided I run the test, eyeball the result, and I am happy with what > >> I see, I presume I have to record it using[6]. First of all, it would be > >> good to add some comments/annotations to that ontology because it is not > >> 100% clear what the various terms mean. Also, the premise was that the > >> implementer does not understand FS, which makes it a bit of a challenge > >> for him/her... > > > > I've modified the page to include a description of the example and > > provided a link to the ontology in RDF/XML. Hopefully that makes it > > more approachable. > > Yes, thank you. But that was not really my point. What I am wondering if > it was possible to add an extra field to, say, the > > http://km.aifb.uni-karlsruhe.de/projects/owltests/index.php/DisjointClasses >-001 > > that provides most of the necessary answer data. Ie, a field saying: > > [[[ > @prefix xsd: <http://www.w3.org/2001/XMLSchema#> . > @prefix : <http://www.w3.org/2007/OWL/testResultOntology> . > [] > a :PositiveEntailmentRun , ADDYOURRESULT ; > > :test [ test:identifier "DisjointClasses-001"^^xsd:string ] ; > :runner ADDIDTOYOURTEST . > > ]]] > > So that the tester can just copy/edit this. The field that is the most > complicated to 'find' for this test is :PositiveEntailementRun; I would > expect a number of responses going wrong... > > Thanks! > > Ivan -- Markus Krötzsch Institut AIFB, Universität Karlsruhe (TH), 76128 Karlsruhe phone +49 (0)721 608 7362 fax +49 (0)721 608 5998 mak@aifb.uni-karlsruhe.de www http://korrekt.org
Received on Wednesday, 3 June 2009 10:15:50 UTC