- From: Ivan Herman <ivan@w3.org>
- Date: Fri, 12 Aug 2016 18:09:59 +0200
- To: W3C Public Annotation List <public-annotation@w3.org>
- Message-Id: <B8DFC42C-78C6-491F-AE2F-99A9F10037D7@w3.org>
Minutes are here, text version below https://www.w3.org/2016/08/12-annotation-minutes.html ---- Ivan Herman, W3C Digital Publishing Lead Home: http://www.w3.org/People/Ivan/ mobile: +31-641044153 ORCID ID: http://orcid.org/0000-0003-0782-2704 [1]W3C [1] http://www.w3.org/ Web Annotation Working Group Teleconference 12 Aug 2016 [2]Agenda [2] https://lists.w3.org/Archives/Public/public-annotation/2016Aug/0117.html See also: [3]IRC log [3] http://www.w3.org/2016/08/12-annotation-irc Attendees Present Rob Sanderson (azaroth), Tim Cole, Nick Stenning (nickstenn), Jacob Jett, Ivan Herman, Dan Whaley, Ben De Meester, Benjamin Young (bigbluehat), Shane McCarron Regrets Chair Rob Sanderson, Tim Cole Scribe Tim_Cole, nickstenn, azaroth Contents * [4]Topics 1. [5]Agenda review 2. [6]Minutes Approval 3. [7]Announcements? 4. [8]Issues 5. [9]Model Testing 6. [10]Protocol Testing * [11]Summary of Action Items * [12]Summary of Resolutions __________________________________________________________ Agenda review <TimCole> azaroth: minutes, announcements, internationalization issues (brief), testing <TimCole> ... any other topics for today? Minutes Approval <azaroth> PROPOSED RESOLUTION: Minutes of the previous call are approved: [13]https://www.w3.org/2016/08/05-annotation-minutes.html [13] https://www.w3.org/2016/08/05-annotation-minutes.html <ivan> +1 <azaroth> +1 <TimCole> +1 <Jacob> +1 RESOLUTION: Minutes of the previous call are approved: [14]https://www.w3.org/2016/08/05-annotation-minutes.html [14] https://www.w3.org/2016/08/05-annotation-minutes.html Announcements? <TimCole> None Issues azaroth: The discussion around I18N has continued. We now have six (6) open issues around this topic. ... To summarise: ... #335: WONTFIX -- long thread with social web WG ... last state was I18N folks were discussing with Activity Streams folks and would get back to us ... #342: A suggestion from Sergiu to add a note to say if dc:language is specified and processingLanguage is not, the latter should be assumed to be the former ... seems like a good editorial note ... Similarly with #343. ... which is about whether processingLanguage should be required to be a language in dc:language ... Haven't had much of a chance to look at #345, also about processingLanguage ... #341 also about processingLanguage for multilingual resources, and we've decided to postpone <ivan> Issue #345 is an attempt from Richard to close an issue and discussion with gsergiu via an editorial change proposal azaroth: suggest we spend the time on the call discussing testing rather than going into the details on some of these I18N issues ... any problems? Model Testing TimCole: regarding the creation of the underlying schemas for the tests -- we've captured everything 3.1, 3.2, and 3.3 excepting agents, and most of section 4 ... these schemas are in the "definitions" folder ... and are referenced by the schemas we intend to use for assertions <TimCole> [15]http://testdev.spec-ops.io:8000/tools/runner/index.html [15] http://testdev.spec-ops.io:8000/tools/runner/index.html TimCole: I've just started working on the test scripts. Have been hashing out with ShaneM what those look like. ... This [^] is a test environment ShaneM has set up. ... You can use this with "Run tests under path" using the following path: <ShaneM> [16]http://testdev.spec-ops.io:8000/tools/runner/index.html?pat h=/annotation-model should work too [16] http://testdev.spec-ops.io:8000/tools/runner/index.html?path=/annotation-model TimCole: "/annotation-model" ... you can paste JSON[-LD] in and run through through the test suite ... that's all working. ... Having some small issues for SHOULD requirements, where we're not necessarily expecting you to pass the test. Currently if you *do* pass the requirement, the test fails. <ShaneM> try going to this URI now: <ShaneM> [17]http://testdev.spec-ops.io:8000/annotation-model/annotation s/3.1-model-musts-v3-manual.html [17] http://testdev.spec-ops.io:8000/annotation-model/annotations/3.1-model-musts-v3-manual.html <azaroth> Example annotations to play with: [18]https://github.com/w3c/web-annotation/tree/gh-pages/model/w d2/examples/correct [18] https://github.com/w3c/web-annotation/tree/gh-pages/model/wd2/examples/correct TimCole: ShaneM: [walking us through how to use the test tool at the link above to test annotations] <ShaneM> Errors: data should have required property '@context'; expected true got false ShaneM: Question: do we want to suppress the output above ^ <Zakim> azaroth, you wanted to ask about display:none in HTML ? azaroth: Would it be possible to have the AJV stack trace be in a display:none; area with a button to reveal it or similar? ShaneM: I don't think so. We just report data back to the test harness, which is responsible for the display. azaroth: the "data should have required property @context" is particularly useful to understand what's going on ShaneM: I'll leave it there, then. ivan: currently can't rerun with updated JSON ShaneM: I'll see if we can fix that <TimCole> [19]http://testdev.spec-ops.io:8000/annotation-model/bodiesTarg ets/3.2-model-manual.html [19] http://testdev.spec-ops.io:8000/annotation-model/bodiesTargets/3.2-model-manual.html TimCole: if you put in a test annotation that *passes* the SHOULD requirements, the output is a little harder to interpret ... not sure quite how to improve the output in that case <azaroth> I used: [20]https://github.com/w3c/web-annotation/blob/gh-pages/model/w d2/examples/correct/anno41.json [20] https://github.com/w3c/web-annotation/blob/gh-pages/model/wd2/examples/correct/anno41.json ivan: [proceeds to find a bug in one of the testing schemas while on the call] TimCole: [paraphrasing heavily] currently for the SHOULD assertions, we expect non-conformance, which means that if it actually is conformant, the test fails <azaroth> Pass if it's not used, with a success message that it SHOULD be there? ShaneM: no way of doing "warning" in the framework ... but we might be able to use "testType" to distinguish between MUST and SHOULD assertion types <azaroth> Result: Pass Message: WARNING: Format SHOULD be included for bodies, if known <ShaneM> use assertionType of must, may, or shold TimCole: we have other scenarios where we say "SHOULD have 1, MAY have more than 1 X" <ShaneM> testType has to do with automation. TimCole: there are other cases where you MUST NOT have more than 1 ... so in these cases we can have multiple assertions for the different cardinalities <ShaneM> [21]http://testdev.spec-ops.io:8000/annotation-model/bodiesTarg ets/3.2-model-manual.html [21] http://testdev.spec-ops.io:8000/annotation-model/bodiesTargets/3.2-model-manual.html ShaneM: If you look at the top of [^] you'll notice that the page fills in as the page loads. ... In the description at the top of the page, there's a list of things the test will check ... do we want to include SHOULD/MUST/MAY information in these descriptions? [noises of general agreement] TimCole: going to spend the next little while cleaning this up with a view to sharing it more widely ... how are the test outputs recorded? ShaneM: anyone with an implementation can record their JSON test output and add it to a git repository which contains all the results TimCole: how do people want to break up the various test for bodies/targets/optional keys/etc.? ShaneM: speaking as "not an implementer" -- the smallest number of manual tests that get us the information we need is probably a reasonable guideline TimCole: probably a discussion for the mailing list <azaroth> [22]https://github.com/w3c/web-annotation/blob/gh-pages/model/w d2/examples/correct/anno41.json [22] https://github.com/w3c/web-annotation/blob/gh-pages/model/wd2/examples/correct/anno41.json azaroth: example 41 [^] is a completely contrived example at the end of the spec ... it seems unlikely (in the short term at least) that any client would generate such an annotation ... but perhaps not outside the realms of possibility ... putting in the 3.2 set of tests, it passes 8 but fails 5 ... wondering what those are: problems with the test harness, the SHOULD problem, or problems with the data? TimCole: would need to have a look, but it's probably the SHOULD issue with multiple formats <Zakim> azaroth, you wanted to ask about fails for example 41 azaroth: we should probably spend some time talking about protocol testing Protocol Testing bigbluehat: have mostly passed the work I've done onto ShaneM ShaneM: the server tests bigbluehat are awesome, but let's talk about client tests for a second ... the server runs in the WPT environment <azaroth> ( issue [23]https://github.com/w3c/web-annotation/issues/344 ) [23] https://github.com/w3c/web-annotation/issues/344 ShaneM: the way annotations work is that an annotation collection lives at a IRI, and thus the server needs to serve at some named route within WPT <bigbluehat> PUT to overwrite ShaneM: but we need to work out how create/update/destroy operations work in the test server ... in particular because we don't want ephemeral data created on the test server ... so we're going to be arranging things such that data created by clients is destroyed as soon as it is read back from the server ivan: how is that going to work if you want to prepare a bunch of data and then run a load of tests against that? ShaneM: in those cases the client will access some static collection of annotations rather than data they created <azaroth> scribenick: azaroth nickstenn: Question about how this is going to work -- the protocol spec doesn't say what the server is supposed to do with the data that you give it, even reasonable things ... for example, in a distributed annotation system, you POST to create it, but you might not be able to read it back again straight away ... want us to be careful that we're not testing that you can read something straight away bigbluehat: The protocol says that it comes back with the full representation ivan: Comes back with an id nickstenn: That's a different point though. Could return it straight away, but the server doesn't necessarily have state beyond that ... intuitively reasonable assumptions are fine, but that's not in the spec ... need to be careful to not write tests on our understanding of the spec, but what the spec actually says bigbluehat: e.g. there's no guarantee that you'll be able to get the annotation back after you create it ShaneM: Wondering if the server tests you wrote rely on creating and then immediately retrieving it bigbluehat: would need to go through the tests to see if it requests the same ones later, I don't think so ... it puts stuff in but I don't think it checks again later ShaneM: If that's the case, then we're good <bigbluehat> this is the thing ShaneM's been mentioning btw: [24]https://github.com/BigBlueHat/web-annotation-protocol-teste r [24] https://github.com/BigBlueHat/web-annotation-protocol-tester ShaneM: in our last 2m, let's just agree on what we think is going to happen over the next week ... I have a couple of next actions from this conversation and will get on those straight away ... I have another to ensure people know how to run tests and upload results ... also working on getting the protocol stuff implemented TimCole: going to fill in a few more schemas to cover sections 3 and 4 ... if people could help make some invalid annotations to help test the failure cases, that would be helpful azaroth: I can do some of that ivan: When do you folks think we can begin to pester implementers to provide reports? TimCole: I think we need to do a reality check next Friday before we start inviting implementers -- maybe the week after that? Adjourn Summary of Action Items Summary of Resolutions 1. [25]Minutes of the previous call are approved: https://www.w3.org/2016/08/05-annotation-minutes.html [End of minutes] __________________________________________________________ Minutes formatted by David Booth's [26]scribe.perl version 1.144 ([27]CVS log) $Date: 2016/08/12 16:07:50 $ [26] http://dev.w3.org/cvsweb/~checkout~/2002/scribe/scribedoc.htm [27] http://dev.w3.org/cvsweb/2002/scribe/
Received on Friday, 12 August 2016 16:10:12 UTC