- From: Ivan Herman <ivan@w3.org>
- Date: Fri, 2 Sep 2016 18:03:43 +0200
- To: W3C Public Annotation List <public-annotation@w3.org>
- Message-Id: <2547A00D-5F7B-4FA7-A1F4-0F117713E6C4@w3.org>
Minutes are here: https://www.w3.org/2016/09/02-annotation-minutes.html Text version below Ivan ---- Ivan Herman, W3C Digital Publishing Lead Home: http://www.w3.org/People/Ivan/ mobile: +31-641044153 ORCID ID: http://orcid.org/0000-0003-0782-2704 [1]W3C [1] http://www.w3.org/ Web Annotation Working Group Teleconference 02 Sep 2016 [2]Agenda [2] http://www.w3.org/mid/05b001d204a2$1334dbd0$399e9370$@illinois.edu See also: [3]IRC log [3] http://www.w3.org/2016/09/02-annotation-irc Attendees Present Rob Sanderson (azaroth), Dan Whaley, Tim Cole, Ben De Meester (bjdmeest), Jacob Jett, Ivan herman, ShaneM, Takeshi Kanai, Nick Stenning, Randall Leeds (tilgovi) Regrets TB_Dinesh Chair Tim, Rob Scribe bjdmeest Contents * [4]Topics 1. [5]Minutes 2. [6]CRs update 3. [7]Extension request 4. [8]Testing * [9]Summary of Action Items * [10]Summary of Resolutions __________________________________________________________ <azaroth> trackbot, start meeting <trackbot> Meeting: Web Annotation Working Group Teleconference <trackbot> Date: 02 September 2016 <azaroth> Chair: Tim_Cole, Rob_Sanderson <TimCole> Meeting: Web Annotation Working Group Teleconference <trackbot> Sorry, ivan, I don't understand 'trackbot does all the rest for you'. Please refer to <[11]http://www.w3.org/2005/06/tracker/irc> for help. [11] http://www.w3.org/2005/06/tracker/irc <ShaneM> working on it <trackbot> Sorry, dwhly, I don't understand 'trackbot, get coffee'. Please refer to <[12]http://www.w3.org/2005/06/tracker/irc> for help. [12] http://www.w3.org/2005/06/tracker/irc <ivan> scribenick: bjdmeest TimCole: Let's get started ... first, we'll talk about the exit criteria of CR ... then, about extending the WG to get through CR, PR.. ... then, we'll talk about testing ... other topics? <TimCole> PROPOSED RESOLUTION: Minutes of the previous call are approved: [13]https://www.w3.org/2016/08/26-annotation-minutes.html [13] https://www.w3.org/2016/08/26-annotation-minutes.html Minutes <azaroth> +1 <ivan> +1 <TimCole> +1 <Jacob> +1 +1 <takeshi> +1 RESOLUTION: Minutes of the previous call are approved: [14]https://www.w3.org/2016/08/26-annotation-minutes.html [14] https://www.w3.org/2016/08/26-annotation-minutes.html CRs update azaroth: we had a request ... we should publish the exit criteria ... that's required ... we have done that ... there are new versions of the 3 specs (each with an appendix about the exit criteria) ... implementations of the model, implementations that the vocabulary is internally consistent and can be used to go from json-ld to json ... for the protocol, 2 implementations of all the interactions ... retrieving an annotation, deleting, etc... ... they will be republished on 6th of September ivan: we also wanted to link to the test cases themselves, but they are not clearly available yet ... everything is done, the publications are checked, they will be published on Tuesday ... that's that for CR Extension request TimCole: we are trying to do an extension request to extend the WG to get through CR and PR ivan: I gave Ralph(?) an overview ... we hope to be able to cover all the exit criteria by the end of October ... that's one month extra ... that + the problem of Christmas in the middle ... my pessimistic deadline would be to publish the recommendation by the end of January, so I asked to extend until the end of February ... hopefully, we will get it ... in any case, the more we can show as readiness, the better ... we should get initial implementation reports on our pages ... they don't need to be complete ... but at the moment, the reports are placeholders ... if we have (partially) tested implementations (e.g., Rob's, Benjamin's) ... showing them is critical ... ideally by next week, realistically by the week after TimCole: test reports will show, preferably next week ivan: they will look at those test reports, as they are in the CR documents ShaneM: about results: I can now merge to the repo <TimCole> [15]https://github.com/w3c/test-results/pulls [15] https://github.com/w3c/test-results/pulls ShaneM: I will push results for our implementation, right now <TimCole> [16]https://github.com/w3c/test-results/tree/gh-pages/annotatio n-model [16] https://github.com/w3c/test-results/tree/gh-pages/annotation-model TimCole: there's a W3C test results repo on github ... there's a small typo: for ==> fork ... There's an open pull request Testing TimCole: Model testing: ... we have about 100 assertions covering body, target, .. ... I need to add a separate folder for specificResource ... those are in the test-dev repository ... you can now use those tests ... you go to the w3c test site ... you input annotations ... you get reports ... those reports, you can add using a pull request to the test-results repo ivan: what ends up in the test-results/implementation reports are a set of json files? <TimCole> [17]https://github.com/spec-ops/wptreport [17] https://github.com/spec-ops/wptreport ShaneM: that and a report TimCole: the current report doens't mention the implementation, you do know who did the pull request <ShaneM> CH53.json ShaneM: as a convention, tests name the file as the name of the implementation and the version ... I jusked as that to the current pull request ivan: all implementers we currently have, should get some kind of name? Shane: whatever name that makes sense is fine ... I'll modify the instructions so that is clear TimCole: the downloadable portion of the generator requires two characters and two numbers for the file.json ShaneM: apparently yes <Zakim> azaroth, you wanted to discuss names azaroth: is it possible to have additionale information about the things with names? ... e.g. a link for every implementation? a registry? ShaneM: we can put that in the readme TimeCole: The pull requester could add extra files, no? Then we could tell them what we want extra ivan: does the report make an automatic count, i.e., how many implementations per test, for the CR, or do we have to create that afterwards? ShaneM: it creates as separate report ... if we want to make changes we can, but I don't want to change the environment too much ... there are other players in the field TimCole: we have about 45 assertions that we expect every annotation to pass, the MUSTs ... and then we have about 100, which are designed to catch optionals ... so, if someone only implements an optional body, and a simple target, it seems as if they fail a lot of tests (the optional target tests) ... can we catch that some way, explain that to people, that they don't 'fail' as much as it seems? ShaneM: this is a meta-conversation about what to do about optional features <azaroth> +1 to that reduction TimCole: I reduced the tests a bit, e.g. for text direction, it doesn't depend on which type of body, so that helps a bit ivan: how do we do the testing and reporting on the vocabulary? ShaneM: by hand <TimCole> for example, we may not decide to consider each kind of selector a separate feature requiring testing, this would reduce the number of tests. ShaneM: we take a template that looks like the current report, and fill in the rows ivan: we need to decide which validation tools we use ... for RDF vs JSON azaroth: there are tools, the Python RDFlib, and the JSON-LD tool from digital bazaar ivan: what would be the other independent toolset? ... what's the situation with json-ld tools? azaroth: it has implementations in most languages ... ruby is pretty good, also for RDF ivan: maybe we can ask greg? from json-ld POV, he would be a logical choice azaroth: what about javascript-based? ivan: RubenVerborgh has a lot of JavaScript tools ... if he could run those few tests, via his toolkit ... then we have 3 mature toolsets ... azaroth, can u ask greg? azaroth: yes ShaneM: I don't care about how you would give them, we just need to input them into the html file <TimCole> [18]http://w3c-test.org/tools/runner/index.html [18] http://w3c-test.org/tools/runner/index.html ShaneM: we need implementations for testing the annotation model TimCole: two parts of the question ... could you generate annotations conforming to the annotation model ... if so, could you input those json-ld in the test runner, generate the json file test results, and do the pull request? nickstenn: I'm not sure our client will spit out the correct JSON-LD in the near future ... but our server could render them as JSON-LD ... I'm very happy to test those using the test runner tilgovi: if it's important to have client-side javascript that generates conforming json TimCole: you have to do one annotation at a time tilgovi: ... I'll have a look at that <ShaneM> Updated result reporting instructions at [19]https://github.com/w3c/test-results/tree/gh-pages/annotatio n-model and [20]https://github.com/w3c/test-results/tree/gh-pages/annotatio n-protocol [19] https://github.com/w3c/test-results/tree/gh-pages/annotation-model [20] https://github.com/w3c/test-results/tree/gh-pages/annotation-protocol TimCole: it's important to have test results published bigbluehat: about protocol testing: it's about exercising a server, and exercising a client ... there's a pull request pending ... there is one test, you give it the url to your annotation server, and a url to one annotation in that server ShaneM: I've only ever run that against the basic python server ... https is a should, and the python server doesn't implement that ... about client-side protocol testing ... there are basically no requirements ... I found one about sending a pref header for a certain use case, but that doens't really have anything to do with the client azaroth: because HTTP doesn't require a specific format, and we don't extend HTTP, there are no testable assertions for the client ShaneM: I would like to either have someone test against a server, or give me links to a server, and I'll run the tests ivan: so we need to reach out to the various implementers, such as Europeana azaroth: they have one, after a slight update ... it would take some time to have it up and running somewhere accessible <ShaneM> [21]http://testdev.spec-ops.io:8000/tools/runner/index.html?pat h=/annotation-protocol [21] http://testdev.spec-ops.io:8000/tools/runner/index.html?path=/annotation-protocol ShaneM: you can do it yourself, they're in test-dev right now <ivan> adjourned <TimCole> Adjourn TimCole: hopefully, by next week, we have some reports, and more specifics about the vocabulary testing <ivan> trackbot, end telcon Summary of Action Items Summary of Resolutions 1. [22]Minutes of the previous call are approved: https://www.w3.org/2016/08/26-annotation-minutes.html [End of minutes] __________________________________________________________ Minutes formatted by David Booth's [23]scribe.perl version 1.144 ([24]CVS log) $Date: 2016/09/02 16:01:47 $ [23] http://dev.w3.org/cvsweb/~checkout~/2002/scribe/scribedoc.htm [24] http://dev.w3.org/cvsweb/2002/scribe/
Received on Friday, 2 September 2016 16:03:58 UTC