- From: Shane McCarron <shane@spec-ops.io>
- Date: Thu, 8 Sep 2016 09:25:37 -0500
- To: "Cole, Timothy W" <t-cole3@illinois.edu>
- Cc: W3C Public Annotation List <public-annotation@w3.org>
Received on Thursday, 8 September 2016 14:27:03 UTC
You are correct - there is no way to record metadata like that in WPT as it
stands. What I would suggest is that the implementor submit
XXNN-inputN.json or something as part of the PR to demonstrate what input
was tested. These could even be in a folder with an input file per test
case (this would lead to automation in the future). In that case something
like:
XXNN-input/
annotations/
annotationAgentOptionals-manual.jsonld
annotationMusts-manual.jsonld
annotationOptionals-manual.jsonld
Optional, of course. But if the vendor likes, that would be most helpful.
On Thu, Sep 8, 2016 at 8:25 AM, Cole, Timothy W <t-cole3@illinois.edu>
wrote:
> Shane-
>
> As best I can tell, the json-ld submitted through our test suite does not
> record anywhere the json-ld that is tested. Is this correct?
>
> (We do now have a place to record a mapping between codes and clients
> http://w3c.github.io/test-results/annotation-model/README.md)
>
> If so, would it be worth providing a means by which implementers can
> submit the json-ld tested alongside the test results resports they submit?
> This would be useful if bugs are discovered in our test schemas (likely).
> Tests could then be rerun.
>
> Thanks,
>
> Tim Cole
>
--
Shane McCarron
Projects Manager, Spec-Ops
Received on Thursday, 8 September 2016 14:27:03 UTC