- From: Tim Thompson <timathom@gmail.com>
- Date: Wed, 10 Oct 2018 18:09:40 -0400
- To: t-cole3@illinois.edu
- Cc: byoung@bigbluehat.com, gklyne@googlemail.com, sgharms@stevengharms.com, public-openannotation@w3.org
- Message-ID: <CAPPeUfgNGLNU7Jcn3wMXfyuuGgEscnJK5ZvXd9fZL8azjps+6A@mail.gmail.com>
Ha! Thanks, Tim. From what I recall, there weren't any extensions to the WA model needed for the project. I think that oa:TextQuoteSelector was the only SpecificResource selector used. I see that Princeton's Digital Humanities Center has launched a more ambitious Derrida project that includes image annotations, but I'm not sure to what extent they are using WA: https://derridas-margins.princeton.edu/. -- Tim A. Thompson Discovery Metadata Librarian Yale University Library On Wed, Oct 10, 2018 at 5:58 PM Cole, Timothy W <t-cole3@illinois.edu> wrote: > Of possible interest on this topic might be a small experiment that Tim > Thompson (then at Princeton University Library, now at Yale University > Library) and his colleagues did attempting to use the Web Annotations model > to capture digitally the handwritten dedications (annotations) found in a > collection of physical rare books at Princeton. They were particularly > focused on identifying references in the bodies of the dedications to > people, places and events that are represented in some way on the Web. It > was only an experiment and as I recall it required some extensions to make > Web Annotation work as they wanted, but if interested, see: > > https://wiki.duraspace.org/display/LD4P/Princeton+-+Derrida%27s+library > > https://github.com/pulcams/ld4p > > > > and if you are a near a library that has a subscription to Journal of > Library metadata, Thompson, et al. published an article about the project > here: > > https://doi.org/10.1080/19386389.2016.1258908 > > > > Thanks, > > Tim Cole > > University of Illinois at Urbana-Champaign > > > > > > *From:* Benjamin Young <byoung@bigbluehat.com> > *Sent:* Wednesday, October 10, 2018 4:16 PM > *To:* Graham Klyne <gklyne@googlemail.com>; sgharms@stevengharms.com > *Cc:* public-openannotation@w3.org > *Subject:* Re: Web annotations for physical texts > > > > Glad to hear this is being explored! And thank you both for adding your > thoughts here. Steven, special thanks to you for kicking this off (and > great post, btw!). :) > > > > One option for books is to use URN's for the targets: > > https://www.iana.org/assignments/urn-namespaces/urn-namespaces.xhtml > <https://urldefense.proofpoint.com/v2/url?u=https-3A__www.iana.org_assignments_urn-2Dnamespaces_urn-2Dnamespaces.xhtml&d=DwMFAg&c=OCIEmEwdEq_aNlsP4fF3gFqSN-E3mlr2t9JcDdfOZag&r=1SG5aXop0rkp_O1wnbIf7XweTLUSe9Z3aj8AqEIzm8c&m=EVNW3iuSYmx3j_7RAp0RMh4rdkIKXldktdocVtWbKG0&s=BKisInvstmeLre454YrWfH0dFILIEJT62OzcdyOzWvE&e=> > > > > Of the ones registered there you'll find `isbn` and `issn` among several > others. > > > > The next steps (for physical things) is how you refine that target--into > what Web Annotation Data Model calls a SpecificResource: > > https://www.w3.org/TR/annotation-vocab/#specificresource > <https://urldefense.proofpoint.com/v2/url?u=https-3A__www.w3.org_TR_annotation-2Dvocab_-23specificresource&d=DwMFAg&c=OCIEmEwdEq_aNlsP4fF3gFqSN-E3mlr2t9JcDdfOZag&r=1SG5aXop0rkp_O1wnbIf7XweTLUSe9Z3aj8AqEIzm8c&m=EVNW3iuSYmx3j_7RAp0RMh4rdkIKXldktdocVtWbKG0&s=wyn-RdSf622lNSwYGZjBmh7OV-jcrholGaVa9jMywdM&e=> > > > > Essentially, those are a target.source (the URN above), and some sort of > "selector" (or locator). That's where the invention and exploration > probably need the most work. :) Usually, for print, that's some combination > of page number, line number, and/or exact quote (or text range, etc). > > > > I'd be very curious to see this explored more, and hope that I and others > here can help in that in some way. :) > > > > Thanks for sharing your hopes here! > > Benjamin > > -- > > http://bigbluehat.com/ > <https://urldefense.proofpoint.com/v2/url?u=http-3A__bigbluehat.com_&d=DwMFAg&c=OCIEmEwdEq_aNlsP4fF3gFqSN-E3mlr2t9JcDdfOZag&r=1SG5aXop0rkp_O1wnbIf7XweTLUSe9Z3aj8AqEIzm8c&m=EVNW3iuSYmx3j_7RAp0RMh4rdkIKXldktdocVtWbKG0&s=i1vcJ2tH7kGn_4pKM-4ZVVameImu5dHBNDZqdzPe5nA&e=> > > http://linkedin.com/in/benjaminyoung > <https://urldefense.proofpoint.com/v2/url?u=http-3A__linkedin.com_in_benjaminyoung&d=DwMFAg&c=OCIEmEwdEq_aNlsP4fF3gFqSN-E3mlr2t9JcDdfOZag&r=1SG5aXop0rkp_O1wnbIf7XweTLUSe9Z3aj8AqEIzm8c&m=EVNW3iuSYmx3j_7RAp0RMh4rdkIKXldktdocVtWbKG0&s=Q2E4JxOSh_eO1fdSZj8dTTqMi878AvhC_ZXGqr2aJJ8&e=> > ------------------------------ > > *From:* Graham Klyne <gklyne@googlemail.com> > *Sent:* Wednesday, October 10, 2018 5:04 PM > *To:* sgharms@stevengharms.com > *Cc:* public-openannotation@w3.org > *Subject:* Re: Web annotations for physical texts > > > > I'm currently doing some linked data work using Web Annotations applied to > physical places, and I'm not seeing any real problems with this (just a > need to be clear what a URI is referring to). I'm actually finding them to > be quite a powerful tool for capturing contextualised descriptions in > linked data (with a modest additions). > > > > #g. > > > > On Wed, 10 Oct 2018, 21:46 Steven Harms, <sgharms@stevengharms.com> wrote: > > Greetings, > > > > I am interested in creating annotations on physical books [1 > <https://urldefense.proofpoint.com/v2/url?u=https-3A__stevengharms.com_research_semweb-2Dtopic_problem-5Fstatement_&d=DwMFAg&c=OCIEmEwdEq_aNlsP4fF3gFqSN-E3mlr2t9JcDdfOZag&r=1SG5aXop0rkp_O1wnbIf7XweTLUSe9Z3aj8AqEIzm8c&m=EVNW3iuSYmx3j_7RAp0RMh4rdkIKXldktdocVtWbKG0&s=I-gf2hS5jLbpw2TPKImJ9xzyw6kCkh6YCcI8tjnnmNI&e=> > ]. > > > > As the name "web annotations" suggests, the default target of the Web > Annotation Working Group would be, of course, to annotation IRI-referable > targets with IRI-identifiable Annotations. > > > > 1. Is there a model whereby we could point to a physical resource in a URI > / IRI format (and thus join the existing Web Annotation universe, *or* > > 2. Is there a framework that might support referring to physical books > that I've simply not found > 3. Or should I plan to use JSON-LD to create "forge my own path?" > > > > I hope to post an example of what #3 might look like, but I'd like to > double check my understanding before engaging in in such an effort, *tabula > rasa*. > > > > Regards, > > > > Steven > > > > > > [1]: https://stevengharms.com/research/semweb-topic/problem_statement/ > <https://urldefense.proofpoint.com/v2/url?u=https-3A__stevengharms.com_research_semweb-2Dtopic_problem-5Fstatement_&d=DwMFAg&c=OCIEmEwdEq_aNlsP4fF3gFqSN-E3mlr2t9JcDdfOZag&r=1SG5aXop0rkp_O1wnbIf7XweTLUSe9Z3aj8AqEIzm8c&m=EVNW3iuSYmx3j_7RAp0RMh4rdkIKXldktdocVtWbKG0&s=I-gf2hS5jLbpw2TPKImJ9xzyw6kCkh6YCcI8tjnnmNI&e=> > > > > -- > > Steven G. Harms > > PGP: E6052DAF > <https://urldefense.proofpoint.com/v2/url?u=https-3A__pgp.mit.edu_pks_lookup-3Fop-3Dget-26search-3D0x337AF45BE6052DAF&d=DwMFAg&c=OCIEmEwdEq_aNlsP4fF3gFqSN-E3mlr2t9JcDdfOZag&r=1SG5aXop0rkp_O1wnbIf7XweTLUSe9Z3aj8AqEIzm8c&m=EVNW3iuSYmx3j_7RAp0RMh4rdkIKXldktdocVtWbKG0&s=F8nzYd_i33TWUHsfqDg39OREr6mRQO5Z5ZqR7CdGBYI&e=> > >
Received on Wednesday, 10 October 2018 22:10:29 UTC