- From: Miles, AJ \(Alistair\) <A.J.Miles@rl.ac.uk>
- Date: Thu, 2 Jun 2005 19:19:04 +0100
- To: "Dan Brickley" <danbri@w3.org>
- Cc: <public-esw-thes@w3.org>, <mf@w3.org>
Trying to turn this into a concrete proposal ... What about a property 'skos:audioLabel' which points to an audio/media resource? Would we need additional properties e.g. 'skos:audioDescription' to differentiate between a label and a fuller description/definition of the concept? What about a property 'skos:voiceLabel' which points to an XML literal containing some VoiceXML [1]? Or should it be left to the application to construct some voice representations of a concept from the content of the skos:prefLabel, skos:altLabel &c. properties? Thinking about this also suggests to me that there should be a super-property for skos:prefSymbol and skos:altSymbol, e.g. 'skos:symbolicLabel'. Thoughts? Cheers, Al. [1] http://www.w3.org/TR/voicexml20/ --- Alistair Miles Research Associate CCLRC - Rutherford Appleton Laboratory Building R1 Room 1.60 Fermi Avenue Chilton Didcot Oxfordshire OX11 0QX United Kingdom Email: a.j.miles@rl.ac.uk Tel: +44 (0)1235 445440 > -----Original Message----- > From: public-esw-thes-request@w3.org > [mailto:public-esw-thes-request@w3.org]On Behalf Of Dan Brickley > Sent: 17 May 2005 16:46 > To: Miles, AJ (Alistair) > Cc: public-esw-thes@w3.org; mf@w3.org > Subject: Re: FW: Audio labels (was RE: comment: FOAF Depiction and > Symbolic Labelling) > > > > * Miles, AJ (Alistair) <A.J.Miles@rl.ac.uk> [2005-05-17 14:47+0100] > > > > > > > Other extensions could also be interesting. While we could > > > debate which > > > things go in core vocab and which in other namespaces, it > > > might be more > > > fun to set that aside for now (while noting that SKOS is only > > > a Working > > > Draft at this stage, and could change), and explore possibilities > > > for such extensions. Was there something specific you had in > > > mind? Audio > > > I think could be very interesting, particularly for SKOS > > > concept that is > > > close to the electronic dictionary space, eg. lexical > databases such > > > as Wordnet (although SWBPD WG isn't using SKOS for > Wordnet currently). > > > Where a concept is lexicalised, we could point to sound clips, or > > > Speech Synth markup (eg. see > > > http://www.w3.org/TR/2004/REC-speech-synthesis-20040907/) > > > ...could have > > > interesting application to accessibility, voice/mobile > and perhaps > > > language learning apps... > > > > I like the idea of 'audio labels' ... Can anyone describe > a relatively concrete use case? > > Autogenerating voice-browser menus? > > http://www.w3.org/Voice/ > http://www.w3.org/Voice/#intro > http://www.w3.org/Voice/Guide/ > [[ > VoiceXML isnt HTML. HTML was designed for visual Web pages > and lacks the > control over the user-application interaction that is needed for a > speech-based interface. With speech you can only hear one thing at a > time (kind of like looking at a newspaper with a times 10 magnifying > glass). VoiceXML has been carefully designed to give authors full > control over the spoken dialog between the user and the > application. The > application and user take it in turns to speak: the > application prompts > the user, and the user in turn responds. > > VoiceXML documents describe: > > * spoken prompts (synthetic speech) > * output of audio files and streams > * recognition of spoken words and phrases > * recognition of touch tone (DTMF) key presses > * recording of spoken input > * control of dialog flow > * telephony control (call transfer and hangup) > ]] > > Annotation of SKOS concept descriptions with voice data > (speech markup, > or audio files, ...) could allow content tagged with those concepts to > be made navigable through VoiceXML-based interactions. Example: a > collection of blog feeds, where the RSS was augmented with > skos:subject > tagging, and the different blog drew on the same (or mapped) concept > schemes. > > [I'm working on some tools to enable blogs to pick up their SKOS > categories from their neighbours (eg. when I go to add a category to > my blog, it reminds me what categories my friends and colleagues are > using, and allows links to be expressed, sub-trees to be imported).] > > So, why would one want to navigate blogs by having the computer read > out labels for their categories? (and eg. also navigating by voice > too). > > - maybe you're driving your car, and in a traffic jam > - RSI or other accessibility reasons for not using mouse/keyboard > - you're walking around wearing some fancy bluetooth headset, > looking all Flash Gordon modern, and want to read what people are > writing about you... > - maybe you're navigating some content collection via your TV, with > menus, and prefer audio to reading of (even large) fonts > on the TV screen. > - maybe you're navigating a content collection in audio labels made > available in your native spoken language, even if the content is > in a language you're less profficient in. > - maybe you can't read the textual labels (in the language they're > available in; or in any language). > - maybe you're navigating a collection of Creative Commons-licensed > Ogg/MP3 'talking book' files on your iPod-like-thing, and > someone has > written a study guide that lets you jump around the texts based on > SKOS-indexed themes that have been collaboratively indexed against > the collection. Ok handwaving a bit here, but I think that > could be > interesting... > > Whether the final end document is read in classic Web browser, > or also via text to speech (Max was looking at this...) is a separable > choice I think. Being able to navigate around the content database > using audio labels doesn't require you to digest the content in audio > form too. > > cheers, > > Dan > > ps. is anyone on this list set up to run student projects? > maybe on in this > area could be interesting...? > >
Received on Thursday, 2 June 2005 18:19:05 UTC