- From: David Pautler <david@intentionperception.org>
- Date: Tue, 14 Apr 2015 21:30:09 +1200
- To: public-vxmlcg@w3.org
Received on Tuesday, 14 April 2015 09:30:40 UTC
Hi VXMLers, I've been searching W3C and reading papers on multimodal UI looking for ways to extend VXML 2.1 for multi-channel grammar and interpretation (e.g., by adding gesture channels). Two questions for you: 1. Is there any candidate spec or informal work that extends vxml in this way? I've read several papers on speech and gesture simultaneous recognition, and all assume multi-channel grammars so that one channel can disambiguate others. (The only W3C spec in this space I'm aware of is SCXML, but it seems to require each channel to be fully disambiguated.) 2. Is there an official dtd or schema for the vxml 2.1 data element to allow the ecmaxmltype attribute? It would be great to use the w3c validator with such a resource (instead of creating my own). Cheers, David Pautler
Received on Tuesday, 14 April 2015 09:30:40 UTC