- From: Kazuyuki Ashimura <ashimura@w3.org>
- Date: Tue, 23 Jun 2009 23:20:41 +0900
- To: www-voice <www-voice@w3.org>
The following is the summary of the Speaker Biometrics Workshop held
on 5-6 March 2009 in Menlo Park, Califrornia, US.
We are very sorry it took so long for us to send this out.
---
Summary of the Workshop on Speaker biometrics and VoiceXML 3.0
On 5-6 March 2009, the W3C Voice Browser Working Group held a Workshop
on Speaker biometrics and VoiceXML 3.0 in Menlo Park, California, US,
hosted by SRI International.
The minutes of the workshop are available on the W3C Web server:
http://www.w3.org/2008/08/siv/minutes.html
Also the HTML version of this summary, which includes the figures of
(1) Resource control and distributed decision making and (2) Menlo
Park model, an SIV architecture, is available at:
http://www.w3.org/2008/08/siv/summary.html
There were 16 attendees from the following organizations:
* SRI International
* Recognition Technologies, Inc.
* J. Markowitz, Consultants
* Intervoice
* EIG
* Deutsche Telekom AG, Laboratories
* Centrelink
* iBiometrics, Inc.
* Daon
* Cisco Systems, Inc.
* iBiometrics, Inc.
* Voxeo
* Nuance
* General Motors
* W3C
This workshop was narrowly focused on identifying and prioritizing
directions for SIV standards work as a means of making SIV more useful
in current and emerging markets. Topics discussed during the workshop
includes:
* SIV use cases (Application requirements for SIV in VoiceXML 3.0)
* SIV users (design philosophy, uncertainty, security, identity)
* Audio formats for SIV (Wav, PCM, alaw, ulaw, OGG, etc.)
* Data format for multimodal applications (EMMA, etc.)
* SIV related standards other than W3C (CBEFF, INCITS 456, BIAS, BioAPI)
* SIV and MRCP V2
* Architecture and functionality (features, configuration, APIs, etc.)
During the workshop we have clarified "Why SIV functionality should be
added to VoiceXML" as follows.
* The system would be more responsive, so VoiceXML could shorten
customer perceived latency and provide performance benefits to
the users.
* It would be easier for developers to generate applications,
because programming interface would be consistent with the way
they use other VoiceXML resources [1] and low-level operations
would be hidden to them.
[1] http://www.w3.org/2008/08/siv/resource_control.png
* Adding SIV to a standard would make it portable and facilitate
integration with Web model, because it makes SIV applications
consistent with the model and provide efficiencies of scale in
hosted environment.
* Standardizations of easy to use API would minimize vendor
lock-in and grow the market.
* Support in VoiceXML enables SIV use (without the application
server) with intermittent/offline connectivity.
* Standards are a sign of technology maturity.
The major "takeaway" is our confirming SIV fits into the VoiceXML
space and generating the "Menlo Park Model" [2], an SIV available
VoiceXML architecture.
[2] http://www.w3.org/2008/08/siv/MenloParkModel-v003.png
The discussion on the above "Menlo Park Model" includes:
1. Main hidden security issues that people have concern about are
idendified, and ways in which they can be realistically addressed
are discussed. Those issues don't disappear but now we know we can
address them.
2. VoiceXML 3.0 could be an example of the Interaction Managers
within the W3C's MMI Architecture. The synchronization and markup
integration of multiple modalities should be addressed. There are
likely to be multiple modalities/factors involved in an interaction
using VoiceXML. Consequently, developers need a way to not
completely separate those modalities.
3. Collaboration with other W3C Working Groups and other standard
bodies, e.g., OASIS/BIAS, is expected.
Note that the discussion during the workshop was mainly focused on the
application side and provided little guidance for engine providers. So
a good standard that would create a nice wrapper around speech engines
would be needed. Classification, segmentation and identification
should be also considered and the group needs to determine whether or
not to include them in VoiceXML V3.0.
Judith Markowitz, Ken Rehor and Kazuyuki Ashimura, Workshop co-Chairs
--
Kazuyuki Ashimura / W3C Multimodal & Voice Activity Lead
mailto: ashimura@w3.org
voice: +81.466.49.1170 / fax: +81.466.49.1171
Received on Tuesday, 23 June 2009 14:21:19 UTC