- From: James A Larson <jlarson@infotoday.com>
- Date: Wed, 18 May 2022 10:58:40 -0700
- To: Kazuyuki Ashimura <ashimura@w3.org>, Dave Raggett <dsr@w3.org>, jeff@w3.org, public-voiceinteraction@w3.org
- Cc: Deborah Dahl <ddahl@infotoday.com>, "jon.stine@openvoicenetwork.org" <jon.stine@openvoicenetwork.org>
- Message-ID: <6c66c9e9-cb7e-8585-5097-fb67bbdeeab0@infotoday.com>
Kaz, Dave, and Jeff,
Interoperable Voice Agents on the web is a hot topic these days. I
encourage the W3C to host a workshop to determine how W3C can
participate in this world- wide effort. Below is a draft description of
a proposed workshop.
Regards
Jim Larson
Open Voice Forum
(and former chair of W3C Voice Browser Working Group)
*Interoperable Voice Agents on the Web *
**
*General background*
* Thanks to the advancement of HTML5 and related Web technologies, Web
applications with speech capability are getting more popular for
ease of user-interaction and richer user experience such as Apple's
SIRI, Google's Voice Assistant, Amazon's Alexa, and others.
Plug-ins, such as Speechly, https://www.speechly.com/, enable users
to speak directly into any web application.
* Users interact with voice agents by exchanging information via
spoken (or textual) dialogs. Users speak and listen by using speech
recognition, speech synthesis and natural language processing
technologies.
* Voice agents are one of the essential applications available through
various devices including smart speakers, mobile phones, tablet
devices, eBook readers, and gaming platforms, automobiles, and IOT
devices such as TV’s, household appliances, security systems and
garage door openers.
* During the breakout session, [NextDirections for Voice and the Web
Breakout – 18 October 2021 (w3.org)
<https://www.w3.org/2021/10/18-voice-minutes.html>, there was
discussion about potential needs for improved interaction among
voice agents for web services.
* W3D recently published “Natural Language Interface Accessibility
User Requirements [Natural Language Interface Accessibility User
Requirements (w3.org) <https://www.w3.org/TR/naur/>] identifies
several requirements for interoperable voice agents
*Focus*
* The current status of integration of the voice-enabled agents from
multiple vendors.
* What is needed for the voice interaction technology to be deployed
globally for all the languages throughout the world.
* How voice agents interact with each other—how they can share data
and processes.
* Discovery, location, and connectivity to voice agents throughout the
world, including DNS, well-formed meta data, schema.org meta data,
and meta data aggregation
* Privacy and security
* Accessibility and usability
* Internationalization and compatibility with region-specific technology
*Applicable areas of application*
* Voice agent
* Connected car
* Smart homes/Smart factories/Smart cities
* Smart speakers/Smartphones as a portal/user device
* IoT
*Workshop deliverables*
·Barriers to voice agent interoperability.
·Summary of current efforts on voice agent interoperability.
·Mission statement for W3C's future efforts on voice and the Web.
*Who should attend?*
* Many possible stakeholders including:
o Service providers/System implementers
o Govt (like Singapore)
o Users from various countries/communities
* Liaisons
o Open Voice Network
o MPIA Community
o Stanford University Open Assistant Lab
o Technology providers (e.g., Amazon, Apple, Google)
o oneM2M
o Singapore Govtech
Received on Wednesday, 18 May 2022 17:58:56 UTC