- From: Marc Schroeder <marc.schroeder@dfki.de>
- Date: Fri, 26 Aug 2011 11:59:05 +0200
- To: Janina Sajka <janina@rednote.net>, W3C WAI Protocols & Formats <w3c-wai-pf@w3.org>
- CC: w3c-mmi-wg <w3c-mmi-wg@w3.org>, www-multimodal@w3.org
Hello Janina,
this is to reply to one of the comments on the EmotionML by W3C WAI PF
that you sent on 20 June [1], specifically:
ISSUE-184 Accessibility use cases for EmotionML
The group has discussed the issue and proposes the following solution:
ACCEPT
We have included the use case examples you sent into section 1.1
"Reasons for defining an Emotion Markup Language" of the specification;
the text describing use cases now reads as listed below.
Please let us know if you agree with this resolution within 14 days,
i.e. by 9 September. Should we not hear from you by that date, we will
consider this to represent implicit approval, but explicit feedback is
always better.
Thanks and best regards,
Marc
Excerpt of current editor's draft of EmotionML including the proposed
change:
"""
Concrete examples of existing technology that could apply EmotionML include:
- Opinion mining / sentiment analysis in Web 2.0, to automatically
track customer's attitude regarding a product across blogs;
- Affective monitoring, such as ambient assisted living applications
for the elderly, fear detection for surveillance purposes, or using
wearable sensors to test customer satisfaction;
- Character design and control for games and virtual worlds;
- Social robots, such as guide robots engaging with visitors;
- Expressive speech synthesis, generating synthetic speech with
different emotions, such as happy or sad, friendly or apologetic;
expressive synthetic speech would for example make more information
available to blind and partially sighted people, and enrich their
experience of the content;
- Emotion recognition (e.g., for spotting angry customers in speech
dialog systems);
- Support for people with disabilities, such as educational programs
for people with autism. EmotionML can be used to make the emotional
intent of content explicit. This would enable people with learning
disabilities (such as Asperger's Syndrome) to realise the emotional
context of the content;
- EmotionML can be used for media transcripts and captions. Where
emotions are marked up to help deaf or hearing impaired people who
cannot hear the soundtrack, more information is made available to enrich
their experience of the content.
"""
[1] http://lists.w3.org/Archives/Public/www-multimodal/2011Jun/0004.html
--
Dr. Marc Schröder, Senior Researcher at DFKI GmbH
Project leader for DFKI in SSPNet http://sspnet.eu
Team Leader DFKI TTS Group http://mary.dfki.de
Editor W3C EmotionML Working Draft http://www.w3.org/TR/emotionml/
Portal Editor http://emotion-research.net
Homepage: http://www.dfki.de/~schroed
Email: marc.schroeder@dfki.de
Phone: +49-681-85775-5303
Postal address: DFKI GmbH, Campus D3_2, Stuhlsatzenhausweg 3, D-66123
Saarbrücken, Germany
--
Official DFKI coordinates:
Deutsches Forschungszentrum fuer Kuenstliche Intelligenz GmbH
Trippstadter Strasse 122, D-67663 Kaiserslautern, Germany
Geschaeftsfuehrung:
Prof. Dr. Dr. h.c. mult. Wolfgang Wahlster (Vorsitzender)
Dr. Walter Olthoff
Vorsitzender des Aufsichtsrats: Prof. Dr. h.c. Hans A. Aukes
Amtsgericht Kaiserslautern, HRB 2313
Received on Friday, 26 August 2011 10:00:27 UTC