- From: Michael Cooper <cooper@w3.org>
- Date: Wed, 23 Sep 2015 12:17:23 -0400
- To: www-multimodal@w3.org
- Cc: PF <public-pfwg@w3.org>, Gottfried Zimmermann <zimmermann@accesstechnologiesgroup.com>
The Protocols and Formats Working Group has the following two comments
on Discovery & Registration of Multimodal Modality Components: State
Handling http://www.w3.org/TR/2015/WD-mmi-mc-discovery-20150611/
1) All figures in the doc are available as high-res images upon mouse
click, but this functionality is not available by keyboard. E.g. <img
onclick="window.open('http://upload.wikimedia.org/wikipedia/commons/3/38/Push_high.png','Push');">.
This should be made an <a> element to allow for keyboard interaction.
There are Web users who use no mouse, but would like to see the images
in high resolution. Also, all figures lack a long description.
2) The most interesting part for PF is the context object. However, the
doc contains no specification for context. In [1], context data is
simply not defined: "The format and meaning of this data is
application-specific." If [1] was to be revised, it would be good to
provide examples of context data which can be used to define a user's
preferences (e.g. pointer to a GPII personal preference set), a device's
characteristics, and situational parameters. However, it is not clear
if the context data could also include information on dynamic aspects of
the interaction, e.g. a sudden increased noise level around the user.
[1] Multimodal Architecture and Interfaces. W3C Recommendation 25
October 2012. http://www.w3.org/TR/2012/REC-mmi-arch-20121025/
Michael Cooper
PFWG staff contact
Received on Wednesday, 23 September 2015 16:17:26 UTC