- From: Jeanne Spellman <jeanne@w3.org>
- Date: Mon, 13 Jul 2015 13:57:51 -0400
- To: public-wai-rd@w3.org
- CC: Henny Swan <hswan@paciellogroup.com>, Susann Keohane <skeohane@us.ibm.com>
Mark,
Thanks for the invitation to attend the RDWG meeting for the discussion
of IoT. Henny Swan (TPG), Susann Keohane (IBM) and I have been
discussing IoT Accessibility with an informal goal of creating IoT
Accessibility use cases to share with groups working on IoT and WoT.
This is a quick synopsis of our ideas to date:
1) Make the output of the sensors standardized and available to the
accessibility APIs:
Examples:
* smart watch that has multi functions like tracking activity or
blood pressure - the device itself may not have an accessible UI but
the user can access the data through a mobile app or Website.
* self-driving cars where the interface in the car may not be
accessible but the car itself could be managed through an app.
* home thermostat which is challenging due to many of the devices
having a multi-function UI of displaying the temp and setting the temp
*
Smart thermostat out of reach of a wheelchair user, the thermostat
data needs to be able to be sent to a mobile or web app.
*
Blind person needs the data and controls sent to a mobile or web
app that is accessible on their mobile device, even if the original
controller is not accessible.
Interaction with the sensors needs to be available to the AAPIs
* Example of the automated home where an alert needs to blink the
lights for someone who is deaf. The user needs to be able to program
a pattern of light blinking to have a unique visual indication to
replace different audio alerts. (e.g. the door alarm should have a
different light blinking pattern than the oven timer.)
Data sent to the user must be text and not an image of text
* Example of a smart thermostat with an image of the temperature which
cannot be spoken when sent to a mobile apps
Data sent should be modality independent (e.g. a battery charger should
send "discharged, charging, full" not "red, yellow, green"
Signage:
Airport sign: A blind person is looking for a restaurant. Web App of
the airport. Sign is in a location it knows, and has data about finding
the restaurant. The data on the sign location could be sent to a user’s
navigation app on their phone. Microsoft has a new glasses that will do
in-building navigation. Buzz a wearable for someone who is deaf, or
give text directions to a wearable.
Cognitive issues:
* Important to have simplified interfaces for controllers with clear
resets when a user disables something important. (e.g. an elderly
person accidentally turning off the heat in a smart house on a cold
winter night. )
Controller Interfaces:
* Providing skins for controller apps with varying accessibility needs
-- magnified for low vision, low contrast, high contrast, simplified
for different purposes.
We think we are just starting to skim the surface. I'm glad to see
there is a WAI group that is looking at these issues.
jeanne
Received on Monday, 13 July 2015 17:57:57 UTC