- From: Gregory J. Rosmaita <oedipus@hicom.net>
- Date: Tue, 11 Dec 2007 03:09:55 +0000
- To: wai-xtech@w3.org, wai-liaison@w3.org
no one nor any working group has directly addressed or is able to answer
the question:
"In the absence of a screen-reader or screen-magnification program or
any other application that tracks a UA cursor and provides flow,
whose responsibility is it to fire aural cues?"
(last attempt to rekindle this discussion:
<http://lists.w3.org/Archives/Public/wai-xtech/2007May/0000.html>)
an extremely important related question is:
"Shouldn't aural styling be extended to control embedded or designated
onAction/onMouseOver/onFocus/onHover/onWhatEver audio events so as to
ensure user control over what are currently usually javascripted
hacks?"
Why are these issues important to consider?
In the use scenario of a screen-magnifier with an aural overlay of
earcons, the screen-magnifier may trigger or request the playing of a
sound-clip, but what application is supposed to provide that aural
feedback? The User Agent? The underlying operating system?
It should be a backplane operation, but someone or something must take
responsibility for the actual aural rendering of the earcon -- we need
to discuss where to address this point in terms of conformance criteria
some (mostly recycled) thoughts on triggering and rendering:
"who plays the target of cue, play-during and other aural icons?"
i firmely believe this needs to be a backplane event, which means that
the
triggering mechanism needs access to the device's default sound renderer
(i.e. that which produces system sounds and other sounds under-the-hood,
without opening or depending on a third party audio file player)
play-during play-before are event driven, as are sounds associated with
MouseOver events, such as the :hover; pseudo-element. The problem with
Aural CSS alone is that it paints to an event responsive timeline, and
therefore it needs to get permission from divergent operating systems
and divergent user agents, to layer sounds, such as in the example of an
audio file set to :play-during while speech is simultaneously being
output)
In the use scenario of a screen-magnifier with an aural overlay of
earcons, but no synthesized speech, the screen-magnifier may trigger or
request the playing of a sound-clip, but what application is supposed to
provide that aural feedback? The user agent? The underlying operating
system? It should be a backplane operation, but someone or something
must
take responsibility for the actual aural rendering of the earcon.
And that means that there is the need for a standardized triggering
mechanism, but where to address this need? is this as much a WebAPI issue
as much as a CSS issue?
gregory.
-------------------------------------------------------
lex parsimoniae:
* entia non sunt multiplicanda praeter necessitatem.
-------------------------------------------------------
the law of succinctness:
* entities should not be multiplied beyond necessity.
-------------------------------------------------------
Gregory J. Rosmaita, oedipus@hicom.net
Camera Obscura: http://www.hicom.net/~oedipus/
-------------------------------------------------------
Received on Tuesday, 11 December 2007 03:10:07 UTC