- From: T.V Raman <raman@google.com>
- Date: Tue, 11 Sep 2007 11:22:25 -0700
- To: lisa@ubaccess.com
- Cc: unagi69@concentric.net, chaals@opera.com, st@isoc.nl, public-html@w3.org, wai-xtech@w3.org
+Gregory's summary though written in a nice logical progression starts with a couple of strong assumptions which leads to the conclusions he arrives at. I've used a self-voicing app -- Emacspeak -- for the last 12 years, and everyone here who knows me knows that I am more than a casual computer user. So Gregory -- in future, when you write something like this, clearly document your opening assumption. In this particular case, your opening assumptions were: A) User is victim to a locked-down "user environment" B) You confused "operating system" with "user environment". In the Linux / Emacspeak case, neither (A) and (B) are true, which consequently debunks most of what you wrote --- Lisa Seeman writes: > > Gregory wrote: self-voicing apps have their place in the overall scheme of > things but they are NOT substitutes for screen readers. > > Two places were they have an important role is for people with learning or > language disabilities. Another use is for people who do not have their own > computer, and can use firevox on a shared computer, such as the library > (which may not be prepared to install a bulky program such as Jaws but will > be prepared to help someone get started). Also as people develop vision > problem (such as associated with diabetes and aging) may often use the self > voicing apps for reading print. This group do not need a screen reader for > selecting icons at the start up but they may not want to, or even be able > to, manage a screen reader (which take quite good memory skills). Another > huge group who need to be taken into account are third word computer users > who may be unable to afford a screen reader or may not read well. > > So they are not a substitutes for screen readers but self-voicing apps > have an important place in the world. > > All the best > > Lisa > > > > > > > -----Original Message----- > From: wai-xtech-request@w3.org [mailto:wai-xtech-request@w3.org] On Behalf > Of Gregory J. Rosmaita > Sent: Monday, September 10, 2007 9:29 PM > To: Charles McCathieNevile; Sander Tekelenburg; public-html@w3.org; > wai-xtech@w3.org > Subject: screen-reader versus self-voicing app (was: Re: Screen-reader > behaviour) > > > aloha! > > as a screen-reader user, let me attempt to explain why there is no > groundswell of support for "self-voicing" applications by those dependent > upon speech output... > > 1) unavoidable black holes: > self-voicing applications cannot replace a dedicated screen reader, for > self-voicing applications often cannot interpret key parts of the chrome, > espeically if the chrome does not reuse standard control sets for the OS on > which it is running -- download interfaces, view source interfaces (that > open up a new browser instance or tab), the ability to "browse" files to be > uploaded to a web site, etc. this is because the self-voicing application > exists solely to voice the application which is currently running; > > 2) one can put one's screen-reader into "sleep" mode for a particular app, > so that the self-voicing app doesn't conflict with the screen reader, but > this often leads to unexpected and undesired results; > > for example, in order to use FireVox, i set JAWS to become inactive whenever > FireVox is loaded -- however, since FireVox is an extension, and not a > seperate app, i can no longer run FireFox with a screen reader, because the > screen reader cannot differentiate between the synonymic executable files > when invoked, and therefore, disables screen reader interaction when a > partucular executable is loaded; > > 3) self-voicing apps can still conflict with a scren reader due to events > from the self-voicing apps firing whilst one is in a plain text document > checking one's credit card or banking information; which is also why > self-voicing applications have limited appeal and why they CANNOT be run > without a screen reader -- if i am using a self-voicing app, once i switch > tasks, i have no wey of knowing what is currently running -- even when doing > something as "trivial" as copying the contents of a page to the clipboard > and pasting it to an empty plain text document -- without a screen-reader at > the ready to "awaken" when the user switches from the self-voicing app, the > speech-dependent user is left without a means of ensuring that previous > information has not been overwritten, nor what directory into which the file > is going to be saved, nor access to any system calls ("do you want to > overwrite..." or "error - nothing selected" > > self-voicing apps have their place in the overall scheme of things, but they > are NOT substitutes for screen readers; what we should be concentrating upon > is NOT how does current assisstive technology handle current markup, but how > to enable assisstive technology to handle markup better, by providing more > explicit association patterns and as much semantic information as possible > > THAT is the goal -- to improve bi-directional communication between > applications, in this case, between user agents and screen readers -- not to > critique the current state of support -- it must ALSO be realized that HTML > 4.01 did not proclaim that it had addressed all accessibility problems, only > those that emergency triage units identified as the most crucial problems in > the late nineteen-nineties -- it was NEVER intended to be the be all or end > all in web accessibility, but an effort to provide a means of breaching > perceptual black holes and the sort of device dependence and modality > dependence that breaks assisstive technologies... even for a self-voicing > app to work well, it must rely upon the semantics built into the markup > language it supports... > > this is why this whole thread is a red herring in my opinion -- we cannot > "break" what was done in the past to promote accessibility, useability, > internationalization and device independence, nor should we be bound to > putting old wine into new bottles -- where superior mechanisms are > available, they should be implemented, but those mechanisms implemented in > HTML 4.01 specifically for accessibility, device independence and > internationalization MUST be supported as part of the "backwards > compatibility" principle, hence my suggested verbiage for the design > principle document: > > "Browsers should retain residual markup designed for a specific > purpose, such as accessibility. internationalization or device > independence. Simply because new technologies and superior > mechanisms have been identified, not all of them have been > implemented. Moreover, disabled users are more likely to be > users of "legacy technology" because it is the only technology > that interacts correctly with third-party assistive technologies" > > or words to that effect... > > gregory. > > > -- > "He who lives on Hope, dies farting." > -- Benjamin Franklin, Poor Richard's Almanack > -- > Gregory J. Rosmaita, unagi69@concentric.net Camera Obscura: > http://www.hicom.net/~oedipus/ > > > > -- Best Regards, --raman Title: Research Scientist Email: raman@google.com WWW: http://emacspeak.sf.net/raman/ Google: tv+raman GTalk: raman@google.com, tv.raman.tv@gmail.com PGP: http://emacspeak.sf.net/raman/raman-almaden.asc
Received on Tuesday, 11 September 2007 18:23:25 UTC