- From: eric hansen <ehansen@ets.org>
- Date: Tue, 27 Apr 1999 03:58:01 -0400 (EDT)
- To: w3c-wai-gl@w3.org
Re: PRI-4 Animation and Movies This memo discusses suggested changes in wording to checkpoint 1.3 and 1.4 to address concerns of commenters. Gregg Vanderheiden summarized the comment as follows: "One of the commenters felt that many of the messages in the public archive address important issues (in particular, Eric Hanson's revision to checkpoint 1.3) "They also felt that animation should be included in many places in the document to accompany "movie" e.g. in Guideline 1, checkpoint 1.3. They thought perhaps, "video and animation" would be better terms. Reason for the suggestion, is that they are referred to as video and animation in multimedia literature." To review Eric Hansen's suggested revisions, see: http://lists.w3.org/Archives/Public/w3c-wai-gl/1999AprJun/0042.html PART A: COMMENTS AND SUGGESTED GLOSSARY EXPLANATION The 4/26/99 version of checkpoint 1.3 provided by G. Vanderheiden reads: "LATEST SUGGESTED CHECKPOINT 1.3 "1.3 Until most audio-visual player technologies can use speech synthesis to generate an auditory description of the video track from a text equivalent (see checkpoint 1.1), provide an auditory description of the video track (synchronized with the audio track). This also applies to animations that are accompanied by sound or time based user interaction. [Priority 1] "[Rationale: there is no need for an audio description of a silent animation unless it is tied to time based interaction with the page (e.g. they need to do something when a particular visual event occurs) in which case it is important to know when something is happening visually. ]" I think that a few issues should be noted: 1. This definition of "auditory description" differs from the definition found in the glossary. The passage indicates that an audio description can be synthesized. Yet the glossary (4/16/99 document) reads: "One example of a non-text equivalent is an auditory description, pre-recorded speech that describes the key visual elements of a presentation and is synchronized with the audio track of the presentation (e.g., during natural pauses in the audio track). Auditory descriptions include information about actions, body language, graphics, and scene changes." My suggested revision to it is as follows: {EH: Revised, 4/27/99:}One example of a non-text equivalent is an _auditory description_, which is an auditory equivalent of a visual track of a video or animation, or multimedia presentation. To function properly, an audio description is usually synchronized with the audio track. (Information is spoken during the natural pauses in the audio track.) Auditory descriptions may include, for example, information about actions, body language, graphics, and scene changes. Following are my assumptions: I assume that an auditory description is always auditory but that it does not matter what _kind_ of audio (synthesized speech OR prerecorded audio OR real-time audio OR can it be a mix of all). I assume that an auditory description should ordinarily be synchronized with the auditory track. Yet it can be called an audio description even when not synchronized. I assume that the text version of the audio description is simply the text equivalent of the video track. A text equivalent of the visual track is NOT an auditory equivalent although, through synthesized speech, a text equivalent can produce an auditory equivalent (e.g., an audio description). 2. Gregg Vanderheiden's definition treats animations differently from video. I am assuming that the requirements that apply to videos also apply to animations. The term animation, in my mind could be something a simple as an animated gif or as rich and extensive that it is essentially indistinguishable from a video. Thus, the terms "video" and "animation" have some overlap in their meaning. Both "video and animation" presentations can include audio as well as graphical content. PART B: MY SUGGESTION FOR CHECKPOINT 1.3 My suggested wording for checkpoint 1.3 is as follows: "1.3 Until most audio-visual player technologies can use speech synthesis to generate an auditory description from a text equivalent of the visual track (see checkpoint 1.1), provide an auditory description of the video track for each audio-visual presentation (video or animation, or multimedia presentation) and synchronize it with the audio track. This does not apply to audio-visual presentations that are silent, unless the presentation requires a time-based user interaction [Priority 2] " For comparison, the old (4/26/99) version is: "1.3 Until most audio-visual player technologies can use speech synthesis to generate an auditory description of the video track from a text equivalent (see checkpoint 1.1), provide an auditory description of the video track (synchronized with the audio track). This also applies to animations that are accompanied by sound or time based user interaction. [Priority 2] {EH: New. 4/27/99} Changes to Checkpoint 1.1. There may need to be a change in checkpoint 1.1 (e.g., "video and animation" instead of "video" AND "animation". I assume that a "multimedia presentation" simply composed of the other elements.). I invite others to see if they think changes are needed there. PART C: MY SUGGESTION FOR CHECKPOINT 1.4 Revisions to Checkpoint 1.4 Note below that I have added reference to audio, since it, too, is a time-based presentation. I also added reference to text transcript of audio. Without these examples, checkpoint 1.4 only seems to cover the same ground as checkpoint 1.3. Priority 2 seems correct. Failure to synchronize does not make access impossible. One question I have is whether synchronization between standalone audio (i.e., without video) and the text transcript is important enough and intended. {EH: 4/27/99, 2:50 hrs.} 1.4 For any time-based presentation (e.g., video or animation, audio {EH:4/27/99. Added}, or multimedia presentation), synchronize equivalent alternatives {EH: "alternatives" may be OK here.}(e.g., captions, auditory descriptions, text transcript of stand-alone audio{EH:4/27/99. Added}) with the presentation. [Priority 2] Techniques for checkpoint 1.4 {EH: New 4/26/99:} Note. Logical sequence (e.g., reading order) for text equivalents is Priority 1. (See note X for checkpoint 1.1) {EH: Old from 4/16/99: 1.4 For any time-based presentation (e.g., a movie, animation, or multimedia presentation), synchronize equivalent alternatives (e.g., captions or auditory descriptions) with the presentation. [Priority 1] PART D: OTHER ISSUES Synchronization is an important accessibility issue (e.g., synchronizing captions and video for people who are deaf; synchronizing audio descriptions and audio for people who are blind) and deserves to have a Priority 2 rating. As others have stated the document sets standards for accessibility. It does not prescribe an implementation schedule for Web content developers. ============================= Eric G. Hansen, Ph.D. Development Scientist Educational Testing Service ETS 12-R Rosedale Road Princeton, NJ 08541 (W) 609-734-5615 (Fax) 609-734-1090 E-mail: ehansen@ets.org
Received on Tuesday, 27 April 1999 03:59:44 UTC