- From: Madeleine Rothberg <Madeleine_Rothberg@wgbh.org>
- Date: 2 Dec 1999 11:12:39 -0500
- To: "W3C-WAI-UA" <w3c-wai-ua@w3.org>
I previously submitted these techniques with all the wrong checkpoint numbers. Here they are, revised to match the last call draft. Also, and more exciting, the SYMM WG has in the meantime released a public working draft of SMIL-Boston with significant enhancements for access. I am including techniques that reference that document, even though it is a draft, because the changes for access seem to be stable. For example, a test attribute for audio description has been added that parallels the one for captioning. I expect this will survive to become part of the final reccomendation. -Madeleine 2.1 Ensure that the user has access to all content, including alternative representations of content. [Priority 1] Techniques for checkpoint 2.1: For SMIL, see specific techniques for Checkpoints 2.2, 2.4, and 2.5. 2.2 If a technology allows for more than one continuous equivalent tracks (e.g., closed captions, auditory descriptions, video of sign language, etc.), allow the user to choose from among the tracks. [Priority 1] Techniques for checkpoint 2.2: Provide an interface which displays all available tracks, with as much identifying information as the author has provided, and allow users to choose which tracks are rendered. For example, if the author has provided "alt" or "title" for various tracks, use that information to construct the list of tracks. Provide an interface which allows users to indicate their preferred language separately for each kind of continuous equivalent. Users with disabilities may need to choose the language they are most familiar with in order to understand a presentation which may not include all equivalent tracks in all desired languages. In addition, international users may prefer to hear the program audio in its original language while reading captions in their first language, fulfulling the function of subtitles or to improve foreign language comprehension. In classrooms, teachers may wish to control the language of various multimedia elements to achieve specific eduational goals. 2.4 Provide time-independent access to time-dependent active elements or allow the user to control the timing of changes. [Priority 1] Techniques for checkpoint 2.4: Provide access to a static list of time dependent links, including information about the context of the link. For example, provide the time at which the link appeared along with a way to easily jump to that portion of the presentation. Provide easy-to-use controls (including both mouse and keyboard commands) to allow viewers to pause the presentation and advance and rewind by small and large time increments. Provide a mode in which all active elements are highlighted in some way and can be navigated sequentially. For example, use a status bar to indicate the presence of active elements and allow the user to navigate among them with the keyboard or mouse to identify each element when the presentation is moving and when it is paused. 2.5 Allow the user to specify that continuous equivalent tracks (e.g., closed captions, auditory descriptions, video of sign language, etc.) be rendered at the same time as audio and video tracks. [Priority 1] Techniques for checkpoint 2.5: It is important that any continuous equivalent tracks be rendered synchronously with the primary content. This ensures that users with disabilities can use the primary and equivalent content in combination. For example, if a hard-of-hearing user is watching a video and reading captions, it is important for the captions to be in sync with the audio so that the viewer can use any residual hearing. For audio description, it is crucial that the primary audio track and the audio description track be kept in sync to avoid having them both play at once, which would reduce the clarity of the presentation. SMIL User agents which play SMIL presentations should take advantage of a variety of access features defined in SMIL. A W3C note on access features of SMIL 1.0 documents those features currently recommended [reference http://www.w3.org/TR/SMIL-access/]. A future version of SMIL (known currently as SMIL Boston) is in development and additional access features may be available when this specification becomes a W3C Recommendation. The following techniques reference features in SMIL 1.0 as well as features included in the November 15, 1999 public working draft of SMIL Boston [reference http://www.w3.org/TR/smil-boston/]. As defined in SMIL 1.0, SMIL players should allow users to turn closed captions on and off by implementing the test attribute system-captions which takes the values "on" and "off." For example, include in the player preferences a way for users to indicate that they wish to view captions, when available. SMIL files with captions available should use the following syntax: <textstream alt="English captions for My Favorite Movie" system-captions="on" src="closed-caps.txt"/> In this case, when the user has requested captions, this textstream should be rendered, and when they have not it should not be rendered. (Note that the use of hyphens has been deprecated in SMIL Boston. This test attribute will be known in upcoming recommendations as systemCaptions.) SMIL 1.0 does not provide a test attribute to control audio description in the same way as captions. However, a test attribute has been proposed for SMIL Boston to meet this need. User agents should consider implementing systemAudioDesc or be aware that when SMIL Boston becomes a W3C Recommendation they will be urged to implement it then. A test attribute to turn audio description on and off should be implemented in parallel to the implementation of the systemCaptions attribute. Users should be able to indicate the preference to receive audio description, when content authors make it available, through the standard preferences-setting section of the UA's user interface. Another test attribute, systemOverdubOrSubtitle (system-overdub-or-captions in SMIL 1.0), allows the user to choose between alternate language text or sound. This attribute specifies whether subtitles or overdub should be rendered for people who are watching a presentation where the audio may be in a language in which they are not fluent. This attribute can have two values: "overdub", which selects for substitution of one voice track for another, and "subtitle", which means that the user prefers the display of subtitles. However, this attribute should not be used to determine if users need captions. When both are available, deaf users will prefer to view captions, which contain additional information on music, sound effects, and who is speaking which are not included in subtitles since those are intended for hearing people. QuickTime User agents which play QuickTime movies should provide the user with a way to turn on and off the different tracks embedded in the movie. Authors may use these alternate tracks to provide synchronized equivalents for use by viewers with disabilities. The Apple QuickTime player currently provides this feature through the menu item "Enable Tracks." Microsoft Windows Media Object User agents which play Microsoft Windows Media Object presentations should provide support for Synchronized Accessible Media Interchange (SAMI), a protocol for creating and displaying caption text synchronized with a multimedia presentation. Users should be given a way to indicate their preference for viewing captions. In addition, user agents which play Microsoft Windows Media Object presentations should enable viewers to turn on and off other alternative equivalents, including audio description and alternate video tracks. Other Formats Other video or animation formats should incorporate similar features. At a minimum, users who are blind and users who are deaf need to be able to turn on and off audio description and captions. General Comments The interface to set these preferences must be accessible. Information on how to author accessible tracks must be included in documentation about how to author for the media player. 2.6 If a technology allows for more than one audio track, allow the user to choose from among tracks. [Priority 1] Techniques for checkpoint 2.6: See techniques for checkpoint 2.2. Madeleine Rothberg The CPB/WGBH National Center for Accessible Media madeleine_rothberg@wgbh.org
Received on Thursday, 2 December 1999 11:09:16 UTC