W3C home > Mailing lists > Public > public-html-a11y@w3.org > November 2010

Re: Categorization of media a11y requirements

From: Silvia Pfeiffer <silviapfeiffer1@gmail.com>
Date: Fri, 5 Nov 2010 13:11:25 +1100
Message-ID: <AANLkTin9rLWav9itpJ_5AZO8yG3gnm7XQS-1Fcb-4OWG@mail.gmail.com>
To: Geoff Freed <geoff_freed@wgbh.org>
Cc: Frank Olivier <Frank.Olivier@microsoft.com>, "public-html@w3.org" <public-html@w3.org>, "public-html-a11y@w3.org" <public-html-a11y@w3.org>
I actually used TTEXT (notice the 'E') and I first thought about TEXT,
but it needed something with time in it. How about CUE?

Silvia.

On Fri, Nov 5, 2010 at 12:56 PM, Geoff Freed <geoff_freed@wgbh.org> wrote:
>
> Hi, Frank and Silvia:
>
> I actually suggest that you *not* call it TTXT because, ironically, TTXT is
> already in use as the name of another timed-text format:
>  http://gpac.sourceforge.net/doc_ttxt.php#ttxt .  Maybe call it TRACKDETAIL?
>
> Geoff/NCAM
>
>
>
> On 11/4/10 5:19 PM, "Silvia Pfeiffer" <silviapfeiffer1@gmail.com> wrote:
>
> Maybe this classification is more useful in the Checklist [1] than the
> one we have there currently for "types of technologies affected"?
>
> I would, however, suggest calling the Text track format details TTEXT
> and not TRACK, since track is now the keyword for the markup in HTML5.
>
> [1] http://www.w3.org/WAI/PF/HTML/wiki/Media_Accessibility_Checklist
>
> Cheers,
> Silvia.
>
>
> On Fri, Nov 5, 2010 at 3:55 AM, Frank Olivier
> <Frank.Olivier@microsoft.com> wrote:
>>
>> Re http://www.w3.org/2010/11/04-html-wg-minutes.html
>>
>> Members of HTML WG and media a11y reviewed the
>> http://www.w3.org/WAI/PF/HTML/wiki/Media_Accessibility_Requirements
>> requirements
>>
>> We sorted the requirements into the following categories:
>> *       UX: User agent user experience requirement
>> *       SPECNEW: New requirements for the HTML5 specification
>> *       SPECCED: Already in the HTML5 specification
>> *       TRACK: Text track format detail - something that should be specced
>> in the track format, not the HTML spec
>> *       NO: Not an issue we will address in the W3C
>>
>>
>> Of the 119 requirements in the document:
>> SPECNEW 11 items (9% of total)
>> SPECCED 21 (18%)
>> TRACK           24 (20%)
>> NO              4 (3%)
>> UX              73 (61%)
>>
>> Detailed list:
>> NO: (DV-9) Allow the author to use a codec which is optimized for voice
>> only, rather than requiring the same codec as the original soundtrack.Does
>> not seem like a UA issue:
>> NO: (KA-3) The author would be able to choose any/all of the controls,
>> skin them and position them. Needs discussion' kill
>> NO: (KA-5) The scripted and native controls must go through the same
>> platform-level accessibility framework (where it exists), so that a user
>> presented with the scripted version is not shut out from some expected
>> behavior.
>> NO: (PP-1) This necessitates a clear and unambiguous declared format, so
>> that existing authoring tools can be configured to export finished files in
>> the required format.
>> SPECCED, UX: (CA-2) Support the synchronisation of multitrack audio either
>> within the same file or from separate files - preferably both.
>> SPECCED: (API-1) The existence of alternative-content tracks for a media
>> resource must be exposed to the user agent.
>> SPECCED: (API-2) Since authors will need access to the alternative content
>> tracks, the structure needs to be exposed to authors as well, which requires
>> a dynamic interface.
>> SPECCED: (API-3) Accessibility APIs need to gain access to alternative
>> content tracks no matter whether those content tracks come from within a
>> resource or are combined through markup on the page.
>> SPECCED: (CA-1) Support clear audio as a separate, alternative audio track
>> from other audio-based alternative media resources.
>> SPECCED: (CC-22) Support captions that are provided inside media resources
>> as tracks, or in external files.
>> SPECCED: (CC-26) Support multiple tracks of foreign-language subtitles in
>> different languages.
>> SPECCED: (CC-27) Support live-captioning functionality. Addressed via API
>> SPECCED: (CN-4) Support third-party provided structural navigation markup.
>> SPECCED: (CNS-4) Producers and authors may optionally provide additional
>> access options to identified structures, such as direct access to any node
>> in a table of contents. [May be done with cue range]
>> SPECCED: (DAC-4) Synchronized alternatives for time-based media (e.g.,
>> captions, descriptions, sign language) can be rendered at the same time as
>> their associated audio tracks and visual tracks (UAAG 2.0 3.1.3).
>> SPECCED: (DV-3) Support multiple description tracks (e.g., discrete tracks
>> containing different levels of detail).
>> SPECCED: (DV-4) Support recordings of real human speech as a track of the
>> media resource, or as an external file.
>> SPECCED: (KA-3) All functionality available to native controls must also
>> be available to scripted controls.
>> SPECCED: (PP-1) Support existing production practice for alternative
>> content resources, in particular allow for the association of separate
>> alternative content resources to media resources. Browsers cannot support
>> all forms of time-stamp formats out there, just as they cannot support all
>> forms of image formats (etc.).
>> SPECCED: (PP-4) Typically, alternative content resources are created by
>> different entities to the ones that create the media content. They may even
>> be in different countries and not be allowed to re-publish the other one's
>> content. It is important to be able to host these resources separately,
>> associate them together through the Web page author, and eventually play
>> them back synchronously to the user.
>> SPECCED: (SL-4) Support multiple sign-language tracks in several sign
>> languages.
>> SPECCED: (T-1) Support the provisioning of a full text transcript for the
>> media asset in a separate but linked resource. where the linkage is
>> programmatically accessible to AT.
>> SPECNEW, SPECCED: (SL-1) Support sign-language video either as a track as
>> part of a media resource or as an external file.
>> SPECNEW, SPECCED: (SL-2) Support the synchronized playback of the
>> sign-language video with the media resource.
>> SPECNEW, TRACK: (CC-5) Support positioning in all parts of the screen -
>> either inside the media viewport but also possibly in a determined space
>> next to the media viewport. This is particularly important when multiple
>> captions are on screen at the same time and relate to different speakers, or
>> when in-picture text is avoided.
>> SPECNEW, TRACK: (CN-1) Provide a means to structure media resources so
>> that users can navigate them by semantic content structure.
>> SPECNEW, UX: (CC-25) Support edited and verbatim captions, if available.
>> SPECNEW, UX: (DV-8) Allow the author to provide fade and pan controls to
>> be accurately synchronized with the original soundtrack.
>> SPECNEW: (CN-3) Support both global navigation by the larger structural
>> elements of a media work, and also the most localized atomic structures of
>> that work, even though authors may not have marked-up all levels of
>> navigational granularity.
>> SPECNEW: (CN-6) Support direct access to any structural element, possibly
>> through URIs. [Media fragment-like issue]
>> SPECNEW: (CNS-1) All identified structures, including ancillary content as
>> defined in "Content Navigation" above, must be accessible with the use of
>> "next" and "previous," as refined by the granularity control. [May be
>> handled with cue ranges]
>> SPECNEW: (DAC-2) The user has a global option to specify which types of
>> alternative content by default and, in cases where the alternative content
>> has different dimensions than the original content, how the layout/reflow of
>> the document should be handled. (UAAG 2.0 3.1.2). [Probably minimal spec
>> text required: Media queries would work nicely here; also UX issue (user
>> sets media query to match)]
>> SPECNEW: (DAC-5) Non-synchronized alternatives (e.g., short text
>> alternatives, long descriptions) can be rendered as replacements for the
>> original rendered content (UAAG 2.0 3.1.3).
>> TRACK, UX: (CC-16) Use conventions that include inserting left-to-right
>> and right-to-left segments within a vertical run (e.g. Tate-chu-yoko in
>> Japanese), when rendered as text in a top-to-bottom oriented language.
>> TRACK, UX: (CC-19) Present the full range of typographical glyphs, layout
>> and punctuation marks normally associated with the natural language's
>> print-writing system.
>> TRACK, UX: (CC-21) Permit the distinction between different speakers.
>> TRACK, UX: (ECC-2) Support hyperlinks and other activation mechanisms for
>> supplementary data for (sections of) caption text.
>> TRACK, UX: (ECC-3) Support text cues that may be longer than the time
>> available until the next text cue and thus provide overlapping text cues.
>> TRACK, UX: (ECC-4) It needs to be possible to define timed text cues that
>> are allowed to overlap with each other in time and be present on screen at
>> the same time
>> TRACK: (CC-10) Render a background in a range of colors, supporting a full
>> range of opacities.
>> TRACK: (CC-11) Render text in a range of colors.
>> TRACK: (CC-14) Allow the use of mixed display styles-- e.g., mixing
>> paint-on captions with pop-on captions-- within a single caption cue or in
>> the caption stream as a whole.
>> TRACK: (CC-17.1) Represent content of different natural languages. In some
>> cases the inclusion of a few foreign words form part of the original
>> soundtrack, and thus need to be in the same caption resource.
>> TRACK: (CC-18) Represent content of at least those specific natural
>> languages that may be represented with [Unicode 3.2], including common
>> typographical conventions of that language (e.g., through the use of
>> furigana and other forms of ruby text).
>> TRACK: (CC-2) Allow the author to specify erasures, i.e., times when no
>> text is displayed on the screen (no text cues are active).
>> TRACK: (CC-20) Permit in-line mark-up for foreign words or phrases.
>> TRACK: (CC-3) Allow the author to assign timestamps so that one
>> caption/subtitle follows another, with no perceivable gap in between.
>> TRACK: (CC-4) Be available in a text encoding.
>> TRACK: (CC-8) Allow the author to specify line breaks.
>> TRACK: (CC-9) Permit a range of font faces and sizes.
>> TRACK: (CN-2) The navigation track should provide for hierarchical
>> structures with titles for the sections.
>> TRACK: (DV-14) Support metadata, such as copyright information, usage
>> rights, language, etc.
>> TRACK: (ECC-1) Support metadata markup for (sections of) timed text cues.
>> TRACK: (PP-2) Support the association of authoring and rights metadata
>> with alternative content resources, including copyright and usage
>> information. [Move to ATAG?]
>> TRACK: (PP-3) Support the simple replacement of alternative content
>> resources even after publishing.
>> UX, SPECCED: (MD-5) If the user can modify the state or value of a piece
>> of content through the user interface (e.g., by checking a box or editing a
>> text area), the same degree of write access is available programmatically
>> (UAAG 2.0 2.1.5).
>> UX: (CA-3) Support separate volume control of the different audio tracks.
>> UX: (CC-1) Render text in a time-synchronized manner, using the media
>> resource as the timebase master.
>> UX: (CC-12) Enable rendering of text with a thicker outline or a drop
>> shadow to allow for better contrast with the background.
>> UX: (CC-13) Where a background is used, it is preferable to keep the
>> caption background visible even in times where no text is displayed, such
>> that it minimises distraction. However, where captions are infrequent the
>> background should be allowed to disappear to enable the user to see as much
>> of the underlying video as possible.
>> UX: (CC-15) Support positioning such that the lowest line of captions
>> appears at least 1/12 of the total screen height above the bottom of the
>> screen, when rendered as text in a right-to-left or left-to-right language
>> UX: (CC-17.2) Also allow for separate caption files for different
>> languages and on-the-fly switching between them. This is also a requirement
>> for subtitles.
>> UX: (CC-23) Ascertain that captions are displayed in sync with the media
>> resource.
>> UX: (CC-24) Support user activation/deactivation of caption tracks.
>> UX: (CC-6) Support the display of multiple regions of text simultaneously.
>> UX: (CC-7) Display multiple rows of text when rendered as text in a
>> right-to-left or left-to-right language.
>> UX: (CN-10) Support that in bilingual texts both the original and
>> translated texts can appear on screen, with both the original and translated
>> text highlighted, line by line, in sync with the audio narration.
>> UX: (CN-5) Keep all content representations in sync, so that moving to any
>> particular structural element in media content also moves to the
>> corresponding point in all provided alternate media representations
>> (captions, described video, transcripts, etc) associated with that work.
>> UX: (CN-7) Support pausing primary content traversal to provide access to
>> such ancillary content in line.
>> UX: (CN-8) Support skipping of ancillary content in order to not interrupt
>> content flow.
>> UX: (CN-9) Support access to each ancillary content item, including with
>> "next" and "previous" controls, apart from accessing the primary content of
>> the title.
>> UX: (CNS-2) Users must be able to discover, skip, play-in-line, or
>> directly access ancillary content structures.
>> UX: (CNS-3) Users need to be able to access the granularity control using
>> any input mode, e.g. keyboard, speech, pointer, etc.
>> UX: (DAC-1) The user has the ability to have indicators rendered along
>> with rendered elements that have alternative content (e.g., visual icons
>> rendered in proximity of content which has short text alternatives, long
>> descriptions, or captions). In cases where the alternative content has
>> different dimensions than the original content, the user has the option to
>> specify how the layout/reflow of the document should be handled. (UAAG 2.0
>> 3.1.1).
>> UX: (DAC-3) The user can browse the alternatives and switch between them.
>> UX: (DAC-6) Provide the user with the global option to configure a cascade
>> of types of alternatives to render by default, in case a preferred
>> alternative content type is unavailable.
>> UX: (DAC-7) During time-based media playback, the user can determine which
>> tracks are available and select or deselect tracks. These selections may
>> override global default settings for captions, descriptions, etc. (UAAG 2.0
>> 4.9.8)
>> UX: (DAC-8) Provide the user with the option to load time-based media
>> content such that the first frame is displayed (if video), but the content
>> is not played until explicit user request.
>> UX: (DV-1) Provide an indication that descriptions are available, and are
>> active/non-active.
>> UX: (DV-10) Allow the user to select from among different languages of
>> descriptions, if available, even if they are different from the language of
>> the main soundtrack.
>> UX: (DV-11) Support the simultaneous playback of both the described and
>> non-described audio tracks so that one may be directed at separate outputs
>> (e.g., a speaker and headphones).
>> UX: (DV-12) Provide a means to prevent descriptions from carrying over
>> from one program or channel when the user switches to a different program or
>> channel.
>> UX: (DV-13) Allow the user to relocate the description track within the
>> audio field, with the user setting overriding the author setting. The
>> setting should be re-adjustable as the media plays.
>> UX: (DV-2) Render descriptions in a time-synchronized manner, using the
>> media resource as the timebase master.
>> UX: (DV-6) Allow the user to independently adjust the volumes of the audio
>> description and original soundtracks, with the user's settings overriding
>> the author's.
>> UX: (DV-7) Permit smooth changes in volume rather than stepped changes.
>> The degree and speed of volume change should be under provider control.
>> UX: (ECC-5) Allow users to define the reading speed and thus define how
>> long each text cue requires, and whether media playback needs to pause
>> sometimes to let them catch up on their reading.
>> UX: (EVD-1) Support detailed user control as specified in (TVD-4) for
>> extended video descriptions.
>> UX: (EVD-2) Support automatically pausing the video and main audio tracks
>> in order to play a lengthy description.
>> UX: (EVD-3) Support resuming playback of video and main audio tracks when
>> the description is finished.
>> UX: (KA-1) Support operation of all functionality via the keyboard on
>> systems where a keyboard is (or can be) present (Needs better text), and
>> where a unique focus object is employed. This does not forbid and should not
>> discourage providing mouse input or other input methods in addition to
>> keyboard operation. (UAAG 2.0 4.1.1)
>> UX: (KA-2) Support a rich set of native controls for media operation,
>> including but not limited to play, pause, stop, jump to beginning, jump to
>> end, scale player size
>> UX: (KA-4) It must always be possible to enable native controls regardless
>> of the author preference to guarantee that such functionality is available
>> UX: (MD-2) Ensure accessibility of all user-interface components including
>> the user interface, rendered content, and alternative content; make
>> available the name, role, state, value, and description via a
>> platform-accessibility architecture. (UAAG 2.0 2.1.2)
>> UX: (MD-3) If a feature is not supported by the accessibility
>> architecture(s), provide an equivalent feature that does support the
>> accessibility architecture(s). Document the equivalent feature in the
>> conformance claim. (UAAG 2.0 2.1.3)
>> UX: (MD-4) If the user agent implements one or more DOMs, they must be
>> made programmatically available to assistive technologies. (UAAG 2.0 2.1.4)
>> This assumes the video element will write to the DOM.
>> UX: (MD-6) If any of the following properties are supported by the
>> accessibility-platform architecture, make the properties available to the
>> accessibility-platform architecture
>> UX: (MD-7) Ensure that programmatic exchanges between APIs proceed at a
>> rate such that users do not perceive a delay. (UAAG 2.0 2.1.7).
>> UX: (SL-3) Support the display of sign-language video either as
>> picture-in-picture or alpha-blended overlay, as parallel video, or as the
>> main video with the original video as picture-in-picture or alpha-blended
>> overlay.
>> UX: (SL-5) Support the interactive activation/deactivation of a
>> sign-language track by the user.
>> UX: (T-2) Support the provisioning of both scrolling and static display of
>> a full text transcript with the media resource, e.g. in a area next to the
>> video or underneath the video, which is also AT accessible.
>> UX: (TSM-1) The user can adjust the playback rate of the time-based media
>> tracks to between 50% and 250% of real time.
>> UX: (TSM-2) Speech whose playback rate has been adjusted by the user
>> maintains pitch in order to limit degradation of the speech quality.
>> UX: (TSM-3) All provided alternative media tracks remain synchronized
>> across this required range of playback rates.
>> UX: (TSM-4) The user agent provides a function that resets the playback
>> rate to normal (100%).
>> UX: (TSM-5) The user can stop, pause, and resume rendered audio and
>> animation content (including video and animated images) that last three or
>> more seconds at their default playback rate.
>> UX: (TVD-1) Support presentation of text video descriptions through a
>> screen reader or braille device, with playback speed control and voice
>> control and synchronisation points with the video.
>> UX: (TVD-2) TVDs need to be provided in a format that contains the
>> following information: (A) start time, text per description cue (the
>> duration is determined dynamically, though an end time could provide a cut
>> point)
>> UX: (TVD-3) Where possible, provide a text or separate audio track
>> privately to those that need it in a mixed-viewing situation, e.g., through
>> headphones.
>> UX: (TVD-4) Where possible, provide options for authors and users to deal
>> with the overflow case: continue reading, stop reading, and pause the video.
>> (One solution from a user's point of view may be to pause the video and
>> finish reading the TVD, for example.) User preference should override
>> authored option.
>> UX: (TVD-5) Support the control over speech-synthesis playback speed,
>> volume and voice, and provide synchronisation points with the video.
>> UX: (VP-1) It must be possible to deal with three different cases for the
>> relation between the viewport size, the position of media and of alternative
>> content:
>> UX: (VP-2) The user can change the following characteristics of visually
>> rendered text content, overriding those specified by the author or
>> user-agent defaults
>> UX: (VP-3) Provide the user with the ability to adjust the size of the
>> time-based media up to the full height or width of the containing viewport,
>> with the ability to preserve aspect ratio and to adjust the size of the
>> playback viewport to avoid cropping, within the scaling limitations imposed
>> by the media itself.
>> UX: (VP-4) Provide the user with the ability to control the contrast and
>> brightness of the content within the playback viewport.
>> UX: (VP-5) Captions and subtitles traditionally occupy the lower third of
>> the video, where also controls are also usually rendered.
>> UX: [In that this is a user agent issue] (MD-1) Support a
>> platform-accessibility architecture relevant to the operating environment.
>> (UAAG 2.0 2.1.1)
>> UX: (CA-4) Support pre-emphasis filters, pitch-shifting, and other
>> audio-processing algorithms.
>> UX: (DV-5) Allow the author to independently adjust the volumes of the
>> audio description and original soundtracks. [Actually a requirement on the
>> media format]
>>
>>
>>
>>
>
>
>
Received on Friday, 5 November 2010 02:12:32 GMT

This archive was generated by hypermail 2.2.0+W3C-0.50 : Friday, 27 April 2012 04:42:23 GMT