- From: David Singer <singer@apple.com>
- Date: Fri, 06 May 2011 16:50:26 -0700
- To: Silvia Pfeiffer <silviapfeiffer1@gmail.com>
- Cc: Mark Watson <watsonm@netflix.com>, HTML Accessibility Task Force <public-html-a11y@w3.org>
On May 6, 2011, at 16:25 , Silvia Pfeiffer wrote: > On Sat, May 7, 2011 at 9:16 AM, Mark Watson <watsonm@netflix.com> wrote: >> >> On May 6, 2011, at 2:18 PM, Silvia Pfeiffer wrote: >> >> On Sat, May 7, 2011 at 5:22 AM, Mark Watson <watsonm@netflix.com> wrote: >> >> I also have a procedural question: do we consider that we have received the >> >> liaison from 3GPP (mentioned on the above page) ? Are we going to answer it >> >> ? >> >> There should be an email by a 3GPP member with all the relevant >> requests to the WG list. I haven't seen one such. >> >> I doubt the 3GPP staff will post to the list. They will send the liaison to >> the chair of the group, possibly via W3C staff members, and expect them to >> distribute it to the list. I imagine it is stuck somewhere in that path. >> >> Unless that arrives, >> there is nothing to reply to. >> >> We can see that it has been sent from their minutes and we can see the text >> of the document in their archives, so we can at least prepare for its >> arrival ;-) > > > Do poke the chairs to forward it, or ask them to reply to it. It > doesn't seem right for us to just go ahead and try to represent the > HTML WG in a reply to the 3GPP WG. I just poked Philippe as I suggested it be sent to him for him to send on. > > >> One question they ask is whether we will define a URN to identify the >> >> space of kind values defined by W3C. One advantage of doing that is that >> >> these kinds are then immediately supported in 3GPP and MPEG adaptive >> >> streaming manifests, which means that there *is* a media container >> >> supporting those kinds (perhaps addressing one of the editor's concerns >> >> about new kind values). >> >> I don't think HTML needs these values as anything else but short >> strings that we now have. >> >> Agreed. >> >> If 3GPP and MPEG need them as URNs >> >> No, they need a URN that refers to the codepoint space containing these >> values. But I notice W3C doesn't appear to have a top-level URN space. It >> could use urn:fdc:w3.org:2011.... > > Seems to make sense to me... would such a URN be useful for anything else? > > >> The 3GPP and DASH specs allow you to tag a track with any number of "Roles" >> - so it can be tagged with roles from this W3C space (if anyone defines a >> URN) as well as equivalent roles from other spaces if necessary for some >> application. >> >> - and they probably have a swag >> more that they have proposed and will use >> >> Not really - they'd prefer W3C to define some container-format-independent >> ones, as I understand it. >> >> - it makes a lot more sense >> to me for them to define these themselves. >> >> They seem to think the opposite ;-) In particular it's been stated at MPEG >> that the W3C HTML a11y group has more a11y expertise than the MPEG group. >> And there is nothing container-format-specific about this concept of track >> kinds. The abstract kinds and their definitions need to be defined somewhere >> with responsibility for all containers or for none, not in the groups >> focussed on particular containers. The containers (and HTML) just need to >> define how those kinds are labeled in their syntax. > > Well, MPEG caters for a lot more than just the Web and the kinds > required elsewhere may need to be a lot more diverse. So, I agree that > they do well in looking to us for a11y related kinds and it's great > that they want to pick up the ones that we define here, but I highly > doubt that will be the end of their list. > > >> Some of the kinds that >> media containers will expose may just end up creating getLabel() text >> rather than be exposed in getKind(). Ogg already seems to have a few >> of those. >> >> Agreed, but I think this is a symptom of having media container formats lead >> the definition of kinds: they define kinds which are either not very >> well-defined or not universally applicable and so we decide not to expose >> these over HTML because we want the HTML interface to be clean and >> well-defined. The best way to achieve that is to define these things in a >> container-independent place and ask the containers to align. > > > Well, we don't need anyone to align. All we need is a mapping, which > your wiki page provides perfectly. If the names are identical, the > better so. It's ok when not everything that a container can contain is > exposed to HTML. Only what gets exposed needs to be compatible. I > think we're on the best way there anyway. :-) > > Cheers, > Silvia. > > >> >> Finally, we discussed the "commentary" kind here at Netflix and in the end >> >> we are happy to have it dealt with simply as "alternative". I do think >> >> though that in principle there could be other (UI-related) reasons for >> >> exposing a new track kind than triggering default behavior or application of >> >> user preferences. This is certainly the case for accessibility use-cases >> >> where the UI to enable/disable a particular track could usefully be tailored >> >> to the intended users of that track (for example, enabling/disabling tracks >> >> intended for the blind or those with low vision should ideally not involve >> >> complex visual UI elements). >> >> OK. I think we'll cross that bridge when we get to it. >> >> Cheers, >> Silvia. >> >> >> ...Mark >> >> >> On May 3, 2011, at 10:12 PM, Silvia Pfeiffer wrote: >> >> I understand the problem of additive/alternative tracks, too, and have >> >> tried to approach it with markup before. However, I think this is >> >> making something that is supposedly simple much too difficult. The >> >> ultimate choice of active tracks has got to be left to the user. For >> >> this reason, I think @kind (or getKind()) should only ever expose what >> >> content is available in the track, but there should not be an >> >> automatic choice made by the browser. It's up to the user to >> >> activate/deactivate the correct tracks. >> >> Before we dive into anything more complex, we should get some >> >> experience with an implementation of multitrack and the roles. I don't >> >> think we will have much to go by for making a decision beforehand. >> >> Cheers, >> >> Silvia. >> >> >> On Wed, May 4, 2011 at 2:02 PM, Mark Watson <watsonm@netflix.com> wrote: >> >> So, if we are looking for a generic approach, where a track can have >> >> multiple "roles", then I think the correct logic is indeed to pick the >> >> fewest number of tracks which fulfill the intersection of the desired roles >> >> and the available roles such that no role is fulfilled more than once. You >> >> need a priority list of roles to drop from the desired list if that isn't >> >> possible (which would mean some badly authored content, but has to be dealt >> >> with). It may be a mouthful, but I think it would be reasonably >> >> straightforward to implement. >> >> However, I'm still not sure a generic approach is necessary. A "simpler" >> >> approach is to say every track has a single role. But for some applications >> >> (like audio descriptions) there are two distinct role values defined - an >> >> additive one and an alternative one. The problem is addressed at a semantic >> >> level - i.e. people implement support for audio descriptions - and they know >> >> what these are and how to handle them - rather than trying for a generic >> >> descriptor matching algorithm. >> >> Regarding Repetitive Stimulus Safe, I guess that since most content is >> >> unfortunately not labeled one way or the other the default assumption has to >> >> be up to the user themselves. i.e. that user preferences associated with >> >> this aspect should support required, preferred and don't care. In a really >> >> generic approach every role may have a status from { require, prefer, don't >> >> care, prefer not, require not }. >> >> Again, this suggests that a generic approach might be over-ambitious - who >> >> says some new role doesn't come along next week with a sixth user-preference >> >> status of "required unless role Y present" or similar ... I think maybe the >> >> UA needs to understand what these things are and act appropriately. >> >> ...Mark >> >> >> >> >> >> On May 3, 2011, at 11:55 AM, David Singer wrote: >> >> >> On May 2, 2011, at 16:55 , Mark Watson wrote: >> >> >> I think it's evidence that there is something to be solved. >> >> I'd prefer a solution where adding a track to an existing presentation >> >> didn't require me to change the properties of existing tracks, though, since >> >> there is an error waiting to happen in that case. >> >> >> Yes. This idea made some sense when it was the tracks in a multiplex (e.g. >> >> MP4 file), perhaps makes sense when all the tracks are annotated in the >> >> markup (e.g. in HTML5 or DASH MPD) but makes much less sense when some >> >> tracks are in a multiplex and some are added in the markup - a track added >> >> in the markup might need the annotations in a multiplex changed, ugh. >> >> So, thinking out loud here. >> >> Assume the user has a set of roles that they would kinda like to experience. >> >> The default is 'main, supplementary', I think, or something like. >> >> Now, we have a set of tracks, each of which satisfies some roles. Let's >> >> ignore tracks we have discarded because they are the wrong mime type, codec, >> >> language, etc., and focus just on this selection mechanism. What is the >> >> right simple way to get the set of tracks? >> >> It's easy to 'go overboard' and treat this as a very general problem of >> >> finding the minimal set of tracks that will span a set of design roles. I >> >> don't think anyone will author *for the same language* >> >> track - main >> >> track - captions >> >> track - main + captions >> >> so an algorithm designed to pick only (3) instead of (1 + 2) for the >> >> main+captions desiring user is probably overkill. >> >> 'enable the tracks whose roles are a subset of the desired roles, and >> >> disable the rest' may be too simple, unless tracks are ordered from the >> >> most-labelled to the least-labelled. >> >> So, audio-description replacing the main audio: >> >> track - main description >> >> track - main >> >> Audio description adding to the main audio >> >> track - main >> >> track - description >> >> The same works for all the adaptations that might require re-authoring or >> >> might be achievable with an additional track (captions, burned in or >> >> separate, for example). >> >> >> Where this fails is when the 'base content' is good enough for both the >> >> plain user and the user who desires more roles. The obvious case here (Mark >> >> will laugh) is repetitive-stimulus-safeness; we have to assume unlabelled >> >> content is unsafe, but much content is naturally safe and can be labelled as >> >> such. >> >> David Singer >> >> Multimedia and Software Standards, Apple Inc. >> >> >> >> >> >> >> >> >> >> > David Singer Multimedia and Software Standards, Apple Inc.
Received on Friday, 6 May 2011 23:50:55 UTC