- From: Jeremy Doig <jeremydo@google.com>
- Date: Thu, 7 Aug 2008 10:59:18 -0700
how would this work (say) for different avc profile levels and features (eg: PAFF support) ?would we require video creators to know the specific capabilities of every fourCC target ? On Thu, Aug 7, 2008 at 4:23 AM, Tim Starling <tstarling at wikimedia.org>wrote: > Mikko Rantalainen wrote: > > Tim Starling wrote: > > > >> Henri Sivonen wrote: > >> > >>> On Aug 7, 2008, at 09:53, Tim Starling wrote: > >>> > >>> > >>>> xiphQtVersion = videoElt.GetComponentVersion('imdc','XiTh', > >>>> 'Xiph'); > >>>> > >>> This kind of FourCC use is exactly the kind of thing I meant earlier > >>> when I asked if the MIME stuff is really the best match for frameworks. > >>> > >> FourCC and MIME are not mutually exclusive, see RFC 2361. > >> > > > > RFC 2361 doesn't seem to provide a method for describing both video > > codec and audio codec for a resource. The same parameter ("codec") is > > used for both audio and video codec information but I cannot understand > > how a resource that contains both video and audio should be labeled. > > > > RFC 4281 seems to provide a parameter "codecs" that is a comma separated > > list of used codecs. The detail required ("MUST") for the "codecs" > > parameter seems quite high and I'm afraid that this parameter will not > > be correctly used in real world, unfortunately. In theory, this would > > seem a good way to provide the information needed for a resource. > > > > > > Well, I was thinking of an interface which would list the codecs, rather > than the overall file types. But it could be done either way. By file type: > > if ( video.supportsType( 'application/ogg;codecs="vorbis,theora"' ) ) { > ... > } > > By codec/container: > > if ( 'application/ogg' in video.supportedTypes > && 'video/theora' in video.supportedTypes > && 'audio/vorbis' in video.supportedTypes ) > { > ... > } > > The first one looks easier in this application, and allows for multiple > backends with different allowed combinations of container and codec. The > second one allows for complete enumeration of client capabilities. > > Obviously you couldn't provide an interface to enumerate every possible > combination of container, audio codec and video codec. > > The reason this is needed, as opposed to using multiple <source> tags, > is because inevitably, some clients will support certain formats via > <object> (or in our special case, <applet>) and not via <video>. > Querying the browser's <video> capabilities allows JS to decide what > sort of embedding method to use. It's an essential migration feature. > > -- Tim Starling > -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://lists.whatwg.org/pipermail/whatwg-whatwg.org/attachments/20080807/2e29d11e/attachment.htm>
Received on Thursday, 7 August 2008 10:59:18 UTC