W3C home > Mailing lists > Public > public-webrtc@w3.org > October 2011

Re: Track Definition - Was: Re: MediaStreamTrack disabling - effect on downstream MediaStreams?

From: Justin Uberti <juberti@google.com>
Date: Wed, 19 Oct 2011 22:01:55 -0400
Message-ID: <CAOJ7v-04MSMf03DYFi8gJuKvk6owUG20ST0Mhur2yS9fmva+MQ@mail.gmail.com>
To: Cullen Jennings <fluffy@cisco.com>
Cc: public-webrtc@w3.org
On Wed, Oct 19, 2011 at 5:57 PM, Cullen Jennings <fluffy@cisco.com> wrote:

>
> On Oct 5, 2011, at 16:44 , Justin Uberti wrote:
>
> >
> >
> > On Wed, Oct 5, 2011 at 2:07 PM, Cullen Jennings <fluffy@cisco.com>
> wrote:
> >
> > On Sep 26, 2011, at 11:03 AM, Adam Bergkvist wrote:
> >
> > >
> > > Proposed change 2: The MediaStream constructor should take a list of
> MediaStreamTracks, which makes it possible to combine tracks from several
> streams into a new MediaStream (see older mails for use case). Status: Some
> support, some think it may be more powerful than what we need. No real
> objections.
> >
> > I'm not sure we agree what a track is and without a clear answer to that,
> I find it hard to have a position on this. If other think it is clear what a
> track is, perhaps I just need it explained to me.
> >
> > I will propose what I think a track should be ... I think it should be
> the smallest unit stream which can not be further subdivide in it's on the
> wire representation. So effectively this means I think it should represent
> the media flowing in or out of a single codec. So if you you have an audio
> codec that codes the stereo as one set of data and uses joint coding - they
> I view that as one track. But if you you separately encode the right and
> left by passing them through two different codecs and one could be decoded
> without the other, then I would view that as two tracks. Similarly by this
> definition, if you were sending a DTMF and G.711 media, that would be two
> tracks. I'm not sure if there is widespread agreement on this or not but if
> people think tracks are something different than this, it would be great to
> hear what they should be.
> >
> > My understanding is slightly different; I'm looking at a track as an RTP
> source. The main difference from your definition is that I would not see
> DTMF and G.711 as separate media. If DTMF is sent out-of-band using
> telephone-event, or in-band using G.711, that distinction shouldn't be
> evident to the higher layers. Similarly, if a client chooses to switch
> codecs in mid-stream, that wouldn't create a new track IMO. If stereo is
> sent as "left mic source" + "right mic source" with the same CNAME, that
> would result in two tracks in a single MediaStream. But if stereo was sent
> as a single "stereo mic source" with stereo coding, that would be a single
> track.
> >
>
> ok - just trying to get my head around all this.  So would a video and
> audio from the same camera be one track ? It seems like they are the same
> RTP source.
>

If they could be coded as a single interleaved format, I suppose they could
be the same source. In practice, such a camera will generate two independent
streams, with separate codecs, and will result in two different RTP sources
(and hence tracks).


>
>
> > In short, I see that a track indicates a single media source with a
> single clock and a single identifier (SSRC) that allows clear identification
> of which packets belong to which track.
> >
> > I don't know if this is a good definition or not but really what it gets
> down to is I'm a bit unclear on what all these things are.
> >
> >
> >
> >
> >
>
>
Received on Thursday, 20 October 2011 02:02:49 UTC

This archive was generated by hypermail 2.3.1 : Monday, 23 October 2017 15:19:25 UTC