- From: Stefan Håkansson LK <stefan.lk.hakansson@ericsson.com>
- Date: Wed, 28 Sep 2011 15:14:34 +0200
- To: "public-audio@w3.org" <public-audio@w3.org>
Dear Audio WG,
in an earlier chat between the webrtc and the audio chairs, it was
decided that the audio wg should be invited to the next telco of the
webrtc wg. That telco will take place next week (Wed). Details are
available here:
<http://lists.w3.org/Archives/Public/public-webrtc/2011Sep/0099.html>.
To give some background:
In webrtc/rtcweb we have the following use case and requirement
document:
<http://datatracker.ietf.org/doc/draft-ietf-rtcweb-use-cases-and-requirements/?include_text=1>.
I think the use cases
4.2.7. Multiparty video communication . . . . . . . . . . . . 7
4.2.8. Multiparty on-line game with voice communication . . . 8
are most relevant to the Audio WG. These requirements are derived from them:
F13 The browser MUST be able to pan, mix and render
several concurrent audio streams.
----------------------------------------------------------------
F15 The browser MUST be able to process and mix
sound objects (media that is retrieved from another
source than the established media stream(s) with the
peer(s) with audio streams).
There are API requirements as well:
A14 The Web API MUST provide means for the web
application to control panning, mixing and
other processing for streams.
We're also about to add a requirement on determining the level/activity
in audio streams (useful for speaker indication, level corrections,
detecting noise sources).
To me this sounds a lot like Audio WG territory!
In the current API proposal
(<http://dev.w3.org/2011/webrtc/editor/webrtc.html>) there is something
called "MediaStream", and if nothing changes this is the kind of object
that we would like to be able apply mixing, spatialization,
level/activity measurement/setting to.
Stefan (for Harald and Stefan, chairs of the webrtc WG)
Received on Wednesday, 28 September 2011 13:15:16 UTC