- From: Mandyam, Giridhar <mandyam@quicinc.com>
- Date: Tue, 1 Oct 2013 22:28:35 +0000
- To: cowwoc <cowwoc@bbs.darktech.org>, "public-media-capture@w3.org" <public-media-capture@w3.org>
No, not from my perspective. 'zoom' would be a camera-specific constraint that would be invoked as part of the constructor of a VideoStreamTrack (see http://www.w3.org/TR/mediacapture-streams/#idl-def-VideoStreamTrack). It is separately defined from WebRTC. -----Original Message----- From: cowwoc [mailto:cowwoc@bbs.darktech.org] Sent: Tuesday, October 01, 2013 9:15 AM To: public-media-capture@w3.org Subject: Re: [Bug 23220] Add 'zoom' constraint to VideoStreamTrack Sorry to come in on the tail of this conversation... Out of curiosity, when you talk about "zoom", are you referring to accessing camera-specific functionality such as zoom/pan/etc from within WebRTC? Or are we talking about something else? Thanks, Gili On 01/10/2013 5:51 AM, Stefan Håkansson LK wrote: > On 2013-09-30 02:37, Rob Manson wrote: >> I drew this diagram recently in an attempt to put together a clear >> picture of how all the flows work/could work together. It seems to >> resonate with quite a few people I've discussed it with. >> >> NOTE: This was clearly from the perspective of the post-processing >> pipeline scenarios we were exploring. >> Code: >> https://github.com/buildar/getting_started_with_webrtc#image_processi >> ng_pipelinehtml >> Slides: >> http://www.slideshare.net/robman/mediastream-processing-pipelines-on- >> the-augmented-web-platform >> >> But the question marks in this diagram really relate directly to this >> discussion (at least in the write direction). >> >> It would be very useful in many scenarios to be able to >> programatically generate a Stream (just like you can with the Web >> Audio API). Plus all the other scenarios this type of overall >> integration would unlock (zoom, etc). >> >> And then we would be able to use this Stream as the basis of a MediaStream. >> >> And then we would be able to pass this MediaStream to a >> PeerConnection like normal. >> >> There's a lot of discussion in many different threads/working groups >> and amongst many of the different browser developers about exactly >> this type of overall Stream based architecture...and the broader implications of this. >> >> I think this overall integration is a key issue that could gain good >> support and really needs to be looked into. I'd be happy to put >> together a more detailed overview of this model and gather >> recommendations if there's no strong objections? >> >> It seems like this would most sensibly fit under the Media Capture >> TF...is this correct home for this? > If you look at the charter [1], the fit is not that good. The WebRTC > WG charter [2] mentions "API functions for encoding and other > processing of those media streams" - which could be interpreted to fit. > > On the other hand the WebRTC WG has a lot on its plate already and > needs to focus. > > A third option could be to form some new group - community or working > - for this work. > > Perhaps the best thing to do is to use this list to probe the interest > (e.g. by putting together the detailed overview you mention), and if > there is interest we can discuss where to take the work. > > Stefan > > [1] > http://lists.w3.org/Archives/Public/public-media-capture/2013Feb/0012. > html [2] http://www.w3.org/2011/04/webrtc-charter.html > >> PS: I'm about to present this pipeline work as part of a workshop at >> the IEEE ISMAR in Adelaide so I'll share the slides once they're up. >> >> roBman >> >> >> On 30/09/13 09:59, Silvia Pfeiffer wrote: >>> On Mon, Sep 30, 2013 at 8:17 AM, Harald Alvestrand <harald@alvestrand.no> wrote: >>>> On 09/29/2013 06:39 AM, Silvia Pfeiffer wrote: >>>>> FWIW, I'd like to have the capability to zoom into a specific part >>>>> of a screen share and only transport that part of the screen over >>>>> to the peer. >>>>> >>>>> I'm not fuzzed if it's provided in version 1.0 or later, but I'd >>>>> definitely like to see such functionality supported eventually. >>>> I think you can build that (out of strings and duct tape, kind of) >>>> by screencasting onto a canvas, and then re-screencasting a part of >>>> that canvas. I'm sure there's missing pieces in that pipeline, >>>> though - I don't think we have defined a way to create a >>>> MediaStreamTrack from a canvas yet. >>> Interesting idea. >>> >>> It would be possible to pipe the canvas back into a video element, >>> but I'm not sure if we can as yet take video element content and add >>> it to a MediaStream. >>> >>>> It takes more than a zoom constraint to do it, however; at minimum >>>> you need 2 coordinate pairs (for a rectangular viewport). >>> True. I also think we haven't standardised screensharing yet, even >>> if there is an implementation in Chrome. >>> >>> Silvia. >>> >>> >
Received on Tuesday, 1 October 2013 22:29:03 UTC