- From: Kazuyuki Ashimura <ashimura@w3.org>
- Date: Tue, 8 Oct 2019 16:20:28 +0900
- To: public-web-and-tv@w3.org
- Message-ID: <CAJ8iq9UNOnA9zaaqJwVXZokV8TH0Jwx=XYLsc1+HBo8axjiBfQ@mail.gmail.com>
available at: https://www.w3.org/2019/09/16-me-minutes.html also as text below. Very sorry for the delay. I thought I had sent out these draft minutes the other day, but it seems the message didn't go out. Thanks, Kazuyuki --- [1]W3C [1] http://www.w3.org/ - DRAFT - Media and Entertainment IG f2f meeting at TPAC 2019 in Fukuoka 16 Sep 2019 [2]group photo from the WoT F2F meeting on 19-20 September 2019 [2] https://www.w3.org/me-f2f/photos/0916/DSC_3346.JPG Group photo from the WoT F2F meeting on 19-20 September 2019 (Some more photos from at the [3]F2F at Hilton) [3] https://www.w3.org/me-f2f/photos/0916/ [4]Agenda [4] https://www.w3.org/2011/webtv/wiki/Face_to_face_meeting_during_TPAC_2019#Agenda_Monday_16_September_2019 Attendees Present Andreas_Tai, Anssi_Kostiainen, Chris_Needham, Daiki_Matsui, David_Fazio, Eric_Siow, Florent_Castelli, Gary_Katsevman, Greg_Freedman, Hiroaki_Shimano, Hiroshi_Fujisawa, John_Riviello, Josh_O_Connor, Kaz_Ashimura, Li_Lin, Mamoru_Takagi, Mark_Watson, Masaya_Ikeo, Masayoshi_Onishi, Nonoka_Jinushi, Pierre_Lemieux, Scott_Low, Shinya_Abe, Sudeep_Divakaran, Tatsuya_Igarashi, Xu_Song, Yongjun_Wu, Youngsun_Ryu, Yuki_Yamakami, Takio_Yamaoka, Akihiko_Koizuka, Keiichi_Suzuki, Michael_Li, Takahiro_Kumekawa, Chris_Cunningham, Taki_Kamiya, Jeff_Jaffe, David_Singer, Glenn_Adams, Hyojin_Song, Francois_Daoust, Ken_Komatsu, Toshiya_Nakakura, Jonathan_Devlin, Amit_Hilbuch, Steve_Anton, Sebastian_Kaebisch, Daniel_Peintner, Nigel_Megitt, Samira_Hirji, Alan_Bird, Mark_Foltz Remote Kazuhiro_Hoya, Lei_Zhai, Mark_Vickers, Yajun_Chen Regrets Chair Chris, Igarashi, Pierre Scribe kaz, tidoust Contents * [5]Topics 1. [6]Welcome and introduction 2. [7]Hybridcast update 3. [8]Media Timed Events in Hybridcast 4. [9]Media Timed Events Task Force 5. [10]CTA WAVE update 6. [11]Review open issues 7. [12]Frame accuracy synchronization (contd) 8. [13]Professional media workflows on the web 9. [14]Bullet Chatting 10. [15]Joint meeting with Second Screen WG/CG 11. [16]Joint meeting with Timed Text WG 12. [17]Closing and wrap-up * [18]Summary of Action Items * [19]Summary of Resolutions __________________________________________________________ Welcome and introduction <inserted> scribenick: kaz Chris: welcome and let's get started ... this is a one-day meeting today ... starting with the general view about the MEIG ... [W3C Code of Ethics] ... [Media&Entertainment IG: Mission] ... apply Web technology in general to media services ... new use cases/requirements to drive the discussion ... [History of Major Initiatives] ... 1. 2011-2014: HTML5 Media Pipeline ... 2. 2011-2017: Adaptive streaming & content protection ... 3. 2017-2019: Media Web App Platform ... [Charter] ... scope covers almost everything ... end-to-end pipeline ... continuous experience ... increasing interactive media ... including games ... [Charter (cont.)] ... Tasks ... identify requirements ... incubation of technical ... review media-related deliverables ... coordinate with other media-related groups, e.g., MPEG, HbbTV, ... ... internationalization, accessibility, etc., are very important ... [Work Flow] ... new ideas & issues from Members and SDOs ... use cases, requirments and gap analysis ... but not specs themselves ... because this is an IG ... but some of the results could promote new features for other WGs ... we're encouraged to work on more and more new features ... [WICG] ... discourse forum there ... get implementer supporter for your idea ... GitHub repo for your proposed feature ... [Contributing to HTML and DOM] ... HTML WG and WHATWG ... [Task Forces] ... 2018-19: Media Timed Events TF ... 2016-17: Cloud Browser API TF (dormant) ... [Monthly conference call] ... 2019/2018 ... (list of topics) ... [Activities for 2020] ... whole bunch of topics for Media WG ... potential candidates for adaptation ... MEIG can input use cases and requirements ... [Activities for 2020 (cont)] ... what will be new topics for 2020? ... what would be the main things? ... would like to capture ideas ... in the after noon, we'll have more open discussion ... [Schedule] ... (shows the agenda) ... any additions? ... [Resources] ... various links here ... btw, we have a new co-Chair, Pierre, here Pierre: involving standard activity, e.g., IETF ... started with HTML WG for W3C ... co-editor of TTML 1.0 ... feel free to contact me and Chris (and Igarashi-san) Chris: would mention that Mark Vickers stepping down as a co-Chair ... has been leading the group successfully ... really getting the Web as the platform for media ... Mark will continue to participate in the MEIG Mark: one of the founding co-Chairs ... plan to stay involved as an Invited Expert ... the best source consolidated expertise is media, video and audio ... before HTML5 media support in the Web ... we've been a good source ... for W3C, WHATWG, Kronos, etc. ... we don't write specs ourselves ... but see what's the priority for media on the Web ... provide expertise ... communicate with media companies, etc., which are not in the W3C as well ... it takes a lot of work ... glad to help ... a lot of leadership in studio world ... and so on ... have three very strong co-Chairs ... and Pierre is joining ... aiming for HTML5 media 2.0 ... this is really a time to better support for media ... Comcast, my company, provides a new rep John: yes, I'm here Hybridcast update [20]slides [20] https://www.w3.org/2011/webtv/wiki/images/d/d3/RecentAchievementHybridcast_TPAC20190916.pdf Ikeo: welcome to Japan! ... would talk about Hybridcast ... [Today's outline] ... recent achievement of hybridcast ... [Deployment status] ... [History of standardization and experiments] ... 2014-2019 ... hybridcast connect deployed on some of the TV sets ... [Shipment of Hybridcast receivers] ... number of receivers over 10 million ... [Trial deployment "Hybridcast-Connect" ... what is hybridcast connect? ... new functions/services uses new APIs ... new APIs are experimentally implemented in some of the TV sets ... brought an example here ... a number of companies are involved ... [Typical Sequence by additional APIs] ... 5 functions ... 1. MediaAvailabilityAPI ... 2. ChannelsInfoAPI ... 3. StartAITAPI ... 4. TaskStatusAPI ... 5. ReceiverStatusAPI ... [Hybridcast Connect demo] ... will show a demo here ... (brings a TV set in front of the screen ... [Hybridcast-Connect demos] ... two demos ... 1 emergency alert ... 2. smooth guidance of catch-up ... [Supporsed use cases demo (1)] ... (Kaz adds webcam to webex) ... embedded buttons here on my PC ... press a button on the PC and get a notification on the smartphone David: accessibility consideration? ... would be important Ikeo: agree Pierre: API to TV? or smartphone? Ikeo: pairing between TV and smartphone Igarashi: depending on the TV's implementation ... possibly implemented as an application on the TV side ... the Hybridcast connect specification itself just specify the protocol Yongjun: additional latency? Igarashi: remote control for playback, etc.? Ikeo: go back to [Typical Sequence by additional APIs] ... explains the sequence of communication ... using websocket Yongjun: how much latency there? Ikeo: TV-set dependent Igarashi: support all the functions? Ikeo: some specific functions are supported Igarashi: arrow keys are supported? Ikeo: yes ... would like to support all the keys included in the TV remote Igarashi: but there are too many buttons Ikeo: also we need to consider security ... e.g., to avoid unexpected change of volume ... here, all the 5 APIs are implemented based on HTTP Sudeep: infra-red remote vs this API? ... what kind of values are added? Pierre: TV implements some specific capabilities Igarashi: TV vendors have to implement the APIs Lilin: how many TVs could be controlled? Ikeo: more than two ... we'd like to handle more than one TVs ... but TV vendors say hundreds of mobiles can't be connected ... maybe 2-3 ... TV is used within a local network ... user selects which would be the best one to get connected David: the system detects the devices available ... are children notified? Ikeo: application keep it stored in the session information ... the user don't have to mind Pierre: emergency notification itself is not included in the protocol ... its separate Ikeo: right ... these 5 APIs implemented within the device ... so device APIs ... not Web PIs Chris: looking at secure protocol? Ikeo: some solution ... we have two devices using some key Chris: second screen wg works on secure protocol ... so you're aware of that Ikeo: right ... the problem is HTTPs in local network ... tx for your comments! ... [Supposed use cases demo (2)] ... implemented as a service like Netflix or Amazon Prime Video ... application selects a program from a list ... using just one API ... (select a program on his smartphone) <MarkVickers> Q: What would the Hybridcast group like from W3C? New specs? Changes to specs? <MarkVickers> Thx Ikeo: launch the HTMl5 app ... using Dash.js Igarashi: can control playback, forward/backward? Ikeo: can be done using websocket <MarkVickers> While it's always interesting to see what other groups are doing, we have to focus on our goals to drive changes into W3C and increase adoption of W3C standards outside of W3C. Yongjun: device features can be controled? Ikeo: some of the features Yongjun: how many subscribers? Ikeo: In some research, 30% of the TV sets can connect to internet now and we expect these TV sets to implement this protocol near future Chris: what would the Hybridcast group like from W3C? <scribe> ... new specs, gap analysis? Ikeo: would like to combine some Hybridcast APIs to W3C standards ... e.g., playback API, as Igarashi-san mentioned ... between a mobile and a TV Kaz: kind of like the formerly proposed TV control API? Ikeo: yeah... Chris: or something like proposed by the second screen wg? Ikeo: we have to select web standard APIs ... can't create another APIs ourselves ... that's the second demo ... [Conformance Test and verification] ... [Conformance Test for Hybridcast-Connect] ... IPTV Forum Japan provides hybridcast connect standard ... and also test kit ... this is the overview ... (shows a diagram) ... emulator as the test environment Chris: cover the Web application? Ikeo: end-to-end test ... similar to the tests by HbbTV, etc. ... [Service Verification by MIC project and others] ... MIC is Ministry of Internal Affairs and Communications from the Japanese Government ... service verification with Hybridcast-Connect ... in 2018 19 companies ... in 2019 23 companies ... that's it for Hybridcast update ... thank you! Chris: what other specific things to address the gaps? ... relationship with the Web platform test, etc. Ikeo: we require some functions from Web APIs ... TV vendors sometimes want and sometimes not Chris: ok ... move on to the next topic Media Timed Events in Hybridcast Ikeo: [Service Patterns with Hybridcast Connect] ... broadcasters in Japan need trigger message to switch to broadcast service ... pattern1: from mobile app to broadcasing on TV ... pattern2: from another app in TV to broadcasting ... [Media Timed Events with Hybridcast-Connect] ... JP broadcasters interested in media timed events (MTE) ... same function as the trigger message ... there are two possible choices ... (MTE data in video resource + push emergency alert notification) to the smartphone ... another option (MTE data in video resource to another app on TV) ... there are those two possible patterns Chris: is this emsg? ... in ISO container ... in the DASH format Ikeo: yes Igarashi: upper pattern can be realized using mobile API ... but what about the bottom pattern? ... is the TV device at the bottom same as the one on the right? Ikeo: yes ... in the case of Android platform, the mechanism is something like intent Igarashi: general question about MTE ... unclear why you want to embed events within the video stream Ikeo: main reason is the cost of access the message API from mobile Igarashi: cost of notification servers Ikeo: right ... also the accuracy Igarashi: do we share the same requirements? Yongjun: which layer to be handled? ... should be fragment or manifest Igarashi: manifest embedded event? Ikeo: it depends on the needs ... in case of outbound, MTE might be written in the manifest ... and there would be possible delay Igarashi: could be updated frequently Ikeo: related to cost of access transfer ... trade-off of accuracy and cost ... show another demo on MTE ... (select an app on his mobile) ... send a message using hybridcast-connect to the TV ... this is embedded event ... emergency alert shown on the upper-right of the TV Chris: intended to synchronization of media? Ikeo: this mechanism just sends an alert ... and the Hybridcast application on the TV can handle how to display it ... [Usecases of MTE] ... switch broadcasting service from OTT triggered by emergency message ... super-impose time-dependent metadata, e.g., weather icon and event information ... new style of ad-insertion on a broadcasting service ... [MediaTimedEvents demos] ... demo implementations ... use case 1: switch live news program on braodcasting service from OTT service by emergency-warning message ... use case 2: super-impose a weather icon on the Internet video [21]slides [21] https://www.w3.org/2011/webtv/wiki/images/d/d1/MediaTimedEventsInHybridcast_TPAC20190916.pdf Chris: what is the requirements? Ikeo: would like to show the picture on the warning ... but sometimes overlaps with the important content (e.g., peoples faces) Igarashi: depends on the apps ... for some apps, accuracy is not important Ikeo: we have to consider the accuracy of timing for many cases ... that's all Chris: tx! Ikeo: btw, we would like other devices from the app on TV ... during the WoT demo, we'll show home appliance demos Kaz: at the lunch place, Argos on the 1st floor ... and Wednesday breakout Chris: excellent Ikeo: we'd like to use MTE as the basis [break till 11am] Media Timed Events Task Force [22]slides [22] https://docs.google.com/presentation/d/1f8LVFY3shrUsksKWLyBVQk3icDN4zEsgV0NX9oqPXNw/edit Chris: [Topics] ... in-band timed metadata and timed event support ... out-of band timed metadata ... improving synchronization of DOM events triggered on the media timeline ... also MPEG carriage of Web resources in ISO BMFF ... [History] ... our TF started in 2018 ... Giri Mandyam from Qualcomm preesnted work at ATSC and MPEG on eventing ... published use cases and requirements document early this year ... [Use cases for timed metadata and in-band events] ... MPEG-DASH specific use cases ... notification to media player ... another use case about getting matrix during playback ... ID3 tags: title, artist, image URLs ... ad insertion cues: SCTE35, SCTE214-1, 2, 3 David: keeping web page in synch with media ... you got slides and talking about the slide ... flip the slide dec and showing Chris: we have something like that in the explainer Pierre: we heard another use case in the morning Chris: multiple contents and multiple events Pierre: do you know if the cues tied to entire content? ... somebody may remove the trigger Chris: emsg separately handles Pierre: can remove part of the content and it's still relevant Chris: right ... [Recommendations] ... allow web application to subscribe to event streams by event type ... discussion on type of event ... maybe some concern ... something we can discuss ... also allow web applications to create timed event/timed metadata cues ... including start time, end time and data payload Igarashi: in the morning, we had some discussion on in-band message ... wondering if the current W3C standards support it ... only the scope could be in-band events? Chris: there are some implementations ... e.g., for HbbTV ... exposing MPD events ... W3C specs don't say anything about type of events ... next, actual triggering ... when cues are parsed from the media container by the UA ... when the current playback position reaches the cue start/end on the media timeline ... allow cues with unknow end time ... and finally ... improving synchronization (within 20 msec on media timeline) David: covers seeking? ... duration of events to be understood ... what would happen if jump-off? ... very hard to handle spike events Chris: some of that kind of use cases for DASH ... absolutely right Igarashi: requirement might be can detect that kind of delay ... applications would know about the difference between specified timing and actual fired timing ... we need to improved the timing (if possible) ... but should identify the gap ... e.g., based on the timestamp Chris: [Current status] ... almost complete Task Force use cases and requirements ... WICG DataCue explainer in progress ... API spec not started yet Yongjun: need to revise DASH spec, etc.? Chris: we need to have discussion about what kind of mechanism is needed first ... do we ask the UA to give structured data, etc. ... question about how the different formats should match the need Mark: the issue is how to map particular data format ... and how to present it ... the reference by HTML5 ... need to be updated ... based on the newest MPEG spec <cpn> [23]https://dev.w3.org/html5/html-sourcing-inband-tracks/ [23] https://dev.w3.org/html5/html-sourcing-inband-tracks/ Mark: maybe another form Chris: really interesting ... other things reference it ... URL spec ... not really standardized ... definitely right we need handle it ... in more standardized shape Mark: another question ... data cue was implemented in webkit before HTML5 <cpn> [24]https://www.w3.org/TR/media-frags/ [24] https://www.w3.org/TR/media-frags/ Mark: concern about syntax and semantics Chris: don't know the answer now ... a session by the Media WG will be held Mark: sounds like a right place Chris: [MPEG Carriage of Web Resources in ISO-BMFF Containers] ... saw TAG advice ... since then people working on MPEG as ISO/IEC FDIS 23001-15 David: probably public Chris: this topic is welcome to the MEIG David: good to have a workshop including the MEIG, ISO, etc. ... trying to get users of technology at the same time at the same place ... including security experts Pierre: what is the use case? David: two things ... carriage of web pages ... synchronization of media Pierre: but what would be the actual business cases? <MarkVickers> FYI, we have previously gotten permission from MPEG to host MPEG documents on the W3C member-only website. We could ask for MPEG for permission to host CMAF spec for this purpose. Pierre: btw, the community draft is available Igarashi: benefit of web resource embedded to MPEG ... possibly reduce the cost for the web servers ... could be beneficial Pierre: the offline case is weird to me Igarashi: one of the use cases to be addressed ... there are some offline use cases ... for packaged delivery Chris: [Browser support for DataCue] ... current support by browsers ... edge: HTML 5.1 DataCue attribute ArrayBuffer data; ... chrome: no support ... safari: supported ... firefox: no support ... HbbTV: HTL 5.1 (8 Oct 2015 ED) DataCue with native handling of player specific events ... [Player support for DASH and HLS events] ... Shaka Player: shaka.Player.EmsgEvent no internal handling of manifest refresh events. ... (some more examples) ... [Next steps] ... breakout session on Wednesday about "DataCue API and time marches on in HTML" at 11:00am ... raise issues against WHATWG HTML to propose changes to time marches on MarkW: may have people aware of the algorithm ... during the Media breakout Igarashi: in scope of the Meida WG? Chris: right ... the Media WG would take on standardization if the direction is correct ... we started a TF within MEIG and now are looking at WICG Igarashi: any friction? Chris: we're looking at the possible API design Igarashi: what I remember from the previous TPAC ... we were looking at WICG ... but what would be the direction now? Chris: we don't have enough input ... need more concrete feedback for the possible API ... in JavaScript at the moment ... would be good to have more involvement ... also would be good to have more browser vendors ... need to have wider discussion ... if we proposed issues, that should go to WHATWG ... increasing timing accuracy ... [References] ... (links to resources) Igarashi: WICG would have their meeting? Chris: Thu/Fri ... also we have a breakout session ourselves Igarashi: it's good timing to have discussion with them ... should ask the other participants about opinions as well ... need to get opinions from the MEIG guys Pierre: when would be our final report available? ... more input needed? ... anybody have any specific objections? Igarashi: we have not specifically asked the MEIG for opinions ... report itself is about requirements ... it's an IG Note. right? Chris: yes Pierre: the report says something is missing and to be added? ... shouldn't say that explicitly? Chris: solution design to be done by WICG ... our TF could continue editorial changes ... everybody, please join in CTA WAVE update (we're delayed by 30mins) [25]slides [25] https://drive.google.com/file/d/1-mAhZe8s2TRDygCW1fPc-aJkvA1yCTH1/view?usp=sharing John: John Riviello from Comcast ... quick update on CTA WAVE ... [The Web Appliction Video Ecosystem Project] ... aims, focuses, ... ... [Supporting a fragmented OTT world] ... fragmentation impacts content providers and device makers ... [Brief history] ... CEA initiated the GIVE project in 2015 ... CEA becomes CTA in Nov. 2015 ... [Steering Committee] ... technical WG ... CSTF for content specification ... DPCTF for testable requirements ... HATF for reference application framework ... [WAVE bridges media standards & web standards] ... [Curent WAVE Membership] ... many members ... overlapping with W3C Members ... [What is the Common...] ... [WAVE COntent Spec & Published CMAF Media Profiles] ... [Media Profile Approval] ... profiles are added ... typically updated once a year ... [WAVE Content Specification 2018 AMD 1 - Video Profiles] ... [WAVE Content Spec 2018 AMD 1 - Audio Profiles] ... [WAVE Programs and Live Linear...] ... [Anticipated WAVE Content Spec 2019 Updates] ... [Test SUite: COntent Verification Tool] ... verification content ... shared with DASH-IF conformance validator ... [CSTF - Specification Process] ... annual f2f meeting ... [Links] ... links for resources ... [HATF: HTML5 API...] ... What We Do in the HATF] ... playback audio-video media ... [HATF Work Plan] ... W3C Web Media API CG ... [HTML5 APIs: Reference Platform] ... one content format but multiple devices ... [HATF Specs] ... snapshots ... Web Media API snapshop (WMAS) ... CTA and W3C co-publishing Igarashi: what do you mean? John: working on the same document Igarashi: not WG but CG? ... it's not "W3C Recommendation" but "CG Report" Francois: fyi, there will be discussion about W3C process during this week Alan: part of the plenary on Wednesday John: [Anticipated Web Media API 2019 Snapshot Updates] ... update to ECMAScript 7 ... CSS snapshot 2018 ... [HATF Testing Framework] <MarkVickers> FYI on referencing WAVE specs: ATSC references the WAVE WMAS as published by CTA, which is referencable. The W3C version of the WMAS spec, like all CG specs, includes boilerplate language that it should not be referenced. John: WMAS Testing Suite Updates] ... [Abstracted Device Playback Model] ... (skips some slides) ... [Spec Highlights and Outline Dec 2018] ... [Promisses in Spec for 2019 and beyond] ... [Test Sutie: RFPs] ... [Q&A] ... questions? Igarashi: what are you going to talk about "type1 player"? ... any room for W3C standardization? ... if you have any specific requirements, the MEIG can discuss that ... btw, what is the "Content Model Format"? Chris: question around testing ... is the work related to the web platform testing? Pierre: should we put that on the agenda for the afternoon? all: ok Review open issues Chris: we use GitHub to manage issues ... most of the issues will be covered in the afternoon jointly with the other WGs ... but one specific issue here about frame accurate synchronization and seeking Francois: [Related GitHub Issues] ... issue 4, 5, 21 ... the main issue is #4 frame accurate seeking of HTML5 MediaElement ... [Categories of Use cases] ... 2 different use cases ... seeking and rendering ... [Main Seeking Use Cases] ... non-linear edition in a browser ... can be cloud-based ... collaborative review ... evidence playback by camera and video ... [Seeking Gaps] ... currentTime is not precise enough to identify individual frames ... also no way to seek to the next/prev frame in the generic case ... just matter of time ... when is going to be the next frame ... [Main Rendering Use Cases] ... dynamic content insertion (splicing) ... video overlays ... media playback synchronized with map animations ... synchronization between audio and timed text, e.g., karaoke ... synchonized playback across users/devices Igarashi: requirements for time seeking? Francois: this is rather rendering issues Pierre: sample alignment and duration ... current web platform doesn't allow frame-accurate timing Francois: [Rendering Gaps] ... currentTime is not precise enough to identify individual frames ... also timestampOffset is not precise to identify frame boundaries ... it's hard to track media timeline frame by frame ... in any case there is no mechanism to handle frame accuracy ... also synchronization between video and audio ... if you look at global synchronization ... no way to tie the rendering of a video frame to the local wall clock <MarkVickers> Following up on earlier question: It has always been the intention of WAVE to contribute back to W3C any new tests and also any changes to the W3C test runner. WAVE representatives met with the W3C test group at TPAC 2018. There was an issue opened on April 2, 2019: [26]https://github.com/web-platform-tests/wpt/issues/16214 There was a PR entered on June 12, 2019: [27]https://github.com/web-platform-tests/rfcs/pull/23 [26] https://github.com/web-platform-tests/wpt/issues/16214 [27] https://github.com/web-platform-tests/rfcs/pull/23 [lunch till 1:30pm] Frame accuracy synchronization (contd) Francois: continue the slides ... [Rendering Gaps that Remain] ... currentTime is not precise enough ... timestampOffset is not precise enough ... three following requirements deleted ... [Seeking Gaps that remain] ... [Next Steps?] ... what do we want to do? ... follow up on MTE recommendations around synchronization? who? ... wrete a UCR document on frame accurate synch? who? ... feed needs back into WHATWG and Media WG? who? ... different possible groups to bring ideas to ... possibly machine learning group? ... discussion with different people on Wednesday Chris: production use cases? Pierre: put together a presentation ... what's happening about a lot professional asset ... have some demo as well MarkW: the problem is probably how to identify individual frame ... could be ended up with overlaps Chris: we need rationale Igarashi: (asks new comers to sing up with the attendees list) ... btw, this proposal includes two different points ... because the difficulty for realization is quite different ... depending on theperformance of the browsers and the hardware Francois: maybe write two documents or might be simply continue discussion ... there are different use cases ... some of them might be out of scope Pierre: sound and video synchronization is a use case ... not even possible currently ... there is no API for that purpose today Igarashi: that requirement is related to time seeking ... different from synchronization itself Pierre: when you say "seeking", it's API level. right? Igarashi: currently there is no way for requirements to specify how quickly browsers should behave Pierre: it's largely implementation-dependent? Igarashi: yeah ... current time issue and synchronization are different issues ... wonder if any other W3C specs handle that kind of performance ... how fast browsers are expected to render the data ... we need to talk about performance Pierre: sounds like you're interested in making contribution :) Yongjun: need some mechanism to handle frame jumping ... not only at the beginning ... if we care about one case, we may miss another case Igarashi: ad-insertion seamlessly? ... accuracy of seeking is important for many use cases ... but we should distinguish time seeking accuracy Pierre: if there is a stream and also another stream starts, in that case, need for frame accuracy Igarashi: that's true Professional media workflows on the web Pierre: very timely topic ... proposal by MovieLabs ... increasing popularity ... [Web applications are coming to profesisonal media workflows] ... why? ... web applications have become mainstream ... web platform media capabilities are tantalizingly close ... profesisonal audiovisual assets are moving to the cloud ... [Why move audio-visual assets to the cloud?] ... instead of using UPS ... now available immediately on the cloud ... it's more secure actually ... and of course more efficient ... [Today] ... previsualization, visual effects, grading, editing, localization, mastering, qualit check, archival, distribution ... [tomorrow] ... all of them will be on the cloud (and can be accessed via web applications) ... [demo] ... ownzones ... content already on the cloud ... there is an editor here ... (for audio and timed text) ... (going back to the presentation) ... [Some steps of the workflow remain out of reach of web applications] ... gaps exist in the web platform ... what's missing? ... that's it Chris: we have many items, so don't want to dive into the details Yongjun: as far as I know, people use MSE Igarashi: video editing using browser ... requirements for rendering related to multiple video clipping ... handle frames seamlessly Pierre: we need volunteers MarkW: not volunteering myself but support the use cases Scott: folks here might want to consider local content Igarashi: local content using browser? Scott: not necessarily on the cloud ... how to handle frame accuracy on the local devices Pierre: help document the issues? ... take a first step ... somebody needs to take the lead ... this is listing the current issues Gary: interested Chris: thank you Samira: gathering data ... we have a few ideas ... one of them is adding attribute to video tags ... somebody from Google also proposed media container ... my first question is ... would you have any thoughts ... will host a session on Wednesday Chris: related to content representation ... quite a lot of valuables ... we'll talk about caption later Andreas: where to standardize 360 video, etc. ... we have a presentation on that later in the afternoon ... also a session on Wednesday ... possibly tomorrow as well Chris: anybody aware of MPEG format update? David: whole bunch of work <tidoust> [28]https://mpeg.chiariglione.org/standards/mpeg-i/omnidirectio nal-media-format [28] https://mpeg.chiariglione.org/standards/mpeg-i/omnidirectional-media-format <scottlow> [29]https://mpeg.chiariglione.org/standards/mpeg-i [29] https://mpeg.chiariglione.org/standards/mpeg-i <dsinger> [30]https://mpeg.chiariglione.org [30] https://mpeg.chiariglione.org/ Andreas: TTWG has liaison with MPEG ... but just one part of scenarios ... inband information ... doesn't sort the issues about outband captioning <Joshue108> There are also accessibility requirements around how 360 is standardised. Andreas: possibly discuss that tomorrow? Samira: possible ... how many content producers, providers, here? ... what block you? Song: China Mobile Igarashi: VR content protection? Samira: can be represented as VR ... magic window scenario ... just wanted to bring this discussion up Chris: what's the natural home for this discussion? ... first candidate is Timed Text Samira: just wanted to share the ideas since this is an IG Josh: There are accessibility requirements if 360 is to be standardised, around an architecture that will support accessibiity and multimodal requirements. Andreas: would like to come back later in the afternoon ... where to do it ... really difficult to find a right place Chris: related to accessibility Sudeep: Chair of the Web&Networks IG ... will have our meeting tomorrow ... please drop by ... interested in Media Timed Event as well ... network latency ... very happy to give inputs Chris: interesting questions ... very close relationship with this group ... having a Web interface ... webrtc stream for multiple different sources ... it is stuff we've been implementing ... not necessarily synchronized with each other Sudeep: how should we bring back? Chris: GitHub issues ... also we have monthly IG calls ... have media-related topics Josh: particular accessibility issue in synch with video stream Chris: yeah Josh: bunch of stuff <Joshue108> [31]https://www.w3.org/WAI/APA/wiki/Accessible_RTC_Use_Cases [31] https://www.w3.org/WAI/APA/wiki/Accessible_RTC_Use_Cases Josh: can put resource on what I'm working on (above) ... related to this group ... different modality channels based on user's preference, TTS, braille, etc. (kaz remembers the MMI Architecture and SCXML, which are used to integrate multiple modality channels :) Chris: any other issues? Igarashi: local packaging? ... publishing group is working on packaged media ... playback locally ... on a local storage ... might be with very high resolution of time Chris: seems we need another gap analysis <tidoust> [Note the breakout session on Web Packaging planned on Wednesday: [32]https://w3c.github.io/tpac-breakouts/sessions.html#wpack] [32] https://w3c.github.io/tpac-breakouts/sessions.html#wpack Bullet Chatting Song: Song Xu from China Mobile ... would give a presentation about bullet chatting ... Michael from Dwango as well <cpn> [33]https://w3c.github.io/danmaku/index_en.html [33] https://w3c.github.io/danmaku/index_en.html [34]proposal [34] https://w3c.github.io/danmaku/index_en.html <scribe> scribenick: tidoust Song: Interactive tool for video broadcasting over the Internet. Use cases: see reviews of group users. Real-time interaction, engagement for young generation, to show social presence. ... Implementation is difficult because you need to compute the positioning and animation of bullet chatting, rendered in DOM or Canvas and overlaid on top of the video. ... Strong demand for this type of applications, particularly in Asia ... Standardization would improve UX, reduce the difficulty in implementation. ... We suggest to define a standard format for bullet curtain. ... We started an analysis to identify gaps. No specific API introduced for the time being. ... Bullet chatting is basically floating text over the screen with four attributes: ... mode, basic properties, timeline, and container (typically the video) ... [going through Bullet Chatting Proposal document] ... During streaming, two main ways to present: chatting room or bullet chatting. ... Advantages of bullet chatting display are that there is a wider display area and it does not require the user to move her eyes. ... The movement from right to left allows users to read content quickly (and again without moving her eyes). ... Sometimes, it's not only about comments, it can be text to improve the feeling of horror videos for instance. ... Also used to share messages in stadiums on a big wall. Michael: I'm from Dwango. Use cases and requirements for our current service Niconico. ... Niconico is a streaming Web site launched in 2006. Since its inception, its unique feature has been its comment system. ... [showing a demo] ... allows to create a user experience. Pierre: Who specifies at what vertical position the bullet curtain appears? ... Do you foresee that to be done at the client side? Song: No, done on the server side Pierre: So the format has all the positioning information. Michael: In the current implementation, clients do the rendering, and they all have the same algorithm, so deterministic. Pierre: If things were standardized at W3C, would the positioning be imposed by the server? Michael: Currently, we'd like the client to have the ability to position the comments. Pierre: So the client receives the comments and decides where to lay them out. Igarashi: You want to let the browser do the whole rendering? Michael: No, the Web application. ... Goal of the standardization is to have a shared format for bullet curtains, because many providers have a similar comments system (Niconico, Bilibili, etc.) Song: First step is to define an interoperability format. If there is a way to involve the browser vendors, then great, second step. MarkW: Browsers would want to know why something cannot be done in JS. David: And you could possibly do it with WebVTT / TTML. Song: For advanced features, there are things that TTML does not address. Happy to talk with TTML folks though. Michael: Use cases and requirements level for now. Possible solutions are still very early stage. ... Bullet curtain allows to create feelings such as sharing content with friends. ... Comments can be used to improve the video with artwork, or even to flood the video with comments. ... Comments have become an important part of Niconico's culture. ... Part of on-demand and live-streaming services of Niconico. ... Comments move right to left across at set times, based on the media timeline. Chris: If I pause the video, do the comments pause? Michael: Yes. ... Comments are clipped to the edge of the player (or to an arbitrary region). ... When the video loads, comments are loaded from the server and rendered. ... If a user submits a comment, it appears immediately to the user, and gets shared to other viewers. ... Seeking to the same time in the same video will have the same comment appear at the same time and at the same position. ... As if the comments were part of the video, comments scale with the video in particular. ... Comments can be interactive (e.g. context menu) MarkW: Layout problem (HTML is good at it), animation problem (Web Animations), but the thing is Web Animations ties animations to the wall clock, whereas here animation is tied to the media clock. ... That may be a useful gap to identify Chris: Came earlier during Francois' presentation. Tying non-media content rendering to media timeline. Igarashi: Some requirements about positioning the subtitles. ... Client decides arbitrary where to position the comments. Michael: Yes. Igarashi: Content provider does not care about positioning of subtitles. Sangwhan: Aside from Web, do you also want to handle support for native players? ... That would change perspectives. Michael: We do have native apps, so we'd be interested with a solution that covers that space too. Sangwhan: According to Mark's idea, if it's tied to the animation timeline in browsers, you're restricting yourself to Web environment. Kaz: When I talked to Koizuka-san from Niconico, he mentioned extension mechanism named "Nico-script", and that mechanism has capability of specifying style and position of captions. so that capability could be also considered at some point. maybe not now, though. <MarkVickers> I'm not staying connected for the joint meetings. Have a good TPAC all! -mav Joint meeting with Second Screen WG/CG Chris: The Second Screen WG/CG made a lot of progress on the Open Screen Protocol for discovering, authenticating and controlling remote displays on the local network. MarkF: I work for Google. Been involved in Second Screen since 2015. Second screen for the Web is the way we want to enable Web applications to take advantage of connected displays/speakers and render different types of content. ... Content can be a full Web page or specific media. ... The Presentation API enables a web page, called the controller, to request display of an URL on a remote display on the LAN. ... Example of a photo app that displays the loaded picture on a large display. You can play media, do gaming, collaboration tools. Pretty agnostic, but our experience shows that it's mainly used for media playback. ... The Remote Playback API allows a web page on which there is a media element to remote the playback of the media element on a second screen, either through media flinging where the URL to play gets sent to the remote device, or media remoting where the media gets streamed to the second screen. ... Both APIs are in Chrome. ... The APIs were designed to take advantage of proprietary protocols. To get broad adoption, we decided to develop an open set of protocols so that implementers could all support the APIs in an interoperable way. ... We hope to converge at the end of the Second Screen F2F meeting this week to v1.0 of the Open Screen Protocol. ... One use case for the future: enabling Web applications to generate their own media and present it to a connected display, e.g. for gaming. ... The Open Screen Protocol supports all sorts of use cases that we hope to expose to Web applications in the future. Yongsun: Support of QUIC in smart TVs. UDP is not supported in some TVs. Sangwhan: UDP is supported at the kernel level. MarkF: in our library implementation, we expose UDP but that's pretty much the same thing as what you get at the system level. Chris: One of the question that came up in our previous F2F meeting is around synchronization, e.g. ability to provide audio description on their device while they are sharing a media element on a second screen. ... Within that, there is the question of how close the synchronization needs to be. ... We worked on close synchronization between main screen and companion device in HbbTV. MarkF: Does the HbbTV specification rely on clocks? Chris: Yes, clock synchronization and then the devices can make adjustments to playback to stay in sync. MarkF: We need a mechanism for the two sides agree on a wall clock for presentation. ... If the HbbTV covers all of that, we can have a look for OSP. Chris: Yes, it does. <anssik> Open Screen Protocol issue Requirements for multi-device timing while streaming [35]https://github.com/webscreens/openscreenprotocol/issues/195 [35] https://github.com/webscreens/openscreenprotocol/issues/195 Chris: Some implementers have found it difficult to achieve that level of synchronization. It's not so widely implemented for now. ... I can provide information on how that has been done. MarkF: Collaboration between the protocol and the application levels. Chris: And also something that exposes the pipeline delays. MarkF: One of the things that seem very important is the establishment of a secure communication between devices, which could have broader implications, such as connected home scenarios. ... it could be a good foundation for that. Part of the OSP focus has been on authenticating devices, currently based on SPAKE2. ... We're not currently focused on enabling one piece of software to find out attributes of another, for instance who manufactured it, what does it do. <anssik> SPAKE2 [36]https://datatracker.ietf.org/doc/draft-irtf-cfrg-spake2/ [36] https://datatracker.ietf.org/doc/draft-irtf-cfrg-spake2/ MarkF: You could take the chapter on authentication and use it elsewhere. ... We did anticipate that there may be other use cases than the ones we foresee, so have landed an extensibility mechanism. Sangwhan: Is there a registry for these capabilities? MarkF: Yes, it's on GitHub. ... You can be a presentation controller, receiver, send or receive media, that's all negotiable in the OSP. Chris: I suspect remote playback of encrypted content is a use case shared by different members here. MarkF: The API is pretty much agnostic. At the protocol level, we haven't tried to add support for messages to exchange to support encrypted media. ... That seems more to be a use case for the Presentation API where the application can create and exchange application-specific message commands. ... Remote playback of encrypted media is closely tied to credentials, and that's application level. MarkW: The thing that you don't have here is the streaming model where the controlling device has the decryption key and wants to stream the content to the receiver device. ... What happens to the media stream when it reaches the receiver? Goes to a media element or through JS processing? Peter: receiver is handling the decoding. Chris: Is there an IG recommendation that we'd want to make? MarkW: The most likely model for us for doing this would be to have a receiving web application that handles the user's credentials Chris: That would make the sync issue interesting because it is then at the application level. ... One of the issues we have with Remote Playback is that we want to provide a custom UI, which means that we rather want to use the Presentation API for that. ... Didn't we discuss having a Media element through the Presentation API that gets automatically synchronized with local content? MarkF: I believe that's correct. I don't recall the status of it. It came up in May 2018, I think. <anssik> Second Screen May 2019 F2F [37]https://www.w3.org/wiki/Second_Screen/Meetings/May_2019_F2F [37] https://www.w3.org/wiki/Second_Screen/Meetings/May_2019_F2F MarkF: I think we probably agreed that it should be possible. It probably requires a few tweaks to the protocol so that it knows that the remoting is part of a shared presentation. ... We discussed whether everything could be done in script. Same recommendation for synchronization. What you might be missing is the latency of the media rendering pipeline. Chris: I have seen implementations that manage to do synchronized playback across devices through a timing server. Igarashi: I don't follow the discussion on encrypted media. You are not going to define how keys are exchanged in the protocol? MarkF: Someone with more experience on EME might be able to shed some lights as to what would be required. ... One reason we designed an extension system is that people interested in new features can propose them, prototype implementations, and then we can incorporate them in the spec if all goes fine. We don't have the expertise in the group. ... We're not defining the path for encrypted media from one device with another. Might work if both devices support HDCP. ... I think there is an open issue in our GitHub about remote playback and encrypted media. Igarashi: Arbitrary application message passing is supported? MarkF: Yes. ... In the spec, you'll see bindings between the API and the messages exchanged in the protocol. ... For instance, video.remote.prompt() requires exchanges messages between devices MarkW: Could the protocol work on TCP? Peter: You'd have to advertise it differently Igarashi: [question on security during remote playback] MarkF: the Remote Playback API does not require the receiver to be a user agent in the usual sense, it does require the receiver to support media playback as in the HTML spec. MarkW: The Presentation API requires the receiver to be able to render the URL, but the URL could be a non HTTP URL, custom schemes may be supported instead. MarkF: The spec defines processing of HTTPS URL, the rest is undefined. <anssik> Open Screen Protocol [38]https://github.com/webscreens/openscreenprotocol/ [38] https://github.com/webscreens/openscreenprotocol/ MarkF: We have a writeup of how the protocol interacts with custom schemes in the GitHub repo. Chris: That has been one of the extension mechanisms that we've been interested in for opening a Web page that has broadcast capability in HbbTV (perhaps Hybridcast has similar needs) <anssik> Custom Schemes and Open Screen Protocol [39]https://github.com/webscreens/openscreenprotocol/blob/gh-pa ges/schemes.md [39] https://github.com/webscreens/openscreenprotocol/blob/gh-pages/schemes.md [discussion on second screen support in Hybridcast] MarkF: regarding authentication, we looked at J-PAKE and request/response challenges but we had memory concerns there so switched to SPAKE2 following internal discussion with security experts at Google. Peter: The protocol allows for more authentication mechanisms in the future. ... Devices can support their own mechanism. Igarashi: Co-chair of HTTPS in local network CG, meeting on Thursday morning. We haven't reached discussion on authentication. Would be good to align with Open Screen Protocol. Sangwhan: Is there a prototype? MarkF: We recently decided to add streaming to the OSP, which complicated things. We have a first implementation of Presentation API commands. No crypto because we've kept changing that. ... The library is coming. It implements the protocol. It does not do media rendering, it does not have JS bindings, etc. <anssik> Open Screen Library implementation [40]https://chromium.googlesource.com/openscreen/ [40] https://chromium.googlesource.com/openscreen/ Igarashi: If you want to apply the OSP to the broadcast protocol, we need to consider the case where the remote device is not a browser. For instance, channel change is done by the system, not the application. MarkF: Capabilities like supporting channel tuning is not in the OSP. If you think that the communication channel needs to be terminated on channel change, that can be added. Igarashi: In the case that some arbitrary message protocol is still necessary, you'd use the Presentation API, but the receiver may not be a browser agent. MarkF: seems like something for an extension. Chris: OK, thank you for the discussion. MarkF: Mostly, we want input on use cases that we haven't considered yet. We'd love to get feedback on the extension mechanism as well. Pierre: Thank you. Joint meeting with Timed Text WG Andreas: We could start with 360 standardization Nigel: In TTWG, we're in the final stages of rechartering. ... Some things that we're considering such as karaoke. <Joshue108> [41]https://www.w3.org/WAI/APA/wiki/Accessible_RTC_Use_Cases [41] https://www.w3.org/WAI/APA/wiki/Accessible_RTC_Use_Cases Nigel: Quick agenda bashing, any topic you'd like to cover? Josh: accessibility use cases? See accessible RTC use cases document Chris: TTML and MSE? Nigel: Yes, opinions about exposing TextTracks from MSE. <Joshue108> apologises for throwing a curve ball to Nigel, I'm here for the XR bit but think this doc may still be useful as an FYI Andreas: Focus the discussion of the day on standardization of 360 subtitles. Most of the stuff comes from an EU research project. ... To make it short, there have been extensive user tests. For captions, main requirement is to have subtitles that are always in the field of view. It's enough to have them on a 2D plane, no need to have them positioned in 3D. ... There should be some indication of where the audio source is positioned. ... Of course, you also need features present in TTML, TTML-IMSC profile being a good example. ... [demo of an application to test subtitles positioning] ... Lots of activity starting last year at TPAC. We started with a discussion in the Immersive Web CG. Then discussion within the TTWG, Media & Entertainment IG. ... In the end, we realized we needed more people from immersive and browser vendors. ... We wrote a proposal to be discussed in the WICG. ... There has been no comment on the WICG forum yet, so question is how do we proceed? ... Two additional activities worth noting. A colleague from Google proposed the creation of an Immersive Caption Community Group, and XR accessibility W3C workshop in November. ... There is awareness that something needs to be done. ... Hard to get enough resources to get started though. ... How to get time and resources from implementors? <Joshue108> Inclusive Design for Immersive Web Standards W3C Workshop Seattle Nov 5-6 <Joshue108> [42]https://www.w3.org/2019/08/inclusive-xr-workshop/ [42] https://www.w3.org/2019/08/inclusive-xr-workshop/ Andreas: Everything is evolving, nothing really fixed. ... Is it really a web platform topic? ... Important to know when to stop if there is not enough interest. ... Apart from which group should deal with it, the question is also where does this solution fit? ... Authoring environments (Unity, Unreal), Web applications, WebXR API (linked to OpenXR) and 360 / XR device ... How to follow-up? I thought WICG would be the right place, but if there is not enough place, there is still the question of whether that's the right place. Not sure about Immersive Caption CG since it does not exist yet. ... TTWG is the right group but we need more expertise from the XR world. ... Another solution is to continue the work in a "private" repository. <Zakim> nigel, you wanted to ask what is the state of documentation of the requirements right now Nigel: What is the state of documentation in terms of the requirements? ... Describing positioning in 3D space, can I do it with audio? Andreas: There are documented user tests, as part of an European project deliverable. Nigel: I was thinking about requirements documentation. What is the problem that you're trying to solve, user needs. Samira: Who was the person who started the Immersive Caption Community Group? Andreas: Christopher Patnoe at Google Samira: OK. Another comment is that WebXR is becoming more stable. Andreas: Yes, the question for me is where should this go. ... The WebXR API does not know anything about what's inside the WebGL right now. Chris: Is all that's needed a delivery format and then some library can place that in the immersive environment? Igarashi: Do we need to extend APIs in the browser to support this? <igarashi> -q Andreas: OMAF defines a way to multiplex IMSC subtitles with MP4, but then it's all bound to that content format. Not sure it's sufficient for interoperability scenarios. <igarashi> +q Kaz: wondering about the possible relationship with Web Francois: WebVMT is about tracks positioned on a map, not in 360 videos. Kaz: However, there is a possibility of mapping the positioned map to the 360 screen and then put caption on that screen Andreas: It would be an option to have a subtitle format, but burning captions in a frame does not provide good user experience. Josh: Looking at things from an accessibility perspective. APA would seem a good group to talk to. Andreas: We talked a lot with Judy, Janina and so on. <Joshue108> [43]https://www.w3.org/WAI/APA/wiki/Xaur_draft [43] https://www.w3.org/WAI/APA/wiki/Xaur_draft Josh: We created a list of requirements for XR in APA. <samira> IW group is also discussing dom overlays so this is another option for subtitles Pierre: How many people in this group doing 360 videos and XR content? ... One possibility is that this group is not the best group to get feedback from. Andreas: I don't know, that's what all groups say ;) ... We need a critical mass to do it. Pierre: People that build apps for Oculus, are they around? Andreas: I spoke to some of them. They always say that they don't provide subtitles. ... Some discussion in Khronos with Unity and Epic. ... I talked with Immersive Web folks. We'll talk about that on Wednesday 11:00 during Samira's breakout session. ... The issue that we have is that there is not endless time to deal with it. The project is running out. It stops next year. To push a standard, it will take 2-3 more years. <Joshue108> There are very few testing with people with disabilities in this space so this is very interesting. Igarashi: From a content production perspective, I'm interested in a format, but not sure about browser support for this. [44]https://github.com/immersive-web/dom-overlays [44] https://github.com/immersive-web/dom-overlays Francois: Not clear to me what you want to be standardized. DOM overlays could be one building block. Andreas: Yes, DOM overlays may be a good way forward to render captioning thatn burning things in WebGL. <Zakim> nigel, you wanted to wonder what the smallest thing is that we need to standardise first - is it a syntax for expressing a 3D location? <Joshue108> +1 to Nigel Nigel: Same point. Do we have agreement that it's about a syntax for expressing a 3D location? Andreas: Actually, that's not what we need, since we want it to appear on a 2D plane, that is what the users want. ... We need a way to indicate where in the 3D space the audio source is coming from. Gary: So you need some positioning in 3D to make that possible. Andreas: Define a good container is another issue. Josh: in the User requirements document I showed you, we took a modular approach. ... This architecture does not exist yet. <Joshue108> [45]https://www.w3.org/WAI/APA/wiki/Media_in_XR [45] https://www.w3.org/WAI/APA/wiki/Media_in_XR Josh: We're also looking at Media requirements in XR. Not vetted by the APA WG yet. Andreas: Lots of 360 content for the time being, and a lot of it without captioning. <Joshue108> s/module approach/modular approach Gary: WebVTT update. I joined TTWG half a year ago. Trying to get WebVTT to progress. One of the big thing is an implementation report exists right now. ... Something like 6-7 issues with it. <atai> Link to 360 subtitle requirement [46]https://github.com/immersive-web/proposals/issues/40 [46] https://github.com/immersive-web/proposals/issues/40 Gary: Basically, we're looking at features implemented in browsers and in VLC. Then identify features at risk, and possibly remove them to get a V1 out. ... Then hopefully convince browser vendors to implement the features that we may remove. <gkatsev> [47]WebVTT Implementation Report [47] https://www.w3.org/wiki/TimedText/WebVTT_Implementation_Report Glenn: Any SMPTE spec that includes 3d positions of audio sources? Nigel: That's a good question. ... One of the things we're doing around TTML2 is adding new functionality in extension modules. We're trying to constrain the core, and then provide the rest in extensions. ... There are a few ones that are ongoing. ... [details extensions] ... Right now, audio/video comes to MSE but not text. MarkW: My personal position is that things should be symmetrical across media types. ... At least in our application, we prefer to do the rendering of text tracks ourselves. ... It would be advantageous in which the browser is aware of text tracks. Nigel: You said my sentiment much better than I could. Gregg: I would argue that we don't want to render them ourselves, but we still want to control the rendering with our styles. MarkW: Yes, we want to have enough control of the rendering, but we could offload the rendering to the browser, that would be great. Nigel: It's been hard to get statistics about user customization, or people that play back content with captions. MarkW: In terms of rendering, you would still want the site to control enabling/disabling. <atai> +1 Gary: We shouldn't try to do the same thing twice. If there's more support to do the new generic TextTrack thing, then that's good. Pierre: Two different questions: any objection to enabling symmetry in MSE? Are you going to use it? MarkW: First question is whether people think that could be harmful. Nigel: OK, I just wanted to raise it to get feedback. [No concerns expressed regarding question on whether people think that could be harmful] Josh: About accessibility in WebRTC use cases, challenge of synchronizing some of these things together when switching to a different modality. That's one. Nigel: It would make sense to talk about live contribution to see where that fits. How does live contributions actually work, what's the mental model? ... Alright, I think we covered all topics. Closing and wrap-up Chris: Thinking about Media Timed Events, some editorial work. Planned discussion on DataCue. Around bullet chatting, more conversation will happen this week. ... Some possibility to go to Timed Text WG. Nigel: It feels to me that this IG could be the best place to give guidance for that if there's no clarity in TTWG on Friday about that. Andreas: Can you explain again how you want to proceed? ... Draft published in the Chinese IG, what would the ideal next step be? Song: Initially, contributors were from China. Now that NicoNico is engaged in discussions, work could go to TTWG, or perhaps in another group. ... We want the use cases to be approved by the IG, afterwards we'd like to push standardization work on identified gaps. ... Within the next few weeks, we'll have a last version of the use cases. Andreas: OK, so this week would be a good opportunity to decide where this should go. Chris: We had a lot of discussion around synchronization today. Frame accurate rendering. ... Ability to seek accurately within videos. ... Some interest to follow-up, although no one volunteers. ... The media production use case that Pierre presented would be a good perspective to address this. Pierre: With an action on Gary to follow up with Garrett Singer on that. Chris: Secure communications between devices, we heard interesting stuff from Hybridcast and HTTPS in local network, and Second Screen. Interesting set of approaches that could be compared. ... Seems like a good fit for HTTPS in local network CG discussions. ... Clearly the immersive captioning is interesting, but not sure what next step in this group should be. Maybe the Immersive Captioning CG could be the right forum. ... We talked about 360 videos. That's something that the IG could follow on. We have liaison with MPEG. Unless you feel that immersive group would be a better home. Samira: Possibly. At this point, I'm gathering input. Chris: Finally, there's the timed text in MSE proposal. Would that sit in TTWG? MarkW: It would be in scope of the Media WG. Chris: Have I missed anything from the summary? Pierre: One encouragement for you to clarify the scope in Media Timed Events. Chris: And also possibly make more specific recommendations. Pierre: I think it helps to have something concrete. Chris: OK, I think that's everything, thank for your presence today! Summary of Action Items Summary of Resolutions [End of minutes] __________________________________________________________ Minutes manually created (not a transcript), formatted by David Booth's [48]scribe.perl version 1.154 ([49]CVS log) $Date: 2019/10/04 17:02:15 $ [48] http://dev.w3.org/cvsweb/~checkout~/2002/scribe/scribedoc.htm [49] http://dev.w3.org/cvsweb/2002/scribe/
Received on Tuesday, 8 October 2019 07:21:49 UTC