- From: Kazuyuki Ashimura <ashimura@w3.org>
- Date: Tue, 8 Oct 2019 16:20:28 +0900
- To: public-web-and-tv@w3.org
- Message-ID: <CAJ8iq9UNOnA9zaaqJwVXZokV8TH0Jwx=XYLsc1+HBo8axjiBfQ@mail.gmail.com>
available at:
https://www.w3.org/2019/09/16-me-minutes.html
also as text below.
Very sorry for the delay. I thought I had sent out these draft minutes
the other day, but it seems the message didn't go out.
Thanks,
Kazuyuki
---
[1]W3C
[1] http://www.w3.org/
- DRAFT -
Media and Entertainment IG f2f meeting at TPAC 2019 in Fukuoka
16 Sep 2019
[2]group photo from the WoT F2F meeting on 19-20 September 2019
[2] https://www.w3.org/me-f2f/photos/0916/DSC_3346.JPG
Group photo from the WoT F2F meeting on 19-20 September 2019
(Some more photos from at the [3]F2F at Hilton)
[3] https://www.w3.org/me-f2f/photos/0916/
[4]Agenda
[4]
https://www.w3.org/2011/webtv/wiki/Face_to_face_meeting_during_TPAC_2019#Agenda_Monday_16_September_2019
Attendees
Present
Andreas_Tai, Anssi_Kostiainen, Chris_Needham,
Daiki_Matsui, David_Fazio, Eric_Siow, Florent_Castelli,
Gary_Katsevman, Greg_Freedman, Hiroaki_Shimano,
Hiroshi_Fujisawa, John_Riviello, Josh_O_Connor,
Kaz_Ashimura, Li_Lin, Mamoru_Takagi, Mark_Watson,
Masaya_Ikeo, Masayoshi_Onishi, Nonoka_Jinushi,
Pierre_Lemieux, Scott_Low, Shinya_Abe, Sudeep_Divakaran,
Tatsuya_Igarashi, Xu_Song, Yongjun_Wu, Youngsun_Ryu,
Yuki_Yamakami, Takio_Yamaoka, Akihiko_Koizuka,
Keiichi_Suzuki, Michael_Li, Takahiro_Kumekawa,
Chris_Cunningham, Taki_Kamiya, Jeff_Jaffe, David_Singer,
Glenn_Adams, Hyojin_Song, Francois_Daoust, Ken_Komatsu,
Toshiya_Nakakura, Jonathan_Devlin, Amit_Hilbuch,
Steve_Anton, Sebastian_Kaebisch, Daniel_Peintner,
Nigel_Megitt, Samira_Hirji, Alan_Bird, Mark_Foltz
Remote
Kazuhiro_Hoya, Lei_Zhai, Mark_Vickers, Yajun_Chen
Regrets
Chair
Chris, Igarashi, Pierre
Scribe
kaz, tidoust
Contents
* [5]Topics
1. [6]Welcome and introduction
2. [7]Hybridcast update
3. [8]Media Timed Events in Hybridcast
4. [9]Media Timed Events Task Force
5. [10]CTA WAVE update
6. [11]Review open issues
7. [12]Frame accuracy synchronization (contd)
8. [13]Professional media workflows on the web
9. [14]Bullet Chatting
10. [15]Joint meeting with Second Screen WG/CG
11. [16]Joint meeting with Timed Text WG
12. [17]Closing and wrap-up
* [18]Summary of Action Items
* [19]Summary of Resolutions
__________________________________________________________
Welcome and introduction
<inserted> scribenick: kaz
Chris: welcome and let's get started
... this is a one-day meeting today
... starting with the general view about the MEIG
... [W3C Code of Ethics]
... [Media&Entertainment IG: Mission]
... apply Web technology in general to media services
... new use cases/requirements to drive the discussion
... [History of Major Initiatives]
... 1. 2011-2014: HTML5 Media Pipeline
... 2. 2011-2017: Adaptive streaming & content protection
... 3. 2017-2019: Media Web App Platform
... [Charter]
... scope covers almost everything
... end-to-end pipeline
... continuous experience
... increasing interactive media
... including games
... [Charter (cont.)]
... Tasks
... identify requirements
... incubation of technical
... review media-related deliverables
... coordinate with other media-related groups, e.g., MPEG,
HbbTV, ...
... internationalization, accessibility, etc., are very
important
... [Work Flow]
... new ideas & issues from Members and SDOs
... use cases, requirments and gap analysis
... but not specs themselves
... because this is an IG
... but some of the results could promote new features for
other WGs
... we're encouraged to work on more and more new features
... [WICG]
... discourse forum there
... get implementer supporter for your idea
... GitHub repo for your proposed feature
... [Contributing to HTML and DOM]
... HTML WG and WHATWG
... [Task Forces]
... 2018-19: Media Timed Events TF
... 2016-17: Cloud Browser API TF (dormant)
... [Monthly conference call]
... 2019/2018
... (list of topics)
... [Activities for 2020]
... whole bunch of topics for Media WG
... potential candidates for adaptation
... MEIG can input use cases and requirements
... [Activities for 2020 (cont)]
... what will be new topics for 2020?
... what would be the main things?
... would like to capture ideas
... in the after noon, we'll have more open discussion
... [Schedule]
... (shows the agenda)
... any additions?
... [Resources]
... various links here
... btw, we have a new co-Chair, Pierre, here
Pierre: involving standard activity, e.g., IETF
... started with HTML WG for W3C
... co-editor of TTML 1.0
... feel free to contact me and Chris (and Igarashi-san)
Chris: would mention that Mark Vickers stepping down as a
co-Chair
... has been leading the group successfully
... really getting the Web as the platform for media
... Mark will continue to participate in the MEIG
Mark: one of the founding co-Chairs
... plan to stay involved as an Invited Expert
... the best source consolidated expertise is media, video and
audio
... before HTML5 media support in the Web
... we've been a good source
... for W3C, WHATWG, Kronos, etc.
... we don't write specs ourselves
... but see what's the priority for media on the Web
... provide expertise
... communicate with media companies, etc., which are not in
the W3C as well
... it takes a lot of work
... glad to help
... a lot of leadership in studio world
... and so on
... have three very strong co-Chairs
... and Pierre is joining
... aiming for HTML5 media 2.0
... this is really a time to better support for media
... Comcast, my company, provides a new rep
John: yes, I'm here
Hybridcast update
[20]slides
[20]
https://www.w3.org/2011/webtv/wiki/images/d/d3/RecentAchievementHybridcast_TPAC20190916.pdf
Ikeo: welcome to Japan!
... would talk about Hybridcast
... [Today's outline]
... recent achievement of hybridcast
... [Deployment status]
... [History of standardization and experiments]
... 2014-2019
... hybridcast connect deployed on some of the TV sets
... [Shipment of Hybridcast receivers]
... number of receivers over 10 million
... [Trial deployment "Hybridcast-Connect"
... what is hybridcast connect?
... new functions/services uses new APIs
... new APIs are experimentally implemented in some of the TV
sets
... brought an example here
... a number of companies are involved
... [Typical Sequence by additional APIs]
... 5 functions
... 1. MediaAvailabilityAPI
... 2. ChannelsInfoAPI
... 3. StartAITAPI
... 4. TaskStatusAPI
... 5. ReceiverStatusAPI
... [Hybridcast Connect demo]
... will show a demo here
... (brings a TV set in front of the screen
... [Hybridcast-Connect demos]
... two demos
... 1 emergency alert
... 2. smooth guidance of catch-up
... [Supporsed use cases demo (1)]
... (Kaz adds webcam to webex)
... embedded buttons here on my PC
... press a button on the PC and get a notification on the
smartphone
David: accessibility consideration?
... would be important
Ikeo: agree
Pierre: API to TV? or smartphone?
Ikeo: pairing between TV and smartphone
Igarashi: depending on the TV's implementation
... possibly implemented as an application on the TV side
... the Hybridcast connect specification itself just specify
the protocol
Yongjun: additional latency?
Igarashi: remote control for playback, etc.?
Ikeo: go back to [Typical Sequence by additional APIs]
... explains the sequence of communication
... using websocket
Yongjun: how much latency there?
Ikeo: TV-set dependent
Igarashi: support all the functions?
Ikeo: some specific functions are supported
Igarashi: arrow keys are supported?
Ikeo: yes
... would like to support all the keys included in the TV
remote
Igarashi: but there are too many buttons
Ikeo: also we need to consider security
... e.g., to avoid unexpected change of volume
... here, all the 5 APIs are implemented based on HTTP
Sudeep: infra-red remote vs this API?
... what kind of values are added?
Pierre: TV implements some specific capabilities
Igarashi: TV vendors have to implement the APIs
Lilin: how many TVs could be controlled?
Ikeo: more than two
... we'd like to handle more than one TVs
... but TV vendors say hundreds of mobiles can't be connected
... maybe 2-3
... TV is used within a local network
... user selects which would be the best one to get connected
David: the system detects the devices available
... are children notified?
Ikeo: application keep it stored in the session information
... the user don't have to mind
Pierre: emergency notification itself is not included in the
protocol
... its separate
Ikeo: right
... these 5 APIs implemented within the device
... so device APIs
... not Web PIs
Chris: looking at secure protocol?
Ikeo: some solution
... we have two devices using some key
Chris: second screen wg works on secure protocol
... so you're aware of that
Ikeo: right
... the problem is HTTPs in local network
... tx for your comments!
... [Supposed use cases demo (2)]
... implemented as a service like Netflix or Amazon Prime Video
... application selects a program from a list
... using just one API
... (select a program on his smartphone)
<MarkVickers> Q: What would the Hybridcast group like from W3C?
New specs? Changes to specs?
<MarkVickers> Thx
Ikeo: launch the HTMl5 app
... using Dash.js
Igarashi: can control playback, forward/backward?
Ikeo: can be done using websocket
<MarkVickers> While it's always interesting to see what other
groups are doing, we have to focus on our goals to drive
changes into W3C and increase adoption of W3C standards outside
of W3C.
Yongjun: device features can be controled?
Ikeo: some of the features
Yongjun: how many subscribers?
Ikeo: In some research, 30% of the TV sets can connect to
internet now and we expect these TV sets to implement this
protocol near future
Chris: what would the Hybridcast group like from W3C?
<scribe> ... new specs, gap analysis?
Ikeo: would like to combine some Hybridcast APIs to W3C
standards
... e.g., playback API, as Igarashi-san mentioned
... between a mobile and a TV
Kaz: kind of like the formerly proposed TV control API?
Ikeo: yeah...
Chris: or something like proposed by the second screen wg?
Ikeo: we have to select web standard APIs
... can't create another APIs ourselves
... that's the second demo
... [Conformance Test and verification]
... [Conformance Test for Hybridcast-Connect]
... IPTV Forum Japan provides hybridcast connect standard
... and also test kit
... this is the overview
... (shows a diagram)
... emulator as the test environment
Chris: cover the Web application?
Ikeo: end-to-end test
... similar to the tests by HbbTV, etc.
... [Service Verification by MIC project and others]
... MIC is Ministry of Internal Affairs and Communications from
the Japanese Government
... service verification with Hybridcast-Connect
... in 2018 19 companies
... in 2019 23 companies
... that's it for Hybridcast update
... thank you!
Chris: what other specific things to address the gaps?
... relationship with the Web platform test, etc.
Ikeo: we require some functions from Web APIs
... TV vendors sometimes want and sometimes not
Chris: ok
... move on to the next topic
Media Timed Events in Hybridcast
Ikeo: [Service Patterns with Hybridcast Connect]
... broadcasters in Japan need trigger message to switch to
broadcast service
... pattern1: from mobile app to broadcasing on TV
... pattern2: from another app in TV to broadcasting
... [Media Timed Events with Hybridcast-Connect]
... JP broadcasters interested in media timed events (MTE)
... same function as the trigger message
... there are two possible choices
... (MTE data in video resource + push emergency alert
notification) to the smartphone
... another option (MTE data in video resource to another app
on TV)
... there are those two possible patterns
Chris: is this emsg?
... in ISO container
... in the DASH format
Ikeo: yes
Igarashi: upper pattern can be realized using mobile API
... but what about the bottom pattern?
... is the TV device at the bottom same as the one on the
right?
Ikeo: yes
... in the case of Android platform, the mechanism is something
like intent
Igarashi: general question about MTE
... unclear why you want to embed events within the video
stream
Ikeo: main reason is the cost of access the message API from
mobile
Igarashi: cost of notification servers
Ikeo: right
... also the accuracy
Igarashi: do we share the same requirements?
Yongjun: which layer to be handled?
... should be fragment or manifest
Igarashi: manifest embedded event?
Ikeo: it depends on the needs
... in case of outbound, MTE might be written in the manifest
... and there would be possible delay
Igarashi: could be updated frequently
Ikeo: related to cost of access transfer
... trade-off of accuracy and cost
... show another demo on MTE
... (select an app on his mobile)
... send a message using hybridcast-connect to the TV
... this is embedded event
... emergency alert shown on the upper-right of the TV
Chris: intended to synchronization of media?
Ikeo: this mechanism just sends an alert
... and the Hybridcast application on the TV can handle how to
display it
... [Usecases of MTE]
... switch broadcasting service from OTT triggered by emergency
message
... super-impose time-dependent metadata, e.g., weather icon
and event information
... new style of ad-insertion on a broadcasting service
... [MediaTimedEvents demos]
... demo implementations
... use case 1: switch live news program on braodcasting
service from OTT service by emergency-warning message
... use case 2: super-impose a weather icon on the Internet
video
[21]slides
[21]
https://www.w3.org/2011/webtv/wiki/images/d/d1/MediaTimedEventsInHybridcast_TPAC20190916.pdf
Chris: what is the requirements?
Ikeo: would like to show the picture on the warning
... but sometimes overlaps with the important content (e.g.,
peoples faces)
Igarashi: depends on the apps
... for some apps, accuracy is not important
Ikeo: we have to consider the accuracy of timing for many cases
... that's all
Chris: tx!
Ikeo: btw, we would like other devices from the app on TV
... during the WoT demo, we'll show home appliance demos
Kaz: at the lunch place, Argos on the 1st floor
... and Wednesday breakout
Chris: excellent
Ikeo: we'd like to use MTE as the basis
[break till 11am]
Media Timed Events Task Force
[22]slides
[22]
https://docs.google.com/presentation/d/1f8LVFY3shrUsksKWLyBVQk3icDN4zEsgV0NX9oqPXNw/edit
Chris: [Topics]
... in-band timed metadata and timed event support
... out-of band timed metadata
... improving synchronization of DOM events triggered on the
media timeline
... also MPEG carriage of Web resources in ISO BMFF
... [History]
... our TF started in 2018
... Giri Mandyam from Qualcomm preesnted work at ATSC and MPEG
on eventing
... published use cases and requirements document early this
year
... [Use cases for timed metadata and in-band events]
... MPEG-DASH specific use cases
... notification to media player
... another use case about getting matrix during playback
... ID3 tags: title, artist, image URLs
... ad insertion cues: SCTE35, SCTE214-1, 2, 3
David: keeping web page in synch with media
... you got slides and talking about the slide
... flip the slide dec and showing
Chris: we have something like that in the explainer
Pierre: we heard another use case in the morning
Chris: multiple contents and multiple events
Pierre: do you know if the cues tied to entire content?
... somebody may remove the trigger
Chris: emsg separately handles
Pierre: can remove part of the content and it's still relevant
Chris: right
... [Recommendations]
... allow web application to subscribe to event streams by
event type
... discussion on type of event
... maybe some concern
... something we can discuss
... also allow web applications to create timed event/timed
metadata cues
... including start time, end time and data payload
Igarashi: in the morning, we had some discussion on in-band
message
... wondering if the current W3C standards support it
... only the scope could be in-band events?
Chris: there are some implementations
... e.g., for HbbTV
... exposing MPD events
... W3C specs don't say anything about type of events
... next, actual triggering
... when cues are parsed from the media container by the UA
... when the current playback position reaches the cue
start/end on the media timeline
... allow cues with unknow end time
... and finally
... improving synchronization (within 20 msec on media
timeline)
David: covers seeking?
... duration of events to be understood
... what would happen if jump-off?
... very hard to handle spike events
Chris: some of that kind of use cases for DASH
... absolutely right
Igarashi: requirement might be can detect that kind of delay
... applications would know about the difference between
specified timing and actual fired timing
... we need to improved the timing (if possible)
... but should identify the gap
... e.g., based on the timestamp
Chris: [Current status]
... almost complete Task Force use cases and requirements
... WICG DataCue explainer in progress
... API spec not started yet
Yongjun: need to revise DASH spec, etc.?
Chris: we need to have discussion about what kind of mechanism
is needed first
... do we ask the UA to give structured data, etc.
... question about how the different formats should match the
need
Mark: the issue is how to map particular data format
... and how to present it
... the reference by HTML5
... need to be updated
... based on the newest MPEG spec
<cpn> [23]https://dev.w3.org/html5/html-sourcing-inband-tracks/
[23] https://dev.w3.org/html5/html-sourcing-inband-tracks/
Mark: maybe another form
Chris: really interesting
... other things reference it
... URL spec
... not really standardized
... definitely right we need handle it
... in more standardized shape
Mark: another question
... data cue was implemented in webkit before HTML5
<cpn> [24]https://www.w3.org/TR/media-frags/
[24] https://www.w3.org/TR/media-frags/
Mark: concern about syntax and semantics
Chris: don't know the answer now
... a session by the Media WG will be held
Mark: sounds like a right place
Chris: [MPEG Carriage of Web Resources in ISO-BMFF Containers]
... saw TAG advice
... since then people working on MPEG as ISO/IEC FDIS 23001-15
David: probably public
Chris: this topic is welcome to the MEIG
David: good to have a workshop including the MEIG, ISO, etc.
... trying to get users of technology at the same time at the
same place
... including security experts
Pierre: what is the use case?
David: two things
... carriage of web pages
... synchronization of media
Pierre: but what would be the actual business cases?
<MarkVickers> FYI, we have previously gotten permission from
MPEG to host MPEG documents on the W3C member-only website. We
could ask for MPEG for permission to host CMAF spec for this
purpose.
Pierre: btw, the community draft is available
Igarashi: benefit of web resource embedded to MPEG
... possibly reduce the cost for the web servers
... could be beneficial
Pierre: the offline case is weird to me
Igarashi: one of the use cases to be addressed
... there are some offline use cases
... for packaged delivery
Chris: [Browser support for DataCue]
... current support by browsers
... edge: HTML 5.1 DataCue attribute ArrayBuffer data;
... chrome: no support
... safari: supported
... firefox: no support
... HbbTV: HTL 5.1 (8 Oct 2015 ED) DataCue with native handling
of player specific events
... [Player support for DASH and HLS events]
... Shaka Player: shaka.Player.EmsgEvent no internal handling
of manifest refresh events.
... (some more examples)
... [Next steps]
... breakout session on Wednesday about "DataCue API and time
marches on in HTML" at 11:00am
... raise issues against WHATWG HTML to propose changes to time
marches on
MarkW: may have people aware of the algorithm
... during the Media breakout
Igarashi: in scope of the Meida WG?
Chris: right
... the Media WG would take on standardization if the direction
is correct
... we started a TF within MEIG and now are looking at WICG
Igarashi: any friction?
Chris: we're looking at the possible API design
Igarashi: what I remember from the previous TPAC
... we were looking at WICG
... but what would be the direction now?
Chris: we don't have enough input
... need more concrete feedback for the possible API
... in JavaScript at the moment
... would be good to have more involvement
... also would be good to have more browser vendors
... need to have wider discussion
... if we proposed issues, that should go to WHATWG
... increasing timing accuracy
... [References]
... (links to resources)
Igarashi: WICG would have their meeting?
Chris: Thu/Fri
... also we have a breakout session ourselves
Igarashi: it's good timing to have discussion with them
... should ask the other participants about opinions as well
... need to get opinions from the MEIG guys
Pierre: when would be our final report available?
... more input needed?
... anybody have any specific objections?
Igarashi: we have not specifically asked the MEIG for opinions
... report itself is about requirements
... it's an IG Note. right?
Chris: yes
Pierre: the report says something is missing and to be added?
... shouldn't say that explicitly?
Chris: solution design to be done by WICG
... our TF could continue editorial changes
... everybody, please join in
CTA WAVE update
(we're delayed by 30mins)
[25]slides
[25]
https://drive.google.com/file/d/1-mAhZe8s2TRDygCW1fPc-aJkvA1yCTH1/view?usp=sharing
John: John Riviello from Comcast
... quick update on CTA WAVE
... [The Web Appliction Video Ecosystem Project]
... aims, focuses, ...
... [Supporting a fragmented OTT world]
... fragmentation impacts content providers and device makers
... [Brief history]
... CEA initiated the GIVE project in 2015
... CEA becomes CTA in Nov. 2015
... [Steering Committee]
... technical WG
... CSTF for content specification
... DPCTF for testable requirements
... HATF for reference application framework
... [WAVE bridges media standards & web standards]
... [Curent WAVE Membership]
... many members
... overlapping with W3C Members
... [What is the Common...]
... [WAVE COntent Spec & Published CMAF Media Profiles]
... [Media Profile Approval]
... profiles are added
... typically updated once a year
... [WAVE Content Specification 2018 AMD 1 - Video Profiles]
... [WAVE Content Spec 2018 AMD 1 - Audio Profiles]
... [WAVE Programs and Live Linear...]
... [Anticipated WAVE Content Spec 2019 Updates]
... [Test SUite: COntent Verification Tool]
... verification content
... shared with DASH-IF conformance validator
... [CSTF - Specification Process]
... annual f2f meeting
... [Links]
... links for resources
... [HATF: HTML5 API...]
... What We Do in the HATF]
... playback audio-video media
... [HATF Work Plan]
... W3C Web Media API CG
... [HTML5 APIs: Reference Platform]
... one content format but multiple devices
... [HATF Specs]
... snapshots
... Web Media API snapshop (WMAS)
... CTA and W3C co-publishing
Igarashi: what do you mean?
John: working on the same document
Igarashi: not WG but CG?
... it's not "W3C Recommendation" but "CG Report"
Francois: fyi, there will be discussion about W3C process
during this week
Alan: part of the plenary on Wednesday
John: [Anticipated Web Media API 2019 Snapshot Updates]
... update to ECMAScript 7
... CSS snapshot 2018
... [HATF Testing Framework]
<MarkVickers> FYI on referencing WAVE specs: ATSC references
the WAVE WMAS as published by CTA, which is referencable. The
W3C version of the WMAS spec, like all CG specs, includes
boilerplate language that it should not be referenced.
John: WMAS Testing Suite Updates]
... [Abstracted Device Playback Model]
... (skips some slides)
... [Spec Highlights and Outline Dec 2018]
... [Promisses in Spec for 2019 and beyond]
... [Test Sutie: RFPs]
... [Q&A]
... questions?
Igarashi: what are you going to talk about "type1 player"?
... any room for W3C standardization?
... if you have any specific requirements, the MEIG can discuss
that
... btw, what is the "Content Model Format"?
Chris: question around testing
... is the work related to the web platform testing?
Pierre: should we put that on the agenda for the afternoon?
all: ok
Review open issues
Chris: we use GitHub to manage issues
... most of the issues will be covered in the afternoon jointly
with the other WGs
... but one specific issue here about frame accurate
synchronization and seeking
Francois: [Related GitHub Issues]
... issue 4, 5, 21
... the main issue is #4 frame accurate seeking of HTML5
MediaElement
... [Categories of Use cases]
... 2 different use cases
... seeking and rendering
... [Main Seeking Use Cases]
... non-linear edition in a browser
... can be cloud-based
... collaborative review
... evidence playback by camera and video
... [Seeking Gaps]
... currentTime is not precise enough to identify individual
frames
... also no way to seek to the next/prev frame in the generic
case
... just matter of time
... when is going to be the next frame
... [Main Rendering Use Cases]
... dynamic content insertion (splicing)
... video overlays
... media playback synchronized with map animations
... synchronization between audio and timed text, e.g., karaoke
... synchonized playback across users/devices
Igarashi: requirements for time seeking?
Francois: this is rather rendering issues
Pierre: sample alignment and duration
... current web platform doesn't allow frame-accurate timing
Francois: [Rendering Gaps]
... currentTime is not precise enough to identify individual
frames
... also timestampOffset is not precise to identify frame
boundaries
... it's hard to track media timeline frame by frame
... in any case there is no mechanism to handle frame accuracy
... also synchronization between video and audio
... if you look at global synchronization
... no way to tie the rendering of a video frame to the local
wall clock
<MarkVickers> Following up on earlier question: It has always
been the intention of WAVE to contribute back to W3C any new
tests and also any changes to the W3C test runner. WAVE
representatives met with the W3C test group at TPAC 2018. There
was an issue opened on April 2, 2019:
[26]https://github.com/web-platform-tests/wpt/issues/16214
There was a PR entered on June 12, 2019:
[27]https://github.com/web-platform-tests/rfcs/pull/23
[26] https://github.com/web-platform-tests/wpt/issues/16214
[27] https://github.com/web-platform-tests/rfcs/pull/23
[lunch till 1:30pm]
Frame accuracy synchronization (contd)
Francois: continue the slides
... [Rendering Gaps that Remain]
... currentTime is not precise enough
... timestampOffset is not precise enough
... three following requirements deleted
... [Seeking Gaps that remain]
... [Next Steps?]
... what do we want to do?
... follow up on MTE recommendations around synchronization?
who?
... wrete a UCR document on frame accurate synch? who?
... feed needs back into WHATWG and Media WG? who?
... different possible groups to bring ideas to
... possibly machine learning group?
... discussion with different people on Wednesday
Chris: production use cases?
Pierre: put together a presentation
... what's happening about a lot professional asset
... have some demo as well
MarkW: the problem is probably how to identify individual frame
... could be ended up with overlaps
Chris: we need rationale
Igarashi: (asks new comers to sing up with the attendees list)
... btw, this proposal includes two different points
... because the difficulty for realization is quite different
... depending on theperformance of the browsers and the
hardware
Francois: maybe write two documents or might be simply continue
discussion
... there are different use cases
... some of them might be out of scope
Pierre: sound and video synchronization is a use case
... not even possible currently
... there is no API for that purpose today
Igarashi: that requirement is related to time seeking
... different from synchronization itself
Pierre: when you say "seeking", it's API level. right?
Igarashi: currently there is no way for requirements to specify
how quickly browsers should behave
Pierre: it's largely implementation-dependent?
Igarashi: yeah
... current time issue and synchronization are different issues
... wonder if any other W3C specs handle that kind of
performance
... how fast browsers are expected to render the data
... we need to talk about performance
Pierre: sounds like you're interested in making contribution :)
Yongjun: need some mechanism to handle frame jumping
... not only at the beginning
... if we care about one case, we may miss another case
Igarashi: ad-insertion seamlessly?
... accuracy of seeking is important for many use cases
... but we should distinguish time seeking accuracy
Pierre: if there is a stream and also another stream starts, in
that case, need for frame accuracy
Igarashi: that's true
Professional media workflows on the web
Pierre: very timely topic
... proposal by MovieLabs
... increasing popularity
... [Web applications are coming to profesisonal media
workflows]
... why?
... web applications have become mainstream
... web platform media capabilities are tantalizingly close
... profesisonal audiovisual assets are moving to the cloud
... [Why move audio-visual assets to the cloud?]
... instead of using UPS
... now available immediately on the cloud
... it's more secure actually
... and of course more efficient
... [Today]
... previsualization, visual effects, grading, editing,
localization, mastering, qualit check, archival, distribution
... [tomorrow]
... all of them will be on the cloud (and can be accessed via
web applications)
... [demo]
... ownzones
... content already on the cloud
... there is an editor here
... (for audio and timed text)
... (going back to the presentation)
... [Some steps of the workflow remain out of reach of web
applications]
... gaps exist in the web platform
... what's missing?
... that's it
Chris: we have many items, so don't want to dive into the
details
Yongjun: as far as I know, people use MSE
Igarashi: video editing using browser
... requirements for rendering related to multiple video
clipping
... handle frames seamlessly
Pierre: we need volunteers
MarkW: not volunteering myself but support the use cases
Scott: folks here might want to consider local content
Igarashi: local content using browser?
Scott: not necessarily on the cloud
... how to handle frame accuracy on the local devices
Pierre: help document the issues?
... take a first step
... somebody needs to take the lead
... this is listing the current issues
Gary: interested
Chris: thank you
Samira: gathering data
... we have a few ideas
... one of them is adding attribute to video tags
... somebody from Google also proposed media container
... my first question is
... would you have any thoughts
... will host a session on Wednesday
Chris: related to content representation
... quite a lot of valuables
... we'll talk about caption later
Andreas: where to standardize 360 video, etc.
... we have a presentation on that later in the afternoon
... also a session on Wednesday
... possibly tomorrow as well
Chris: anybody aware of MPEG format update?
David: whole bunch of work
<tidoust>
[28]https://mpeg.chiariglione.org/standards/mpeg-i/omnidirectio
nal-media-format
[28]
https://mpeg.chiariglione.org/standards/mpeg-i/omnidirectional-media-format
<scottlow> [29]https://mpeg.chiariglione.org/standards/mpeg-i
[29] https://mpeg.chiariglione.org/standards/mpeg-i
<dsinger> [30]https://mpeg.chiariglione.org
[30] https://mpeg.chiariglione.org/
Andreas: TTWG has liaison with MPEG
... but just one part of scenarios
... inband information
... doesn't sort the issues about outband captioning
<Joshue108> There are also accessibility requirements around
how 360 is standardised.
Andreas: possibly discuss that tomorrow?
Samira: possible
... how many content producers, providers, here?
... what block you?
Song: China Mobile
Igarashi: VR content protection?
Samira: can be represented as VR
... magic window scenario
... just wanted to bring this discussion up
Chris: what's the natural home for this discussion?
... first candidate is Timed Text
Samira: just wanted to share the ideas since this is an IG
Josh: There are accessibility requirements if 360 is to be
standardised, around an architecture that will support
accessibiity and multimodal requirements.
Andreas: would like to come back later in the afternoon
... where to do it
... really difficult to find a right place
Chris: related to accessibility
Sudeep: Chair of the Web&Networks IG
... will have our meeting tomorrow
... please drop by
... interested in Media Timed Event as well
... network latency
... very happy to give inputs
Chris: interesting questions
... very close relationship with this group
... having a Web interface
... webrtc stream for multiple different sources
... it is stuff we've been implementing
... not necessarily synchronized with each other
Sudeep: how should we bring back?
Chris: GitHub issues
... also we have monthly IG calls
... have media-related topics
Josh: particular accessibility issue in synch with video stream
Chris: yeah
Josh: bunch of stuff
<Joshue108>
[31]https://www.w3.org/WAI/APA/wiki/Accessible_RTC_Use_Cases
[31] https://www.w3.org/WAI/APA/wiki/Accessible_RTC_Use_Cases
Josh: can put resource on what I'm working on (above)
... related to this group
... different modality channels based on user's preference,
TTS, braille, etc.
(kaz remembers the MMI Architecture and SCXML, which are used
to integrate multiple modality channels :)
Chris: any other issues?
Igarashi: local packaging?
... publishing group is working on packaged media
... playback locally
... on a local storage
... might be with very high resolution of time
Chris: seems we need another gap analysis
<tidoust> [Note the breakout session on Web Packaging planned
on Wednesday:
[32]https://w3c.github.io/tpac-breakouts/sessions.html#wpack]
[32] https://w3c.github.io/tpac-breakouts/sessions.html#wpack
Bullet Chatting
Song: Song Xu from China Mobile
... would give a presentation about bullet chatting
... Michael from Dwango as well
<cpn> [33]https://w3c.github.io/danmaku/index_en.html
[33] https://w3c.github.io/danmaku/index_en.html
[34]proposal
[34] https://w3c.github.io/danmaku/index_en.html
<scribe> scribenick: tidoust
Song: Interactive tool for video broadcasting over the
Internet. Use cases: see reviews of group users. Real-time
interaction, engagement for young generation, to show social
presence.
... Implementation is difficult because you need to compute the
positioning and animation of bullet chatting, rendered in DOM
or Canvas and overlaid on top of the video.
... Strong demand for this type of applications, particularly
in Asia
... Standardization would improve UX, reduce the difficulty in
implementation.
... We suggest to define a standard format for bullet curtain.
... We started an analysis to identify gaps. No specific API
introduced for the time being.
... Bullet chatting is basically floating text over the screen
with four attributes:
... mode, basic properties, timeline, and container (typically
the video)
... [going through Bullet Chatting Proposal document]
... During streaming, two main ways to present: chatting room
or bullet chatting.
... Advantages of bullet chatting display are that there is a
wider display area and it does not require the user to move her
eyes.
... The movement from right to left allows users to read
content quickly (and again without moving her eyes).
... Sometimes, it's not only about comments, it can be text to
improve the feeling of horror videos for instance.
... Also used to share messages in stadiums on a big wall.
Michael: I'm from Dwango. Use cases and requirements for our
current service Niconico.
... Niconico is a streaming Web site launched in 2006. Since
its inception, its unique feature has been its comment system.
... [showing a demo]
... allows to create a user experience.
Pierre: Who specifies at what vertical position the bullet
curtain appears?
... Do you foresee that to be done at the client side?
Song: No, done on the server side
Pierre: So the format has all the positioning information.
Michael: In the current implementation, clients do the
rendering, and they all have the same algorithm, so
deterministic.
Pierre: If things were standardized at W3C, would the
positioning be imposed by the server?
Michael: Currently, we'd like the client to have the ability to
position the comments.
Pierre: So the client receives the comments and decides where
to lay them out.
Igarashi: You want to let the browser do the whole rendering?
Michael: No, the Web application.
... Goal of the standardization is to have a shared format for
bullet curtains, because many providers have a similar comments
system (Niconico, Bilibili, etc.)
Song: First step is to define an interoperability format. If
there is a way to involve the browser vendors, then great,
second step.
MarkW: Browsers would want to know why something cannot be done
in JS.
David: And you could possibly do it with WebVTT / TTML.
Song: For advanced features, there are things that TTML does
not address. Happy to talk with TTML folks though.
Michael: Use cases and requirements level for now. Possible
solutions are still very early stage.
... Bullet curtain allows to create feelings such as sharing
content with friends.
... Comments can be used to improve the video with artwork, or
even to flood the video with comments.
... Comments have become an important part of Niconico's
culture.
... Part of on-demand and live-streaming services of Niconico.
... Comments move right to left across at set times, based on
the media timeline.
Chris: If I pause the video, do the comments pause?
Michael: Yes.
... Comments are clipped to the edge of the player (or to an
arbitrary region).
... When the video loads, comments are loaded from the server
and rendered.
... If a user submits a comment, it appears immediately to the
user, and gets shared to other viewers.
... Seeking to the same time in the same video will have the
same comment appear at the same time and at the same position.
... As if the comments were part of the video, comments scale
with the video in particular.
... Comments can be interactive (e.g. context menu)
MarkW: Layout problem (HTML is good at it), animation problem
(Web Animations), but the thing is Web Animations ties
animations to the wall clock, whereas here animation is tied to
the media clock.
... That may be a useful gap to identify
Chris: Came earlier during Francois' presentation. Tying
non-media content rendering to media timeline.
Igarashi: Some requirements about positioning the subtitles.
... Client decides arbitrary where to position the comments.
Michael: Yes.
Igarashi: Content provider does not care about positioning of
subtitles.
Sangwhan: Aside from Web, do you also want to handle support
for native players?
... That would change perspectives.
Michael: We do have native apps, so we'd be interested with a
solution that covers that space too.
Sangwhan: According to Mark's idea, if it's tied to the
animation timeline in browsers, you're restricting yourself to
Web environment.
Kaz: When I talked to Koizuka-san from Niconico, he mentioned
extension mechanism named "Nico-script", and that mechanism has
capability of specifying style and position of captions. so
that capability could be also considered at some point. maybe
not now, though.
<MarkVickers> I'm not staying connected for the joint meetings.
Have a good TPAC all! -mav
Joint meeting with Second Screen WG/CG
Chris: The Second Screen WG/CG made a lot of progress on the
Open Screen Protocol for discovering, authenticating and
controlling remote displays on the local network.
MarkF: I work for Google. Been involved in Second Screen since
2015. Second screen for the Web is the way we want to enable
Web applications to take advantage of connected
displays/speakers and render different types of content.
... Content can be a full Web page or specific media.
... The Presentation API enables a web page, called the
controller, to request display of an URL on a remote display on
the LAN.
... Example of a photo app that displays the loaded picture on
a large display. You can play media, do gaming, collaboration
tools. Pretty agnostic, but our experience shows that it's
mainly used for media playback.
... The Remote Playback API allows a web page on which there is
a media element to remote the playback of the media element on
a second screen, either through media flinging where the URL to
play gets sent to the remote device, or media remoting where
the media gets streamed to the second screen.
... Both APIs are in Chrome.
... The APIs were designed to take advantage of proprietary
protocols. To get broad adoption, we decided to develop an open
set of protocols so that implementers could all support the
APIs in an interoperable way.
... We hope to converge at the end of the Second Screen F2F
meeting this week to v1.0 of the Open Screen Protocol.
... One use case for the future: enabling Web applications to
generate their own media and present it to a connected display,
e.g. for gaming.
... The Open Screen Protocol supports all sorts of use cases
that we hope to expose to Web applications in the future.
Yongsun: Support of QUIC in smart TVs. UDP is not supported in
some TVs.
Sangwhan: UDP is supported at the kernel level.
MarkF: in our library implementation, we expose UDP but that's
pretty much the same thing as what you get at the system level.
Chris: One of the question that came up in our previous F2F
meeting is around synchronization, e.g. ability to provide
audio description on their device while they are sharing a
media element on a second screen.
... Within that, there is the question of how close the
synchronization needs to be.
... We worked on close synchronization between main screen and
companion device in HbbTV.
MarkF: Does the HbbTV specification rely on clocks?
Chris: Yes, clock synchronization and then the devices can make
adjustments to playback to stay in sync.
MarkF: We need a mechanism for the two sides agree on a wall
clock for presentation.
... If the HbbTV covers all of that, we can have a look for
OSP.
Chris: Yes, it does.
<anssik> Open Screen Protocol issue Requirements for
multi-device timing while streaming
[35]https://github.com/webscreens/openscreenprotocol/issues/195
[35] https://github.com/webscreens/openscreenprotocol/issues/195
Chris: Some implementers have found it difficult to achieve
that level of synchronization. It's not so widely implemented
for now.
... I can provide information on how that has been done.
MarkF: Collaboration between the protocol and the application
levels.
Chris: And also something that exposes the pipeline delays.
MarkF: One of the things that seem very important is the
establishment of a secure communication between devices, which
could have broader implications, such as connected home
scenarios.
... it could be a good foundation for that. Part of the OSP
focus has been on authenticating devices, currently based on
SPAKE2.
... We're not currently focused on enabling one piece of
software to find out attributes of another, for instance who
manufactured it, what does it do.
<anssik> SPAKE2
[36]https://datatracker.ietf.org/doc/draft-irtf-cfrg-spake2/
[36] https://datatracker.ietf.org/doc/draft-irtf-cfrg-spake2/
MarkF: You could take the chapter on authentication and use it
elsewhere.
... We did anticipate that there may be other use cases than
the ones we foresee, so have landed an extensibility mechanism.
Sangwhan: Is there a registry for these capabilities?
MarkF: Yes, it's on GitHub.
... You can be a presentation controller, receiver, send or
receive media, that's all negotiable in the OSP.
Chris: I suspect remote playback of encrypted content is a use
case shared by different members here.
MarkF: The API is pretty much agnostic. At the protocol level,
we haven't tried to add support for messages to exchange to
support encrypted media.
... That seems more to be a use case for the Presentation API
where the application can create and exchange
application-specific message commands.
... Remote playback of encrypted media is closely tied to
credentials, and that's application level.
MarkW: The thing that you don't have here is the streaming
model where the controlling device has the decryption key and
wants to stream the content to the receiver device.
... What happens to the media stream when it reaches the
receiver? Goes to a media element or through JS processing?
Peter: receiver is handling the decoding.
Chris: Is there an IG recommendation that we'd want to make?
MarkW: The most likely model for us for doing this would be to
have a receiving web application that handles the user's
credentials
Chris: That would make the sync issue interesting because it is
then at the application level.
... One of the issues we have with Remote Playback is that we
want to provide a custom UI, which means that we rather want to
use the Presentation API for that.
... Didn't we discuss having a Media element through the
Presentation API that gets automatically synchronized with
local content?
MarkF: I believe that's correct. I don't recall the status of
it. It came up in May 2018, I think.
<anssik> Second Screen May 2019 F2F
[37]https://www.w3.org/wiki/Second_Screen/Meetings/May_2019_F2F
[37] https://www.w3.org/wiki/Second_Screen/Meetings/May_2019_F2F
MarkF: I think we probably agreed that it should be possible.
It probably requires a few tweaks to the protocol so that it
knows that the remoting is part of a shared presentation.
... We discussed whether everything could be done in script.
Same recommendation for synchronization. What you might be
missing is the latency of the media rendering pipeline.
Chris: I have seen implementations that manage to do
synchronized playback across devices through a timing server.
Igarashi: I don't follow the discussion on encrypted media. You
are not going to define how keys are exchanged in the protocol?
MarkF: Someone with more experience on EME might be able to
shed some lights as to what would be required.
... One reason we designed an extension system is that people
interested in new features can propose them, prototype
implementations, and then we can incorporate them in the spec
if all goes fine. We don't have the expertise in the group.
... We're not defining the path for encrypted media from one
device with another. Might work if both devices support HDCP.
... I think there is an open issue in our GitHub about remote
playback and encrypted media.
Igarashi: Arbitrary application message passing is supported?
MarkF: Yes.
... In the spec, you'll see bindings between the API and the
messages exchanged in the protocol.
... For instance, video.remote.prompt() requires exchanges
messages between devices
MarkW: Could the protocol work on TCP?
Peter: You'd have to advertise it differently
Igarashi: [question on security during remote playback]
MarkF: the Remote Playback API does not require the receiver to
be a user agent in the usual sense, it does require the
receiver to support media playback as in the HTML spec.
MarkW: The Presentation API requires the receiver to be able to
render the URL, but the URL could be a non HTTP URL, custom
schemes may be supported instead.
MarkF: The spec defines processing of HTTPS URL, the rest is
undefined.
<anssik> Open Screen Protocol
[38]https://github.com/webscreens/openscreenprotocol/
[38] https://github.com/webscreens/openscreenprotocol/
MarkF: We have a writeup of how the protocol interacts with
custom schemes in the GitHub repo.
Chris: That has been one of the extension mechanisms that we've
been interested in for opening a Web page that has broadcast
capability in HbbTV (perhaps Hybridcast has similar needs)
<anssik> Custom Schemes and Open Screen Protocol
[39]https://github.com/webscreens/openscreenprotocol/blob/gh-pa
ges/schemes.md
[39]
https://github.com/webscreens/openscreenprotocol/blob/gh-pages/schemes.md
[discussion on second screen support in Hybridcast]
MarkF: regarding authentication, we looked at J-PAKE and
request/response challenges but we had memory concerns there so
switched to SPAKE2 following internal discussion with security
experts at Google.
Peter: The protocol allows for more authentication mechanisms
in the future.
... Devices can support their own mechanism.
Igarashi: Co-chair of HTTPS in local network CG, meeting on
Thursday morning. We haven't reached discussion on
authentication. Would be good to align with Open Screen
Protocol.
Sangwhan: Is there a prototype?
MarkF: We recently decided to add streaming to the OSP, which
complicated things. We have a first implementation of
Presentation API commands. No crypto because we've kept
changing that.
... The library is coming. It implements the protocol. It does
not do media rendering, it does not have JS bindings, etc.
<anssik> Open Screen Library implementation
[40]https://chromium.googlesource.com/openscreen/
[40] https://chromium.googlesource.com/openscreen/
Igarashi: If you want to apply the OSP to the broadcast
protocol, we need to consider the case where the remote device
is not a browser. For instance, channel change is done by the
system, not the application.
MarkF: Capabilities like supporting channel tuning is not in
the OSP. If you think that the communication channel needs to
be terminated on channel change, that can be added.
Igarashi: In the case that some arbitrary message protocol is
still necessary, you'd use the Presentation API, but the
receiver may not be a browser agent.
MarkF: seems like something for an extension.
Chris: OK, thank you for the discussion.
MarkF: Mostly, we want input on use cases that we haven't
considered yet. We'd love to get feedback on the extension
mechanism as well.
Pierre: Thank you.
Joint meeting with Timed Text WG
Andreas: We could start with 360 standardization
Nigel: In TTWG, we're in the final stages of rechartering.
... Some things that we're considering such as karaoke.
<Joshue108>
[41]https://www.w3.org/WAI/APA/wiki/Accessible_RTC_Use_Cases
[41] https://www.w3.org/WAI/APA/wiki/Accessible_RTC_Use_Cases
Nigel: Quick agenda bashing, any topic you'd like to cover?
Josh: accessibility use cases? See accessible RTC use cases
document
Chris: TTML and MSE?
Nigel: Yes, opinions about exposing TextTracks from MSE.
<Joshue108> apologises for throwing a curve ball to Nigel, I'm
here for the XR bit but think this doc may still be useful as
an FYI
Andreas: Focus the discussion of the day on standardization of
360 subtitles. Most of the stuff comes from an EU research
project.
... To make it short, there have been extensive user tests. For
captions, main requirement is to have subtitles that are always
in the field of view. It's enough to have them on a 2D plane,
no need to have them positioned in 3D.
... There should be some indication of where the audio source
is positioned.
... Of course, you also need features present in TTML,
TTML-IMSC profile being a good example.
... [demo of an application to test subtitles positioning]
... Lots of activity starting last year at TPAC. We started
with a discussion in the Immersive Web CG. Then discussion
within the TTWG, Media & Entertainment IG.
... In the end, we realized we needed more people from
immersive and browser vendors.
... We wrote a proposal to be discussed in the WICG.
... There has been no comment on the WICG forum yet, so
question is how do we proceed?
... Two additional activities worth noting. A colleague from
Google proposed the creation of an Immersive Caption Community
Group, and XR accessibility W3C workshop in November.
... There is awareness that something needs to be done.
... Hard to get enough resources to get started though.
... How to get time and resources from implementors?
<Joshue108> Inclusive Design for Immersive Web Standards W3C
Workshop Seattle Nov 5-6
<Joshue108>
[42]https://www.w3.org/2019/08/inclusive-xr-workshop/
[42] https://www.w3.org/2019/08/inclusive-xr-workshop/
Andreas: Everything is evolving, nothing really fixed.
... Is it really a web platform topic?
... Important to know when to stop if there is not enough
interest.
... Apart from which group should deal with it, the question is
also where does this solution fit?
... Authoring environments (Unity, Unreal), Web applications,
WebXR API (linked to OpenXR) and 360 / XR device
... How to follow-up? I thought WICG would be the right place,
but if there is not enough place, there is still the question
of whether that's the right place. Not sure about Immersive
Caption CG since it does not exist yet.
... TTWG is the right group but we need more expertise from the
XR world.
... Another solution is to continue the work in a "private"
repository.
<Zakim> nigel, you wanted to ask what is the state of
documentation of the requirements right now
Nigel: What is the state of documentation in terms of the
requirements?
... Describing positioning in 3D space, can I do it with audio?
Andreas: There are documented user tests, as part of an
European project deliverable.
Nigel: I was thinking about requirements documentation. What is
the problem that you're trying to solve, user needs.
Samira: Who was the person who started the Immersive Caption
Community Group?
Andreas: Christopher Patnoe at Google
Samira: OK. Another comment is that WebXR is becoming more
stable.
Andreas: Yes, the question for me is where should this go.
... The WebXR API does not know anything about what's inside
the WebGL right now.
Chris: Is all that's needed a delivery format and then some
library can place that in the immersive environment?
Igarashi: Do we need to extend APIs in the browser to support
this?
<igarashi> -q
Andreas: OMAF defines a way to multiplex IMSC subtitles with
MP4, but then it's all bound to that content format. Not sure
it's sufficient for interoperability scenarios.
<igarashi> +q
Kaz: wondering about the possible relationship with Web
Francois: WebVMT is about tracks positioned on a map, not in
360 videos.
Kaz: However, there is a possibility of mapping the positioned
map to the 360 screen and then put caption on that screen
Andreas: It would be an option to have a subtitle format, but
burning captions in a frame does not provide good user
experience.
Josh: Looking at things from an accessibility perspective. APA
would seem a good group to talk to.
Andreas: We talked a lot with Judy, Janina and so on.
<Joshue108> [43]https://www.w3.org/WAI/APA/wiki/Xaur_draft
[43] https://www.w3.org/WAI/APA/wiki/Xaur_draft
Josh: We created a list of requirements for XR in APA.
<samira> IW group is also discussing dom overlays so this is
another option for subtitles
Pierre: How many people in this group doing 360 videos and XR
content?
... One possibility is that this group is not the best group to
get feedback from.
Andreas: I don't know, that's what all groups say ;)
... We need a critical mass to do it.
Pierre: People that build apps for Oculus, are they around?
Andreas: I spoke to some of them. They always say that they
don't provide subtitles.
... Some discussion in Khronos with Unity and Epic.
... I talked with Immersive Web folks. We'll talk about that on
Wednesday 11:00 during Samira's breakout session.
... The issue that we have is that there is not endless time to
deal with it. The project is running out. It stops next year.
To push a standard, it will take 2-3 more years.
<Joshue108> There are very few testing with people with
disabilities in this space so this is very interesting.
Igarashi: From a content production perspective, I'm interested
in a format, but not sure about browser support for this.
[44]https://github.com/immersive-web/dom-overlays
[44] https://github.com/immersive-web/dom-overlays
Francois: Not clear to me what you want to be standardized. DOM
overlays could be one building block.
Andreas: Yes, DOM overlays may be a good way forward to render
captioning thatn burning things in WebGL.
<Zakim> nigel, you wanted to wonder what the smallest thing is
that we need to standardise first - is it a syntax for
expressing a 3D location?
<Joshue108> +1 to Nigel
Nigel: Same point. Do we have agreement that it's about a
syntax for expressing a 3D location?
Andreas: Actually, that's not what we need, since we want it to
appear on a 2D plane, that is what the users want.
... We need a way to indicate where in the 3D space the audio
source is coming from.
Gary: So you need some positioning in 3D to make that possible.
Andreas: Define a good container is another issue.
Josh: in the User requirements document I showed you, we took a
modular approach.
... This architecture does not exist yet.
<Joshue108> [45]https://www.w3.org/WAI/APA/wiki/Media_in_XR
[45] https://www.w3.org/WAI/APA/wiki/Media_in_XR
Josh: We're also looking at Media requirements in XR. Not
vetted by the APA WG yet.
Andreas: Lots of 360 content for the time being, and a lot of
it without captioning.
<Joshue108> s/module approach/modular approach
Gary: WebVTT update. I joined TTWG half a year ago. Trying to
get WebVTT to progress. One of the big thing is an
implementation report exists right now.
... Something like 6-7 issues with it.
<atai> Link to 360 subtitle requirement
[46]https://github.com/immersive-web/proposals/issues/40
[46] https://github.com/immersive-web/proposals/issues/40
Gary: Basically, we're looking at features implemented in
browsers and in VLC. Then identify features at risk, and
possibly remove them to get a V1 out.
... Then hopefully convince browser vendors to implement the
features that we may remove.
<gkatsev> [47]WebVTT Implementation Report
[47] https://www.w3.org/wiki/TimedText/WebVTT_Implementation_Report
Glenn: Any SMPTE spec that includes 3d positions of audio
sources?
Nigel: That's a good question.
... One of the things we're doing around TTML2 is adding new
functionality in extension modules. We're trying to constrain
the core, and then provide the rest in extensions.
... There are a few ones that are ongoing.
... [details extensions]
... Right now, audio/video comes to MSE but not text.
MarkW: My personal position is that things should be
symmetrical across media types.
... At least in our application, we prefer to do the rendering
of text tracks ourselves.
... It would be advantageous in which the browser is aware of
text tracks.
Nigel: You said my sentiment much better than I could.
Gregg: I would argue that we don't want to render them
ourselves, but we still want to control the rendering with our
styles.
MarkW: Yes, we want to have enough control of the rendering,
but we could offload the rendering to the browser, that would
be great.
Nigel: It's been hard to get statistics about user
customization, or people that play back content with captions.
MarkW: In terms of rendering, you would still want the site to
control enabling/disabling.
<atai> +1
Gary: We shouldn't try to do the same thing twice. If there's
more support to do the new generic TextTrack thing, then that's
good.
Pierre: Two different questions: any objection to enabling
symmetry in MSE? Are you going to use it?
MarkW: First question is whether people think that could be
harmful.
Nigel: OK, I just wanted to raise it to get feedback.
[No concerns expressed regarding question on whether people
think that could be harmful]
Josh: About accessibility in WebRTC use cases, challenge of
synchronizing some of these things together when switching to a
different modality. That's one.
Nigel: It would make sense to talk about live contribution to
see where that fits. How does live contributions actually work,
what's the mental model?
... Alright, I think we covered all topics.
Closing and wrap-up
Chris: Thinking about Media Timed Events, some editorial work.
Planned discussion on DataCue. Around bullet chatting, more
conversation will happen this week.
... Some possibility to go to Timed Text WG.
Nigel: It feels to me that this IG could be the best place to
give guidance for that if there's no clarity in TTWG on Friday
about that.
Andreas: Can you explain again how you want to proceed?
... Draft published in the Chinese IG, what would the ideal
next step be?
Song: Initially, contributors were from China. Now that
NicoNico is engaged in discussions, work could go to TTWG, or
perhaps in another group.
... We want the use cases to be approved by the IG, afterwards
we'd like to push standardization work on identified gaps.
... Within the next few weeks, we'll have a last version of the
use cases.
Andreas: OK, so this week would be a good opportunity to decide
where this should go.
Chris: We had a lot of discussion around synchronization today.
Frame accurate rendering.
... Ability to seek accurately within videos.
... Some interest to follow-up, although no one volunteers.
... The media production use case that Pierre presented would
be a good perspective to address this.
Pierre: With an action on Gary to follow up with Garrett Singer
on that.
Chris: Secure communications between devices, we heard
interesting stuff from Hybridcast and HTTPS in local network,
and Second Screen. Interesting set of approaches that could be
compared.
... Seems like a good fit for HTTPS in local network CG
discussions.
... Clearly the immersive captioning is interesting, but not
sure what next step in this group should be. Maybe the
Immersive Captioning CG could be the right forum.
... We talked about 360 videos. That's something that the IG
could follow on. We have liaison with MPEG. Unless you feel
that immersive group would be a better home.
Samira: Possibly. At this point, I'm gathering input.
Chris: Finally, there's the timed text in MSE proposal. Would
that sit in TTWG?
MarkW: It would be in scope of the Media WG.
Chris: Have I missed anything from the summary?
Pierre: One encouragement for you to clarify the scope in Media
Timed Events.
Chris: And also possibly make more specific recommendations.
Pierre: I think it helps to have something concrete.
Chris: OK, I think that's everything, thank for your presence
today!
Summary of Action Items
Summary of Resolutions
[End of minutes]
__________________________________________________________
Minutes manually created (not a transcript), formatted by
David Booth's [48]scribe.perl version 1.154 ([49]CVS log)
$Date: 2019/10/04 17:02:15 $
[48] http://dev.w3.org/cvsweb/~checkout~/2002/scribe/scribedoc.htm
[49] http://dev.w3.org/cvsweb/2002/scribe/
Received on Tuesday, 8 October 2019 07:21:49 UTC