- From: Jeanne Spellman <jspellman@spellmanconsulting.com>
- Date: Mon, 13 Jul 2020 10:20:34 -0400
- To: Silver Task Force <public-silver@w3.org>
- Message-ID: <eb5388fd-ebf3-e222-22cf-41e4ae80a47e@spellmanconsulting.com>
== Summary ==
* No meeting next week because of XR Access Symposium
* We want to do more outreach to other groups working in this area and
share our User Needs document
* First draft of Functional Outcomes. We disagreed about the
editorial format, but not the concepts
o 1) Auditory information, including speech and key sound
effects, are translated into alternative formats (e.g. captions)
so media can be consumed when sound is unavailable or limited
o 2) Auditory meta-information, including sound directionality, is
conveyed to the viewer to that contextual information is
available when sound is unavailable or limited
o 3) Captions and caption meta-data are capable of being presented
in alternative methods (e.g. second screen) to make the
information more accessible when visual access is unavailable
o 4) Customisation of caption style and position is available to
support users that would benefit from tailored presentation options
o 5) The amount of time that a given caption (and associated
meta-data) spends on screen can be personalised in order to give
additional time to locate the sound that is being presented
* ChrisP provided links to some research from BBC on visual
presentation of captions. He will invite an expert from BBC to our
next meeting on the 27 July.
== Log of Meeting ==
(scribe didn't engage the Minutes bot)
Meeting: Silver XR Subgroup
chair: MikeCrabb
Topic: No meeting next week
MC: Last week we discussed the XR Access Symposium
I am registered for that
https://xraccess.org/symposium/
the symposium and that day 2 session
MC: No meeting next week so that people can attend the Symposium
... Bill Davidson-Curtis is part of the Guidelines & Policies group. He
encouraged us to attend.
<michaelcrabb> https://www.rit.edu/directory/wadnet-wendy-dannels
... Also Wendy Donalds in RIT
... she is the other lead for Guidelines & Policies
Topic: Adding people to this meeting
MC: any progress on making contact with Immersive Captioning?
JS: We also need to include Jason White of Research Questions Task Force
(RQTF)
<CharlesHall> I am inviting other industry people to join the Inclusive
Design for the Immersive Web Community Group – getting ready to register
to have a meeting in TPAC
Topic: Functional Outcomes
<michaelcrabb> 1) Auditory information, including speech and key sound
effects, are translated into alternative formats (e.g. captions) so
media can be consumed when sound is unavailable or limited
<michaelcrabb> 2) Auditory meta-information, including sound
directionality, is conveyed to the viewer to that contextual information
is available when sound is unavailable or limited
<michaelcrabb> 3) Captions and caption meta-data are capable of being
presented in alternative methods (e.g. second screen) to make the
information more accessible when visual access is unavailable
<michaelcrabb> 4) Customisation of caption style and position is
available to support users that would benefit from tailored presentation
options
<michaelcrabb> 5) The amount of time that a given caption (and
associated meta-data) spends on screen can be personalised in order to
give additional time to locate the sound that is being presented
<CharlesHall> this editorial format seems closer to success criteria or
acceptance criteria, that {x} must {y}.
<michaelcrabb>
https://docs.google.com/document/d/1gfYAiV2Z-FA_kEHYlLV32J8ClNEGPxRgSIohu3gUHEA/edit#
<michaelcrabb> When sound is unavailable or limited auditory information
including speech and key sound effects are translated into alternative
formats so that users can understand content.
<CharlesHall> the draft definition of Functional Outcome that is in the
Functional Needs work: A statement that describes a singular objective
of a user has been met – usually in the context of a task or overall
goal – that may need to name or cite a functional need.
<michaelcrabb>
https://w3c.github.io/silver/subgroups/xr/captioning/index.html
<michaelcrabb> User Needs:
https://w3c.github.io/silver/subgroups/xr/captioning/xr-captioning-user-needs.html
JS: What about Minimize Photosensitive Seizure Triggers annd Usage with
Limited Cognition?
MC: Limited cognition is included in #5, and Photosensitive is #4.
MC: What isn't included is Privacy. That is important when
autocaptioning speech to text
... Our group is now using Microsoft Teams, where everything said is
captured in a searchable transcript. Handy, but privacy concerns.
1) We need captions
2) We need meta data of sound effects
3) Second screen
4) Customization of captions
5) Amount of time
MC: I think we need to wait on Functional Outcomes until we get a
recommendation about the editorial format
Topic: Outreach to BBC Blue Room
CP: I have reached out to Spencer Marsden from BBC in Blue Room. He
creates XR content for Blue Room. Captioning is part of what they do.
[discussion of some contacts with Blue Room]
JS: Suggest that you share the User Needs document with them and ask for
their input.
CP: I will invite him to the meeting on the 27th
<Crispy__>
https://www.bbc.co.uk/rd/blog/2017-03-subtitles-360-video-virtual-reality
<Crispy__>
https://www.bbc.co.uk/rd/blog/2014-10-tvx2014-short-paper-enhancing-subtitles
<Crispy__>
https://www.bbc.co.uk/rd/blog/2018-01-accessibility-object-based-media
MC: There is a lot of work there on positioning subtitles
Received on Monday, 13 July 2020 14:20:49 UTC