[tpac-f2f] minutes from the f2f meeting in Lisbon - 19-20 September 2016

Hi group,

Sorry for the delay but it seems I've not yet sent the f2f minutes
out to the group list yet.

The minutes from the f2f meeting during TPAC 2016 in Lisbon are
available at:
  https://www.w3.org/2016/09/19-20-multimodal-minutes.html

also as text below.

Thanks a lot for taking notes, Debbie and Helena!

Kazuyuki

---
   [1]W3C

      [1] http://www.w3.org/

                               - DRAFT -

       Multimodal Interaction Working Group F2F Meeting in Lisbon
                             19-20 Sep 2016

   [2]group photo

      [2] https://www.w3.org/public-photos/multimodal/DSC_0393.JPG

   See also: IRC logs from [3]Day 1 and [4]Day 2

      [3] http://www.w3.org/2016/09/19-multimodal-irc
      [4] http://www.w3.org/2016/09/20-multimodal-irc

Attendees

   Present
          Debbie_Dahl(Invited_Expert),
          Dan_Burnett(Invited_Expert),
          Helena_Rodriguez(Invited_Expert),
          Branimir_Angelov(Wacom;Guest), Kaz_Ashimura(W3C),
          Sebastian_Kaebisch(Siemens), Ningxin_Hu(Intel),
          Uday_Davuluru(RWE), Andrei_Ciortea(Universite_de_Lyon)

   Regrets
   Chair
          Debbie

   Scribe
          Debbie, Helena, Kaz

Contents

     * [5]Topics
         1. [6]MMI Discovery
         2. [7]Joint meeting with the COGA TF
         3. [8]Joint meeting with the WoT IG
     * [9]Summary of Action Items
     * [10]Summary of Resolutions
     __________________________________________________________

   [11]photo

     [11] https://www.w3.org/public-photos/multimodal/DSC_0190.JPG

MMI Discovery

   <helena> [12]http://w3c.github.io/mmi-discovery/vocabulary.html

     [12] http://w3c.github.io/mmi-discovery/vocabulary.html

   <helena> editor's version:

   <helena>
   [13]https://github.com/w3c/mmi-discovery/blob/gh-pages/vocabula
   ry.html

     [13] https://github.com/w3c/mmi-discovery/blob/gh-pages/vocabulary.html

   <ddahl> scribe: ddahl

   (adding reference to example of "greeting service" in a "smart
   environment" to vocabulary

   debbie: this will be useful for cognitive accessibility
   ... also useful for web of things
   ... we need some text relating state management and vocabulary

   <helena> [14]http://www.w3.org/TR/mmi-discovery/

     [14] http://www.w3.org/TR/mmi-discovery/

   <helena> The document explaining the process of discovery and
   the needs on changing the architecture and the vocabulary

   helena: we can add something from the original use cases note

   debbie: first overview, then state management, then vocabulary
   should be in the merged document

   <helena>
   [15]http://www.w3.org/TR/2016/WD-mmi-mc-discovery-20160411/

     [15] http://www.w3.org/TR/2016/WD-mmi-mc-discovery-20160411/

   debbie: vocabulary should all be grounded in standards if
   possible

   helena: we can add metadata in operations to say that the
   component understands, for example, emotionML or BML (behavior
   markup language)

   the "input" and "output" sections under "behavior" list the MMI
   architecture events used for sending input and output

   scribe: as opposed to overall context management
   ... because as an MMI MC, we can assume those events are used

   behavior markup language
   [16]http://www.mindmakers.org/projects/bml-1-0/wiki#BML-10-Stan
   dard

     [16] http://www.mindmakers.org/projects/bml-1-0/wiki#BML-10-Standard

   we should remove the key from the POST because authentication
   is out of scope for MMI

   <helena> [17]http://w3c.github.io/mmi-discovery/vocabulary.html

     [17] http://w3c.github.io/mmi-discovery/vocabulary.html

   updating [18]http://w3c.github.io/mmi-discovery/vocabulary.html
   with discussion

     [18] http://w3c.github.io/mmi-discovery/vocabulary.html

   updating the 'state handling' wd to include the vocabulary
   document

   state handling [19]https://www.w3.org/TR/mmi-mc-discovery/

     [19] https://www.w3.org/TR/mmi-mc-discovery/

   <helena>
   [20]http://w3c.github.io/mmi-discovery/Discovery_and_Registrati
   on.html

     [20] http://w3c.github.io/mmi-discovery/Discovery_and_Registration.html

   (taking a break)

   (back from break)

   helena: we should change the order of the title to
   "Registration and Discovery" because registration comes first
   ... vocabulary should come first, then state handling
   ... we can get some material from the use cases document
   ... it will be hard to edit in real time, we should make a todo
   list
   ... section 4 can be the same
   ... it will be the component used for registration

   <helena> @todolist: Keep the 4 section that describes the
   Resources Manager

   <helena> --- put a 6 section with the vocabulary and rename the
   following sections with the new numbers

   <ddahl_> scribe:ddahl_

   <helena>
   [21]http://w3c.github.io/mmi-discovery/Discovery_and_Registrati
   on.html

     [21] http://w3c.github.io/mmi-discovery/Discovery_and_Registration.html

   helena: leave short examples of vocabulary in place
   ... not all at the end
   ... we could have one running example of the face animation
   throughout
   ... the face synthesizer won't need a lot of states, it will be
   a very concrete example
   ... should state at the beginning that we'll talk about two
   things, registration and then discovery/monitoring
   ... registration -- what does a component have to do to
   describe itself. will just include a paragraph here and then
   point to the use cases document
   ... this is very related to the Internet of Things
   ... we can leverage the IoT registriations like UPnP, but it
   will be necessary to translate because they don't talk about
   modalities

   debbie: will still need information for UI's

   helena: still can get this from different types of devices
   ... maybe WoT can provide an API to give some information like
   the name of the service
   ... the address

   debbie: how would that work for example, a rice cooker?

   helena: visual modality is an LED, haptic modality (button) and
   thermostat
   ... the description has to say that
   ... the only thing we can control is the button on/off
   ... a fancier rice cooker could have more controls

   debbie: what if the rice cooker can be controlled by voice?

   helena: this is a different service
   ... a rice cooker that can be controlled by an app would be
   cognitive
   ... the operation would be the same but the modalities would be
   different (haptic and cognitive)

   debbie: you could turn on the rice cooker or adjust the
   firmness of the rice
   ... those would be different operations

   helena: you have risotto, basmatii, sushi rice -- they might be
   different

   debbie: it would good to talk about leveraging IoT descriptions
   and translation

   we can decide what we should say about WoT and
   Discovery/Registration in our document after we meet with WoT
   tomorrow

   dan: also need a description of the API for data control
   [22]photo

     [22] https://www.w3.org/public-photos/multimodal/DSC_0194.JPG

   [ Day 1 adjourned ]
     __________________________________________________________

Day 2

   <helena> question to kaz: in the table "functions of object" in
   the template, what are Functions of Objects Track and
   Accountability for?
   [23]photo

     [23] https://www.w3.org/public-photos/multimodal/DSC_0196.JPG

   <ddahl> (updating use case document)
   [24]photo       [25]photo

     [24] https://www.w3.org/public-photos/multimodal/DSC_0195.JPG
     [25] https://www.w3.org/public-photos/multimodal/DSC_0197.JPG

   <scribe> ACTION: kaz to clarify in the table "functions of
   object" what Fuctions of Objects Track and Accountability are
   for [recorded in
   [26]http://www.w3.org/2016/09/20-multimodal-minutes.html#action
   01]

     [26] http://www.w3.org/2016/09/20-multimodal-minutes.html#action01]

   <trackbot> Created ACTION-455 - Clarify in the table "functions
   of object" what fuctions of objects track and accountability
   are for [on Kazuyuki Ashimura - due 2016-09-27].

Joint meeting with the Cognitive Accessibility TF

   (visited the Cognitive Accessibility TF's room for the joint
   meeting)

   -> [27]Cognitive Accessibility TF minutes

     [27] https://www.w3.org/2016/09/20-coga-minutes.html

Joint meeting with the WoT IG

   [28]Thing Description examples from the WoT Current Practice
   document

     [28]
http://w3c.github.io/wot/current-practices/wot-practices.html#td-examples

   ddahl: explains what the MMI Architecture is like
   ... the Interaction Manager works with various Modality
   Components like speech, emotion recognition, ink capture
   ... communicate with user for WoT
   ... there is another component named Resource Manager
   ... responsible to maintain the state of the resources
   ... manage their capabilities
   ... shows another diagram
   ... with a rice cooker
   [29]photo

     [29] https://www.w3.org/public-photos/multimodal/DSC_0193.JPG

   helena: UPnP itself doesn't provide device capability
   information
   ... we were thinking the WoT framework should provide that kind
   of information

   seb: explains Thing Description
   ... data types coming from RDF and schema
   ... but how to handle the range, etc.
   ... maybe there are some ways to rely on
   ... but not fixed yet
   ... if the type relies on XML, we can use its schema
   ... JSON Schema is not yet standardized, though
   ... we'll have discussion on Schema.org too

   ddahl: how can a developer get the value like temperature?

   seba: can access the entry point specified by the "uris"
   property
   ... and get the value by the "hrefs"
   ... index.html is the entry point at the uris URL

   ddahl: besides temp, we can use more than one properties?

   helena: each property is atomic?

   seba: we could allow multiple protocols
   ... and multiple hrefs

   helena: question about the end point for multiple protocols

   seba: properties are handled by GET
   ... actions are handled by POST
   ... there is already complains on the usage of "uris"
   ... maybe better to explicitly specify the method in addition
   to protocols within "uris"

   kaz: also it would be even clearer to use different endpoint
   file names for different protocols instead of reuse
   "index.html" for all possible protocols

   seba: right
   ... discussion still ongoing
   ... this is JSON-LD notation based on RDF
   ... can show a bigger example

   helena: do you have any taxonomy to describe things?
   ... about how "things" could be described in addition to
   "devices"
   ... e.g., flower

   (discussion on ontology)

   seba: we're not working on ontology
   ... different kinds of ontologies could be used with the WoT
   framework

   helena: if I want to use some ontology with TD, where I can
   specify that within TD?

   seba: within the @context part
   ... and can use the prefix within the below @context

   ddahl: unit change?
   ... how the manufacture could specify units?

   (after some more joint discussion)

   kaz: MMI should be a promising framework for the expected
   advanced user interface for the Web of Things world.
   ... These days, the group has been working on new use cases for
   that purpose.
   [30]photo       [31]photo

     [30] https://www.w3.org/public-photos/multimodal/DSC_0198.JPG
     [31] https://www.w3.org/public-photos/multimodal/DSC_0194.JPG

   [ Meeting adjourned ]

Summary of Action Items

   [NEW] ACTION: kaz to clarify in the table "functions of object"
   what Fuctions of Objects Track and Accountability are for
   [recorded in
   [32]http://www.w3.org/2016/09/20-multimodal-minutes.html#action
   01]

     [32] http://www.w3.org/2016/09/20-multimodal-minutes.html#action01

Summary of Resolutions

   [End of minutes]
     __________________________________________________________


    Minutes formatted by David Booth's [33]scribe.perl version
    1.144 ([34]CVS log)
    $Date: 2016/10/10 23:40:04 $

     [33] http://dev.w3.org/cvsweb/~checkout~/2002/scribe/scribedoc.htm
     [34] http://dev.w3.org/cvsweb/2002/scribe/



-- 
Kaz Ashimura, W3C Staff Contact for Auto, WoT, TV, MMI and Geo
Tel: +81 3 3516 2504

Received on Monday, 10 October 2016 23:44:16 UTC