[emma] Conversational Technologies Implementation Reports for EMMA 1.0

Conversational Technologies strongly supports the Extensible
MultiModal Annotation 1.0 (EMMA) standard. By providing a standardized
yet extensible and flexible basis for representing user input, we
believe EMMA has tremendous potential for making possible a wide
variety of innovative multimodal applications. In particular, EMMA
provides strong support for applications based on user inputs in human
language in many modalities, including speech, text and handwriting as
well as visual modalities such as sign languages.  EMMA also
supports composite multimodal interactions in which several user
inputs in two or more modalities are integrated to represent a single
user intent.

The Conversational Technologies EMMA implementations are used in
tutorials on commercial applications of natural language processing
and spoken dialog systems.  We report on two implementations. The
first is an EMMA producer (NLWorkbench) which is used to illustrate
statistical and grammar-based semantic analysis of speech and text
inputs. The second implementation is an EMMA consumer, specifically a
viewer for EMMA documents. The viewer can be used in the classroom to
simplify examination of EMMA results as well as potentially in
commercial applications for debugging spoken dialog systems. In
addition, the viewer could also become the basis of an editor which
would support such applications as human annotation of EMMA documents
to be used as input to machine learning applications. For most of the
EMMA structural elements the viewer simply provides a tree structure
mirroring the XML markup. The most useful aspects of the viewer are
probably the graphical representation for EMMA lattices, the ability
to see timestamps as standard dates and the computed durations from
EMMA timestamps. The two implementations will be made available in the
near future as open source software.

Deborah Dahl, Conversational Technologies 

Received on Monday, 14 April 2008 18:33:40 UTC