- From: Felix Sasaki <fsasaki@w3.org>
- Date: Wed, 17 Sep 2008 15:48:17 +0900
- To: Véronique Malaisé <vmalaise@few.vu.nl>
- CC: public-media-annotation@w3.org
Just for relating to issues ...
issue 6077, 6079
Véronique Malaisé さんは書きました:
>
> Sorry about the late fulfillment of my task, I had a writer's blank...
> and I am not really sure yet whether this document is what was
> expected, but I made a try and I'm waiting for your comments! I have
> had a look at the IPTC Standard - Photo Metadata 2008 [1] to draw a
> set of requirements for a MM description ontology. And I have also a
> question for the list: do we consider cataloging information too or
> only "pure content" description?
>
> The document [1] is issued by the International Press
> Telecommunications Council and is the result of a larger
> collaboration; it "specifies metadata properties intended to be used
> primarily but not exclusively with photos". More specifically, "IPTC
> Photo Metadata provides data about photographs and the values can be
> processed by software. Each individual metadata entity is called a
> property and they are grouped into Administrative, Descriptive and
> Rights Related properties." These metadata could be applied to
> describe Multimedia documents too, and some links between different
> vocabularies are made: the metadata are described in natural language
> and show possible links with the “G2-Standard” (see [2] for example)
> and XMP [3] representation format.
>
> As for the link with different vocabularies: for instance, the Title
> property aligns with the Dublin Core "Title" element and the
> properties that have the mention (legacy) should be filled in by
> keywords from different controlled vocabularies.
>
> This set of metadata is aimed primarily at journalists, which explains
> some of the modeling choices. It is, in my opinion, a good starting
> point (amongst others) for listing mandatory description/metadata
> items, nevertheless it contains a number of drawbacks for a generic
> image/multimedia description scheme:
>
> - Ambiguous modeling decisions: the “Keyword” property is supposed to
> get a free text value, and not keyword value as expected ("Keywords to
> express the subject of the content. Keywords may be free text and
> don't have to be taken from a controlled vocabulary."), whereas the
> "Subject Code" field has to be filled with controlled vocabulary from
> the IPTC Subject NewsCodes [4].
>
> - Redundant (and thus ambiguous) modeling decisions: the metadata set
> contains a Title, Header, Caption field that all describe the content
> of the image, but that should/can all be different: it is hard to make
> the distinction between these if you are not one of the expert users
> the Specification is aiming at. In a generic multimedia annotation
> ontology, we could make a selection between these and decide to align
> either with all of these fields (and find a way to define their
> semantics precisely), but most likely only with one subset.
>
> - Lack of relationships between the fields: there are some content
> description fields like Events, Location, Person, Object or Artwork
> Shown on the image, but one image, and moreover one Multimedia
> document, contains often more than one event, person, location;
> multiple Events etc can be specified with this description model, but
> to get satisfactory answers to precise queries, or to be able to
> disambiguate between different documents (particularly relevant in
> large homogeneous document collections), a formal relationship between
> the event, person and location has to be made. For example if a
> picture is about two Heads of State shaking hands at a Summit,
> attended by other Heads of State, an explicit relationship has to be
> made between the ones who are shaking hands and the event “shaking
> hands”.
>
> The StructuredAnnotation of MPEG-7 (see [5] and example below) enables
> to explicit such a relationship; more genrally, I think that an
> annotation system based on graphs explicating relationships between
> the Who/What/When/Where/Why/How would improve browsing and searching
> in Multimedia documents collections.
>
> Example of StructuredAnnotation, taken from [5].
>
> <StructuredAnnotation>
>
> <Who>
>
> <Name xml:lang="en">Zinedine Zidane</Name>
>
> </Who>
>
> <WhatAction>
>
> <Name xml:lang="en">Zinedine Zidane scoring against England.</Name>
>
> </WhatAction>
>
> </StructuredAnnotation>
>
> The NewsML ontology [6], associated with Named Graphs, could also
> enable this type of links. I think that the possibility of such graphs
> should be present in a multimedia annotation schema, to enable as
> precise annotations as possible; the relationship between the
> different metadata elements (person/event/location) could be derived
> automatically in some cases (from text or context in the flow/still
> image), so having the possibility to integrate this context in an
> annotation would bring an added value, in my opinion. And I would be
> very interested to know what you think about this point!
>
> [1]
> http://www.iptc.org/std/photometadata/2008/specification/IPTC-PhotoMetadata-2008_2.pdf
>
>
> [2] http://www.newsml.org/pages/
>
> [3] http://www.adobe.com/products/xmp/
>
> [4] http://www.iptc.org/NewsCodes/
>
> [5] http://www.w3.org/2005/Incubator/mmsem/XGR-mpeg7/
>
> [6] http://homepages.cwi.nl/~troncy/research.html
>
>
Received on Wednesday, 17 September 2008 06:48:55 UTC