- From: Philipp Cimiano <cimiano@cit-ec.uni-bielefeld.de>
- Date: Thu, 13 Mar 2014 11:06:09 +0100
- To: Armando Stellato <stellato@info.uniroma2.it>, 'John McCrae' <jmccrae@cit-ec.uni-bielefeld.de>
- CC: 'Manuel Fiorelli' <fiorelli@info.uniroma2.it>, "public-ontolex@w3.org" <public-ontolex@w3.org>
- Message-ID: <53218311.2070200@cit-ec.uni-bielefeld.de>
Armando, all, yes, I fully agree with you on the views expressed below. Therefore, I am suggesting the following: 1) every lexicon should state with respect to which ontologies it provides references to, this is I guess metadata about the lexicon, right? Could be included in lime thus. Let's call this property lime:hasReferencestoOntology, with domain ontolex:Lexicon and range an ontology URI (which can be a void:Dataset as well) 2) To determine language coverage per ontological resources of the ontologies indicated under the property "hasReferencestoOntology", one could run SPARQL Queries over the ontologies to get the number of classes, properties, individuals, and then run SPARQL queries over the lexicalization or lexicon to determine the average number of lexical entries in a given language in the lexicon. This is a two-stage querying architecture so to speak. Doing the queries in a join is also possible but requires either federation or some physical integration of the files. 3) Given the above, it seems thus reasonable to materialize the results of the queries and include the results in the lexicon and/or lexicalization. One problem I see is certainly that things could get out of date. So for example, assume that the ontologies that a lexicon references get changed. Then in principle the numbers would have to get updated. One could of course assign a timestamp to the metadata, which would to some extent provide a principled solution to the update problem in the sense "I might not be up-to-date, so the least I can do is make explicit the timepoint of the last update". What do you think? Philipp. Am 08.03.14 05:06, schrieb Armando Stellato: > Dear Philipp, John, all, > > Yes, absolutely agree that we should provide the SPARQL constructs, as > an operational way to express their semantics. I think we already have > them, as Manuel has written them in a LIME exporter component (we > should just readjust the output according to the structure we want to > build..see previous emails). > Regarding the fact that the metadata should be obtainable from the > data present in the content, in general I agree, and the general idea > is that those sort of linksets could be subset of different datasets > and thus appear in the most appropriate void file (from the case of > three distinct datasets up to the case in which ontology, lexicon and > lexicalization collapse into one dataset, yet with the lexicalization > being identifiable at the metadata level as a specific subset of the > whole). > Yet, to strictly respect your rule, a problematic case is: providing > that it is important to express the lexical coverage for resource R > (for Rs belonging to an ontology) and that this requires to be able to > enumerate the instances of R, this amounts to say that, in the case of > a separated lexicon and ontology, the metadata should be in the void > of the lexicon, and thus, to iterate Rs, the lexicon should > necessarily owl:import the ontology, or the data from onto would not > be available and this would break your rule for which the metadata > should be always calculable in terms of available data. > > I think our view here is pretty simple: > we consider the triad: <ontology, lexicon, lexicalization>. In this > triad, ontology and lexicon may happen to be independent, but the > lexicalization will always have (at least conceptually) a known > dependency to the content of the other two (I.e. whoever wrote the > lexicalization, did it with the target ontology and chosen lexicon at > hand). Thus I think it should be in any case legal that the metadata > about the lexicalization is able to tell which percentage of (all) the > Rs from the ontology is covered by the lexicalization. > > Would that work? > > Cheers, > > Armando > > > > > ------------------------------------------------------------------------ > Da: Philipp Cimiano <mailto:cimiano@cit-ec.uni-bielefeld.de> > Inviato: 07/03/2014 22.43 > A: Armando Stellato <mailto:stellato@info.uniroma2.it>; 'John McCrae' > <mailto:jmccrae@cit-ec.uni-bielefeld.de> > Cc: 'Manuel Fiorelli' <mailto:fiorelli@info.uniroma2.it>; > public-ontolex@w3.org <mailto:public-ontolex@w3.org> > Oggetto: Re: LIME proposal for the OntoLex W3C Community Group > > Dear Armando, all, > > second email on the metadata, referring in particular to the > aggregating properties. > > Many of the properties that we are proposing in the metadata module > are aggregating properties: number of lexical entries, average number > of lexical entries etc. > > We sort of agreed that these are computed locally for the dataset in > question without consulting external lexica etc. right? > > The values for most of these values could be calculated using SPARQ > construct statements it seems in the sense that some information is > aggregated and added as explicit value of some lime property. Fair > enough, it saves people the effort to run the SPARQL queries over the > dataset themselves, making this information readily accessible. > > However, in order to properly document the semantics of the lime > poperties we introduce, would it not be feasible to indicate a SPARQL > construct query that computes the property value? In that sense we > would clearly define the semantics of these metadata properties. > > What do you think? > > Philipp. > > > Am 06.03.14 20:17, schrieb Armando Stellato: >> >> Dear Philipp and John, >> >> no need to say sorry, you are coordinating a whole community group, >> we cannot say the same on our side, yet we are no quicker than you in >> replying :D >> >> You raise an important point, the solution of which actually raises >> up an interesting opportunity for other important aspects of at the >> level of the web architecture of Ontolex. >> >> Before we delve further into the details, let us ask one more question: >> >> What is the relationship between ontologies and the lexica, is it 1:n >> (an ontology may have multiple lexica) or m:n (as before, plus the >> same lexicon may be connected to multiple ontologies) ?. A strictly >> related question is: “is a lexicon built specifically for an ontology?”. >> >> Having ported WordNet in Ontolex should already give the answer to >> that (WordNet exists a-priori from any ontology, and thus it should >> be one example in favor of the m:n hypothesis, though we may think of >> a Lexicon as something importing WordNet and extending it for being a >> lexicon for a given ontology). >> >> In case the m:n hypothesis is confirmed, we should think about some >> form of binding, as a third object implementing the connection >> between an independent lexicon and an ontology. >> >> I think I already asked something related to that when I had some >> doubts about how to deal with compound terms: if a lexicon exists >> independently, it will probably not contain some compounds needed to >> describe resources of the ontology, so we cannot assume these should >> be always available (at the time of my question, I remember I was >> told: “for things like “red car”, you should foresee a dedicated >> entry in the lexicon, though it can then be decomposed through the >> decomposition module”, thus implying that the lexicon has to exist >> FOR a given ontology. >> >> Probably I’m missing something here, but I think these are >> fundamental aspects which should be made clear in the wiki pages >> about the overall architecture and the usage of the model. >> >> Ok, sorry for the long introduction, but how you will see, it is >> related to our topic…we however maybe managed to handle this >> independently of the above. So, back to the topic… >> >> …Our model relates to a void file, but this file could be, for >> instance, not the void file of an ontology, but the void file of >> (something similar to) a void:linkset which binds a lexicon to an >> ontology. To cover also the need you express at the end of the email, >> we could propose the following changes: >> >> <!--[if !supportLists]-->1)<!--[endif]-->A lime:lexicon property, >> >> <!--[if !supportLists]-->a.<!--[endif]-->domain: >> lime:LanguageCoverage (the class obviously) >> >> <!--[if !supportLists]-->b.<!--[endif]-->range: void:Dataset (or >> an appropriate subclass lime:Lexicon to define a dataset which >> contains linguistic expressions for some dataset. Note that a >> void:Dataset containing both conceptual and lexical info would be the >> lexicon of itself! >> >> <!--[if !supportLists]-->2)<!--[endif]-->lime:lexicalModel (old >> linguisticModel, moved to having domain set to languageConverage) >> >> so we could have a structure like that: >> >> void:Dataset --lime:languageCoverage--> lime:LanguageCoverage >> --lime:lexicon--> void:Dataset >> >> --lime:lexicalModel--> (rdfs:, skos:, skosxl:, ontolex: ) >> >> --lime:resourceCoverage--> (usual stat info) >> >> But then, we would have another issue…what is a lexicon? If a lexicon >> is something independent of the “enrichment” of an ontology with >> respect to a language, and lives on its own, then, here, in our case, >> we are more interested in knowing the third element we were >> mentioning above, that is, the “man in the middle” providing links >> between a conceptual resource and a lexicon. Thus, with just a >> terminological change, (lexicon --> lexicalization), and relying on >> the fact that this representation delegates to the lexicalization the >> pointers to the lexicon: >> >> void:Dataset --lime:languageCoverage--> lime:Lan >> > > [il messaggio originale non è incluso] -- Prof. Dr. Philipp Cimiano Phone: +49 521 106 12249 Fax: +49 521 106 12412 Mail: cimiano@cit-ec.uni-bielefeld.de Forschungsbau Intelligente Systeme (FBIIS) Raum 2.307 Universität Bielefeld Inspiration 1 33619 Bielefeld
Received on Thursday, 13 March 2014 10:06:38 UTC