W3C home > Mailing lists > Public > public-esw-thes@w3.org > January 2014

Re: SKOS Quality Checkers

From: Osma Suominen <osma.suominen@helsinki.fi>
Date: Thu, 16 Jan 2014 17:23:22 +0200
Message-ID: <20140116172322.Horde.ZO98XbrCbqLF0n3_eenxUg1@webmail-basic.helsinki.fi>
To: Vladimir Alexiev <vladimir.alexiev@ontotext.com>, "'Christian Mader'" <c.mader@semantic-web.at>
Cc: public-esw-thes@w3.org

Hi Vladimir!

It's great to hear that you're interested in iur tools!

Quick replies to your questions:

The largest vocabs we analyzed using both qSKOS ans Skosify were over  
1GB (IIRC RAMEAU was 1,5GB) so they should work. Naturally you must  
install locally. It might take an hour or two and a few GB of RAM (8  
should be enough with Skosify, if you install rdflib 4.x).

As for the extensions, you'll have to try what happens. Most vocabs we  
analyzed had some custom extensions.

Skosify supports all input formats that rdflib can parse, so e.g.  
rdf/xml, turtle, n-triples should all work. But you may hit parser  
bugs in some cases - if that happens you can reserialize with Jena  
rdfcat or rapper from Redland.

I'm very interested to hear about your results!

Best regards

Lainaus Vladimir Alexiev <vladimir.alexiev@ontotext.com>:

>> Online Checker: http://qskos.poolparty.biz/
>> qSKOS: https://github.com/cmader/qSKOS/
>> http://demo.seco.tkk.fi/skosify/skosify
> Hi!
> It's really nice to have announcements about 2 validation services  
> in the same day :-)
> Maybe the fastest way to learn about them is this joint paper?
> Osma Suominen and Christian Mader: Assessing and Improving the  
> Quality of SKOS Vocabularies. Journal on Data Semantics, 2013.
> http://www.seco.tkk.fi/publications/2013/suominen-mader-skosquality.pdf
> I think we'll try them out on Getty AAT, which could catch some  
> errors, and may provide useful feedback for the tools.
> But:
> - We use a lot more than SKOS:
>   We use SKOS, SKOS-XL, ISO 25964, DC, DCT, BIBO, FOAF, PROV and a  
> custom ontology (GVP = Getty Vocabulary Program)
> - the file is fairly large: 1Gb (gzips to 63Mb in 10s).
>   We use NTriples because Turtle 1.0 doesn't support numeric  
> localnames (e.g. aat:300012345; Turtle 1.1 does).
> Have you tried the tools with such large files?
> Guess there's no option but to install locally (can't use the online  
> services)?
> What input formats do the tools support?
> Do you have some advice on how to approach this?
> Cheers!
> --
> Vladimir Alexiev, PhD, PMP
> Lead, Data and Ontology Management Group
> Ontotext Corp, www.ontotext.com
> Sirma Group Holding, www.sirma.com
> Email: vladimir.alexiev@ontotext.com, skype:valexiev1
> Mobile: +359 888 568 132, SMS: 359888568132@sms.mtel.net
> Landline: +359 (988) 106 084, Fax: +359 (2) 975 3226
> Calendar: https://www.google.com/calendar/embed?src=vladimir%40sirma.bg

Osma Suominen
Information Systems Specialist
National Library of Finland
P.O. Box 26 (Teollisuuskatu 23)
Tel. +358 50 3199529
Received on Thursday, 16 January 2014 15:24:09 UTC

This archive was generated by hypermail 2.3.1 : Wednesday, 2 March 2016 13:32:19 UTC