W3C home > Mailing lists > Public > public-lod@w3.org > August 2009

Re: [HELP] Can you please update information about your dataset?

From: Jun Zhao <jun.zhao@zoo.ox.ac.uk>
Date: Wed, 12 Aug 2009 08:45:11 +0100
Message-ID: <4A827307.4000504@zoo.ox.ac.uk>
To: Hugh Glaser <hg@ecs.soton.ac.uk>
CC: Aldo Bucchi <aldo.bucchi@gmail.com>, Kingsley Idehen <kidehen@openlinksw.com>, Leigh Dodds <leigh.dodds@talis.com>, "public-lod@w3.org" <public-lod@w3.org>, Anja Jentzsch <anja@anjeve.de>
Hi Hugh,

Thanks for championing voiD here:)

I would love to avoid the manual editing too. As Leigh said, editing esw 
wiki is not the best experience you could have.

Do you have a voiD description creator/editor you can share with the 
community?

Hugh Glaser wrote:
> Please no! Not another manual entry system.
> I had already decided I just haven't got the time to manually maintain this constantly changing set of numbers, so would not be responding to the request to update.
> (In fact, the number of different places that a good LD citizen has to put their data into the esw wiki is really rather high.)
> Last time Anja was kind enough to put a lot of effort into processing the graphviz for us to generate the numbers, but this is not the way to do it.
> In our case, we have 39 different stores, with linkages between them and to others outside.
> There are therefore 504 numbers to represent the linkage, although they don't all meet a threshold.
> For details of the linkage in rkbexplorer see pictures at http://www.rkbexplorer.com/linkage/ or query http://void.rkbexplorer.com/ .
> And these figures are constantly changing, as the system identifies more - there can be more than 1000 a day.
> 
> If any more work is to be put into generating this picture, it really should be from voiD descriptions, which we already make available for all our datasets.
> And for those who want to do it by hand, a simple system to allow them to specify the linkage using voiD would get the entry into a format for the voiD processor to use (I'm happy to host the data if need be).
> Or Aldo's system could generate its RDF using the voiD ontology, thus providing the manual entry system?
> 
> I know we have been here before, and almost got to the voiD processor thing:- please can we try again?

Sure, this will be an interesting experiment.

Regards,

Jun

> 
> Best
> Hugh
> 
> On 11/08/2009 19:00, "Aldo Bucchi" <aldo.bucchi@gmail.com> wrote:
> 
> Hi,
> 
> On Aug 11, 2009, at 13:46, Kingsley Idehen <kidehen@openlinksw.com>
> wrote:
> 
>> Leigh Dodds wrote:
>>> Hi,
>>>
>>> I've just added several new datasets to the Statistics page that
>>> weren't previously listed. Its not really a great user experience
>>> editing the wiki markup and manually adding up the figures.
>>>
>>> So, thinking out loud, I'm wondering whether it might be more
>>> appropriate to use a Google spreadsheet and one of their submission
>>> forms for the purposes of collectively the data. A little manual
>>> editing to remove duplicates might make managing this data a little
>>> more easier. Especially as there are also pages that separately list
>>> the available SPARQL endpoints and RDF dumps.
>>>
>>> I'm sure we could create something much better using Void, etc but
>>> for
>>> now, maybe using a slightly better tool would give us a little more
>>> progress? It'd be a snip to dump out the Google Spreadsheet data
>>> programmatically too, which'd be another improvement on the current
>>> situation.
>>>
>>> What does everyone else think?
>>>
>> Nice Idea! Especially as Google Spreadsheet to RDF is just about
>> RDFizers for the Google Spreadsheet API :-)
> 
> Hehe. I have this in my todo (literally). A website that exposes a
> google spreadsheet as SPARQL endpoint. Internally we use it as UI to
> quickly create config files et Al.
> But It will remain in my todo forever...;)
> 
> Kingsley, this could be sponged. The trick is that the spreadsheet
> must have an accompanying page/sheet/book with metadata (the NS or
> explicit URIs for cols).
> 
>> Kingsley
>>> Cheers,
>>>
>>> L.
>>>
>>> 2009/8/7 Jun Zhao <jun.zhao@zoo.ox.ac.uk>:
>>>
>>>> Dear all,
>>>>
>>>> We are planning to produce an updated data cloud diagram based on
>>>> the
>>>> dataset information on the esw wiki page:
>>>> http://esw.w3.org/topic/TaskForces/CommunityProjects/LinkingOpenData/DataSets/Statistics
>>>>
>>>> If you have not published your dataset there yet and you would
>>>> like your
>>>> dataset to be included, can you please add your dataset there?
>>>>
>>>> If you have an entry there for your dataset already, can you
>>>> please update
>>>> information about your dataset on the wiki?
>>>>
>>>> If you cannot edit the wiki page any more because the recent
>>>> update of esw
>>>> wiki editing policy, you can send the information to me or Anja,
>>>> who is
>>>> cc'ed. We can update it for you.
>>>>
>>>> If you know your friends have dataset on the wiki, but are not on
>>>> the
>>>> mailing list, can you please kindly forward this email to them? We
>>>> would
>>>> like to get the data cloud as up-to-date as possible.
>>>>
>>>> For this release, we will use the above wiki page as the information
>>>> gathering point. We do apologize if you have published information
>>>> about
>>>> your dataset on other web pages and this request would mean extra
>>>> work for
>>>> you.
>>>>
>>>> Many thanks for your contributions!
>>>>
>>>> Kindest regards,
>>>>
>>>> Jun
>>>>
>>>>
>>>> ______________________________________________________________________
> 
> 
>>>> This email has been scanned by the MessageLabs Email Security
>>>> System.
>>>> For more information please visit http://www.messagelabs.com/email
>>>> ______________________________________________________________________
> 
> 
>>>>
>>>
>>>
>>>
>>
>> --
>>
>>
>> Regards,
>>
>> Kingsley Idehen          Weblog: http://www.openlinksw.com/blog/~kidehen
>> President & CEO OpenLink Software     Web: http://www.openlinksw.com
>>
>>
>>
>>
>>
> 
> 
Received on Wednesday, 12 August 2009 07:45:50 UTC

This archive was generated by hypermail 2.3.1 : Sunday, 31 March 2013 14:24:23 UTC