- From: Jun Zhao <jun.zhao@zoo.ox.ac.uk>
- Date: Wed, 12 Aug 2009 08:45:11 +0100
- To: Hugh Glaser <hg@ecs.soton.ac.uk>
- CC: Aldo Bucchi <aldo.bucchi@gmail.com>, Kingsley Idehen <kidehen@openlinksw.com>, Leigh Dodds <leigh.dodds@talis.com>, "public-lod@w3.org" <public-lod@w3.org>, Anja Jentzsch <anja@anjeve.de>
Hi Hugh, Thanks for championing voiD here:) I would love to avoid the manual editing too. As Leigh said, editing esw wiki is not the best experience you could have. Do you have a voiD description creator/editor you can share with the community? Hugh Glaser wrote: > Please no! Not another manual entry system. > I had already decided I just haven't got the time to manually maintain this constantly changing set of numbers, so would not be responding to the request to update. > (In fact, the number of different places that a good LD citizen has to put their data into the esw wiki is really rather high.) > Last time Anja was kind enough to put a lot of effort into processing the graphviz for us to generate the numbers, but this is not the way to do it. > In our case, we have 39 different stores, with linkages between them and to others outside. > There are therefore 504 numbers to represent the linkage, although they don't all meet a threshold. > For details of the linkage in rkbexplorer see pictures at http://www.rkbexplorer.com/linkage/ or query http://void.rkbexplorer.com/ . > And these figures are constantly changing, as the system identifies more - there can be more than 1000 a day. > > If any more work is to be put into generating this picture, it really should be from voiD descriptions, which we already make available for all our datasets. > And for those who want to do it by hand, a simple system to allow them to specify the linkage using voiD would get the entry into a format for the voiD processor to use (I'm happy to host the data if need be). > Or Aldo's system could generate its RDF using the voiD ontology, thus providing the manual entry system? > > I know we have been here before, and almost got to the voiD processor thing:- please can we try again? Sure, this will be an interesting experiment. Regards, Jun > > Best > Hugh > > On 11/08/2009 19:00, "Aldo Bucchi" <aldo.bucchi@gmail.com> wrote: > > Hi, > > On Aug 11, 2009, at 13:46, Kingsley Idehen <kidehen@openlinksw.com> > wrote: > >> Leigh Dodds wrote: >>> Hi, >>> >>> I've just added several new datasets to the Statistics page that >>> weren't previously listed. Its not really a great user experience >>> editing the wiki markup and manually adding up the figures. >>> >>> So, thinking out loud, I'm wondering whether it might be more >>> appropriate to use a Google spreadsheet and one of their submission >>> forms for the purposes of collectively the data. A little manual >>> editing to remove duplicates might make managing this data a little >>> more easier. Especially as there are also pages that separately list >>> the available SPARQL endpoints and RDF dumps. >>> >>> I'm sure we could create something much better using Void, etc but >>> for >>> now, maybe using a slightly better tool would give us a little more >>> progress? It'd be a snip to dump out the Google Spreadsheet data >>> programmatically too, which'd be another improvement on the current >>> situation. >>> >>> What does everyone else think? >>> >> Nice Idea! Especially as Google Spreadsheet to RDF is just about >> RDFizers for the Google Spreadsheet API :-) > > Hehe. I have this in my todo (literally). A website that exposes a > google spreadsheet as SPARQL endpoint. Internally we use it as UI to > quickly create config files et Al. > But It will remain in my todo forever...;) > > Kingsley, this could be sponged. The trick is that the spreadsheet > must have an accompanying page/sheet/book with metadata (the NS or > explicit URIs for cols). > >> Kingsley >>> Cheers, >>> >>> L. >>> >>> 2009/8/7 Jun Zhao <jun.zhao@zoo.ox.ac.uk>: >>> >>>> Dear all, >>>> >>>> We are planning to produce an updated data cloud diagram based on >>>> the >>>> dataset information on the esw wiki page: >>>> http://esw.w3.org/topic/TaskForces/CommunityProjects/LinkingOpenData/DataSets/Statistics >>>> >>>> If you have not published your dataset there yet and you would >>>> like your >>>> dataset to be included, can you please add your dataset there? >>>> >>>> If you have an entry there for your dataset already, can you >>>> please update >>>> information about your dataset on the wiki? >>>> >>>> If you cannot edit the wiki page any more because the recent >>>> update of esw >>>> wiki editing policy, you can send the information to me or Anja, >>>> who is >>>> cc'ed. We can update it for you. >>>> >>>> If you know your friends have dataset on the wiki, but are not on >>>> the >>>> mailing list, can you please kindly forward this email to them? We >>>> would >>>> like to get the data cloud as up-to-date as possible. >>>> >>>> For this release, we will use the above wiki page as the information >>>> gathering point. We do apologize if you have published information >>>> about >>>> your dataset on other web pages and this request would mean extra >>>> work for >>>> you. >>>> >>>> Many thanks for your contributions! >>>> >>>> Kindest regards, >>>> >>>> Jun >>>> >>>> >>>> ______________________________________________________________________ > > >>>> This email has been scanned by the MessageLabs Email Security >>>> System. >>>> For more information please visit http://www.messagelabs.com/email >>>> ______________________________________________________________________ > > >>>> >>> >>> >>> >> >> -- >> >> >> Regards, >> >> Kingsley Idehen Weblog: http://www.openlinksw.com/blog/~kidehen >> President & CEO OpenLink Software Web: http://www.openlinksw.com >> >> >> >> >> > >
Received on Wednesday, 12 August 2009 07:45:50 UTC