W3C home > Mailing lists > Public > public-lod@w3.org > March 2010

Re: Crowdsourcing request: Google People Finder Data as RDF

From: Bill Roberts <bill@swirrl.com>
Date: Thu, 4 Mar 2010 18:12:43 +0000
Cc: Linked Data community <public-lod@w3.org>
Message-Id: <569B56BA-0657-4DFC-B180-AA09B897FEE5@swirrl.com>
To: Aldo Bucchi <aldo.bucchi@gmail.com>
Hi Aldo - I'd like to help, but I see you posted your mail a few hours ago.  Do you have updated information on what still needs done?  Do you have a wiki or similar to coordinate volunteer programming efforts?

Regards

Bill


On 4 Mar 2010, at 14:06, Aldo Bucchi wrote:

> Hi,
> 
> As most of you heard things were a bit shaky down here in Chile. We
> have some requests and hope you guys can help. This is a moment to
> prove what we always boast about: that Linked Data can solve real
> problems.
> 
> Google provides a prople finder service
> (http://chilepersonfinder.appspot.com/) which is right now
> centralizing some ( but not all ) of the missing people data. This
> service is OK but it lacks some features plus we need to integrate
> with other sources to perform analysis and aid our rescue teams /
> alleviate families.
> 
> This is serious matter but it is indeed taken a bti lightly by
> existing software. ( there is a tradeoff between the amount of
> structure you can impose and ease of use in the front-line ).
> 
> What we would love to have is a way to access all feeds from
> <http://chilepersonfinder.appspot.com/> as RDF
> 
> We already have some databases operating on these feeds, but we're
> still far away a clean solution because of its loose structure ( take
> a look and you'll see what I mean ).
> 
> Who wants to take a shot at this?
> 
> Requirements.
> - Take all feeds originating from <http://chilepersonfinder.appspot.com/>
> - Generate an initial RDF dump ( big TTL file )
> - Generate Incremental RDF dumps every hour
> 
> The transfromation should do its best guess at the ideal data
> structure and try not to loose granularity but shield us a bit from
> this feed based model.
> 
> We then take care of downloading this, integrating with other systems,
> further processing, geocoding, etc.
> 
> There's a lot of work to do and the more we can outsource, the bettter.
> 
> On Friday ( tomorrow ) there will be the first nation-wide
> announcement of our search platform and we expect lots of people to
> use our services. So this is something really urgent and really,
> really important for those who need it.
> 
> Ah. Volunteers are moving all this data into a Virtuoso instance that
> will also have more stuff. It will be available soon at
> http://opendata.cl/ so stay tuned.
> 
> We really hope we had something like DBpedia in place by now, it would
> make all this much easier. But now is the time.
> Guys, the tsunami casualties could have been avoided it was all about
> mis-information.
> Same goes for relief efforts. They are not optimal and this is all
> about data in the end.
> 
> I know you know how valuable data is. But it is now that you can
> really make your point! Triple by Triple.
> 
> Thanks!
> A
> 
> -- 
> Aldo Bucchi
> skype:aldo.bucchi
> http://www.univrz.com/
> http://aldobucchi.com/
> 
> PRIVILEGED AND CONFIDENTIAL INFORMATION
> This message is only for the use of the individual or entity to which it is
> addressed and may contain information that is privileged and confidential. If
> you are not the intended recipient, please do not distribute or copy this
> communication, by e-mail or otherwise. Instead, please notify us immediately by
> return e-mail.
> 
Received on Thursday, 4 March 2010 18:13:15 UTC

This archive was generated by hypermail 2.3.1 : Sunday, 31 March 2013 14:24:25 UTC