RE: Potential Home for LOD Data Sets

Another goal would be to allow this pipeline to extend full circle back to
Wikipedia so that users and agents can pass corrections and new content back
to Wikipedia for review and inclusion in future release without editing the
wiki directly (we need to protect our watershed).  Is there another thread
that addresses this somewhere?

-steve

-----Original Message-----
From: Steve Judkins [mailto:steve@wisdomnets.com] 
Sent: Monday, March 23, 2009 2:40 PM
To: 'Kingsley Idehen'; 'Hugh Glaser'
Cc: 'public-lod@w3.org'
Subject: RE: Potential Home for LOD Data Sets

It seems like this has the potential to become a nice collaborative
production pipeline. It would be nice to have a feed for data updates, so we
can fire up our EC2 instance when the data has been processed and packaged
by the providers we are interested in.  For example, if Openlink wants to
fire up their AMI to processes the raw dumps from
http://wiki.dbpedia.org/Downloads32 into this cloud storage, we can wait
until a virtuoso ready package has been produced before we update.  As more
agents get involved in processing the data, this will allow for more
automation notifications of updated dumps or SPARQL endpoints.

-Steve

-----Original Message-----
From: public-lod-request@w3.org [mailto:public-lod-request@w3.org] On Behalf
Of Kingsley Idehen
Sent: Thursday, December 04, 2008 9:20 PM
To: Hugh Glaser
Cc: public-lod@w3.org
Subject: Re: Potential Home for LOD Data Sets


Hugh Glaser wrote:
> Thanks for the swift response!
> I'm still puzzled - sorry to be slow.
> http://aws.amazon.com/publicdatasets/#2
> Says:
> Amazon EC2 customers can access this data by creating their own personal
Amazon EBS volumes, using the public data set snapshots as a starting point.
They can then access, modify and perform computation on these volumes
directly using their Amazon EC2 instances and just pay for the compute and
storage resources that they use.
>   
> Does this not mean it costs me money on my EC2 account? Or is there some
other way of accessing the data? Or am I looking at the wrong bit?
>   
Okay, I see what I overlooked: the cost of paying for an AMI that mounts 
these EBS volumes, even though Amazon is charging $0.00 for uploading 
these huge amounts of data where it would usually charge.

So to conclude, using the loaded data sets isn't free, but I think we 
have to be somewhat appreciative of a value here, right? Amazon is 
providing a service that is ultimately pegged to usage (utility model), 
and the usage comes down to value associated with that scarce resource 
called time.
> Ie Can you give me a clue how to get at the data without using my credit
card please? :-)
>   
You can't you will need someone to build an EC2 service for you and eat 
the costs on your behalf. Of course such a service isn't impossible in a 
"Numerati" [1] economy, but we aren't quite there yet, need the Linked 
Data Web in place first :-)

Links:

1. http://tinyurl.com/64gsan

Kingsley
> Best
> Hugh
>
> On 05/12/2008 02:28, "Kingsley Idehen" <kidehen@openlinksw.com> wrote:
>
>
>
> Hugh Glaser wrote:
>   
>> Exciting stuff, Kingsley.
>> I'm not quite sure I have worked out how I might use it though.
>> The page says that hosting data is clearly free, but I can't see how to
get at it without paying for it as an EC2 customer.
>> Is this right?
>> Cheers
>>
>>     
> Hugh,
>
> No, shouldn't cost anything if the LOD data sets are hosted in this
> particular location :-)
>
>
> Kingsley
>   
>> Hugh
>>
>>
>> On 01/12/2008 15:30, "Kingsley Idehen" <kidehen@openlinksw.com> wrote:
>>
>>
>>
>> All,
>>
>> Please see: <http://aws.amazon.com/publicdatasets/> ; potentially the
>> final destination of all published RDF archives from the LOD cloud.
>>
>> I've already made a request on behalf of LOD, but additional requests
>> from the community will accelerate the general comprehension and
>> awareness at Amazon.
>>
>> Once the data sets are available from Amazon, database constructions
>> costs will be significantly alleviated.
>>
>> We have DBpedia reconstruction down to 1.5 hrs (or less) based on
>> Virtuoso's in-built integration with Amazon S3 for backup and
>> restoration etc..  We could get the reconstruction of the entire LOD
>> cloud down to some interesting numbers once all the data is situated in
>> an Amazon data center.
>>
>>
>> --
>>
>>
>> Regards,
>>
>> Kingsley Idehen       Weblog: http://www.openlinksw.com/blog/~kidehen
>> President & CEO
>> OpenLink Software     Web: http://www.openlinksw.com
>>
>>
>>
>>
>>
>>
>>
>>
>>     
>
>
> --
>
>
> Regards,
>
> Kingsley Idehen       Weblog: http://www.openlinksw.com/blog/~kidehen
> President & CEO
> OpenLink Software     Web: http://www.openlinksw.com
>
>
>
>
>
>
>
>   


-- 


Regards,

Kingsley Idehen	      Weblog: http://www.openlinksw.com/blog/~kidehen
President & CEO 
OpenLink Software     Web: http://www.openlinksw.com

Received on Tuesday, 24 March 2009 07:54:53 UTC