W3C home > Mailing lists > Public > semantic-web@w3.org > September 2005

Fw: RDF tools as workhorse

From: Richard H. McCullough <rhm@volcano.net>
Date: Thu, 15 Sep 2005 04:52:24 -0700
Message-ID: <00a901c5b9ec$076fe040$0100000a@rhm8200>
To: <semantic-web@w3c.org>

I meant to do a reply-all with this email.

P.S. When I looked again at the ISO KB,
there were about 3000 lines of other stuff.
The RDF/OWL part is about 5000 lines.

Dick McCullough
knowledge := man do identify od existent done;
knowledge haspart proposition list;
http://rhm.cdepot.net/
----- Original Message ----- 
From: "Richard H. McCullough" <rhm@volcano.net>
To: "Mailing Lists" <list@thirdstation.com>
Sent: Wednesday, September 14, 2005 6:38 AM
Subject: Re: RDF tools as workhorse


>I use the MKE/MKR system (click on link below my name),
> which is currently integrated with standard GDBM databases.
> My emphasis has been on simplicity and flexibility in the user
> interface.   The interface language, MKR, includes queries,
> n-ary relations, and methods.
> 
> I have not done much work on performance issues, but
> I am confident that they can be solved as well in an 
> MKR-based system as in any other system.
> 
> My biggest projects to date have been a Genealogy KB 
> of about 1000 persons, and an ISO standards KB 
> of about 8000 lines of RDF.
> 
> I also have an MKR interface to the Stanford TAP KB
> and the OpenCyc KB.
> 
> Dick McCullough
> knowledge := man do identify od existent done;
> knowledge haspart proposition list;
> http://rhm.cdepot.net/
> 
> ----- Original Message ----- 
> From: "Mailing Lists" <list@thirdstation.com>
> To: <semantic-web@w3.org>
> Sent: Tuesday, September 13, 2005 1:46 PM
> Subject: RDF tools as workhorse
> 
> 
>> 
>> Hi all,
>> 
>> Does anyone on the list have some real-world stories to share about 
>> using RDF and its tools as a backend technology?  The company I work 
>> for maintains a database of metadata.  I'd like to explore using RDF 
>> instead of our current schemas.
>> 
>> For example:   I have a lot of data about books.  I'd like to translate 
>> the data into RDF/XML and dump it into an RDF database.  Then, taking a 
>> particular book, I'd like to query the database to extract related 
>> information like: other books by the same author, other books with the 
>> same subject code, etc.
>> 
>> My concerns relate to:
>> 1) Performance -- Right now we query the database using SQL.  Sometimes 
>> it is _very_ slow.  That's mainly because the data is distributed 
>> across tables and there are a lot of joins going on.  It seems like 
>> using RDF would allow us to use simple queries.
>> 
>> 2) Scalability -- Our triplestore would be HUGE.  I'd estimate 10-20 
>> Million triples.  Is that small or large in RDF circles?
>> 
>> 3) Productivity -- It's usually easier for me to envision creating RDF 
>> from our source data than massaging the data to fit into our database 
>> schema.  The same goes for when I'm extracting data - it seems like it 
>> would be much easier to express my query as a triple using wildcards 
>> for the data I want.
>> 
>> Any information will be helpful.  I'm interested in learning from other 
>> peoples' experiences.
>> 
>> Thanks,
>> Mark
>> 
>> ..oO  Mark Donoghue
>> ..oO  e: mark@ThirdStation.com
>> ..oO  doi: http://dx.doi.org/10.1570/m.donoghue
>
Received on Thursday, 15 September 2005 11:56:18 UTC

This archive was generated by hypermail 2.3.1 : Tuesday, 1 March 2016 07:41:47 UTC