- From: Danny Ayers <danny@panlanka.net>
- Date: Thu, 29 Mar 2001 12:09:15 +0600
- To: "RDF-Interest" <www-rdf-interest@w3.org>
I'm thinking about a little system, with at least the following resources defined in an RDF doc : data inference engine algorithm processor formatter for result label for resulting resource Use case : --- I have a set of CSV data files, the results of an experiment collected by a legacy tool I want to get a nice table on screen of some statistical analysis of this data. First of all I want to build the RDF containing the instructions - a tool would be nice, but for now a text editor. The resources involved would be : location of the raw data description of the required conversion -> source code (algorithm) for data conversion CSV -> XDF -> compiler for source code -> instructions for compilation and running (rdf + data for intermediate results) description of the required conversion -> source code (algorithm) for statistical analysis XDF -> XDF -> compiler for source code -> instructions for compilation and running (rdf + data for intermediate results) description of the required view -> XSL for view -> instructions for launching viewer the inference engine in the list above is the meta-engine for running through the procedure --- I imagine that by using a named algorithm, the (specified) inference engine could first check to see if a pre-built processor was available for $PLATFORM, otherwise pick up the algorithm as source code suitable for $PLATFORM and build the processor - the resulting processor could be cached and a reference marked in the system. The processor would necessarily have to be on the local systems. The compiler would be specified in the same way, though obviously having a pre-built compiler for at least one language would be a good start - if necessary the system could bootstrap from there. Note that the whole procedure for CSV -> analysis view could be retained as yet another algorithm, specified as the same RDF but without the binding to the source data. There's been some discussion recently about how to deal with the results of processing, I'm not sure if there was any resolution. What I'm suggesting with the above is the (rdf + data for intermediate results) will contain the result of running the algorithm, and a reference to the tools used - this will be the effectively the chain of trust. What I'd like to ask the illustrious correspondents of this list, is what schema are already in existence that could cover some or even all of the above, and also for comments as to the viability and likely problems of such a system. --- Danny Ayers http://www.isacat.net
Received on Thursday, 29 March 2001 01:12:06 UTC