W3C home > Mailing lists > Public > public-xml-processing-model-wg@w3.org > December 2005

Use Cases

From: Alex Milowski <alex@milowski.org>
Date: Thu, 22 Dec 2005 07:30:13 -0800
Message-ID: <43AAC685.8070200@milowski.org>
To: public-xml-processing-model-wg@w3.org

1. A Web Service for US Tide Information


   Here I use an XML pipeline to implement a REST-style web service
   where I pull tide information from the NOAA, manipulate the web page
   that contains the data, and then return an XML representation.

   This is a combination of:

       URL interaction to retrieve and parse a web page.
       Use of the Tagsoup parser to deal with not-so-well-formed XHTML
       Translation of this into a particular XML representation with XSLT

   There is a front-end application that is also a pipeline that interfaces
   with this web service and provides a web front-end for a mobile phone

   Here's the direct links:


2. Parsing/Unparsing embedded XHTML.

   A both Jeteye [1] and as samples for  the small project, I need to be
   able to parse and escape subtrees of a document to interact RSS feeds.
   This allows a processor to deal with elements like the description 
   in RSS.

3.   A replacement for Cocoon

   This was part of the Sylvia project at the CDE 

   The basic use cases are:

    1. Chaining a sequence of XSLT transforms.
    2. Dynamic application of XSLT.  Here the important bit was that there
        is a pipeline style in smallx that lets you compose a vocabulary 
        that specifies the XSLT to run on a subtree:

        <apply-xslt src='myxslt.xsl'>

        The result of the transform replaces the  'apply-xslt' element.

    This is an important use case for web-application use of pipelines 
as they often
    need to be dynamic in their choice of XSLT/etc.

4. A "fat pipe" implementation for mathematical computation

   As part of Monos/Xeerkat/Mathgrid.org, I use XML pipelines to implement
   mathematical computations.  Custom pipelne steps that implement certain
   operations are provided by the Monos software.  These might be simple
   operations like 'transpose a matrix' to more complex algorithms like 
   a groebner basis.  The point here is that I can mix these steps with 
   pipeline steps like XSLT or XQuery.

   The point being is that many computation operations are data manipulation
   tasks.  Many tools have insufficient ability to do this operations.  
In this way,
   XML pipelines excel at letting users mix data manipulation tasks with 

   I have certain steps that are simple element filters.  For example, 
the groebner
   basis pipeline step takes in an XML 'matrix' element and outputs an XML
   'matrix' element.

   See this document for more information:


--Alex Milowski
Received on Thursday, 22 December 2005 15:30:38 UTC

This archive was generated by hypermail 2.4.0 : Friday, 17 January 2020 19:41:29 UTC