W3C home > Mailing lists > Public > www-lib@w3.org > July to September 1998

Re: Feeding the webbot with a list of addresses to check

From: Henrik Frystyk Nielsen <frystyk@w3.org>
Date: Wed, 30 Sep 1998 18:05:29 -0400
Message-Id: <>
To: Bob Racko <bobr@dprc.net>, www-lib@w3.org
Cc: "Lola Garcia Santiago" <mdolo@platon.ugr.es>
At 21:57 9/29/98 -0400, Bob Racko wrote:
>mostly when I want to have the robot check
>multiple URI's I run it multiple times
>from a script or bat file.
>I suppose we can change the interpretation
>of the second and subsequent arguments to mean
>additional URI's. Right now they mean keywords to search for.

Or maybe a line argument where the value is the name of a file with a list
of URIs?

>I personally would prefer (yet another) command line
>option along the lines of -addURI *
>which would add the named link to the list of initial documents to load.
>(The same as finding an <A href="*" > in a document except this link
>would name an alternate root or starting-point).
>This begs the issue of return-status. What does it mean to
>try to fetch one or more documents and they fail to fetch
>(or fail to parse, or fail to... ) ?

Using a "logic engine" may not be a bad idea - then you can say things
like: "get this document only of this and this document fails". Is there a
logic parser out there (with and, or, negate for a start) and would this be
a cool thing to do?


Henrik Frystyk Nielsen,
World Wide Web Consortium
Received on Wednesday, 30 September 1998 18:05:32 UTC

This archive was generated by hypermail 2.3.1 : Tuesday, 6 January 2015 21:33:48 UTC