- From: Mercurial notifier <nobody@w3.org>
- Date: Thu, 17 Mar 2011 21:10:48 +0000
- To: link-checker updates <www-validator-cvs@w3.org>
changeset: 385:74a3e03b3520 user: Ville Skyttä <ville.skytta@iki.fi> date: Thu Mar 17 19:29:34 2011 +0200 files: docs/checklink.html description: Drop unintentional tabs introduced in r377. diff -r 2ef5d1cb8369 -r 74a3e03b3520 docs/checklink.html --- a/docs/checklink.html Mon Mar 14 18:06:13 2011 +0200 +++ b/docs/checklink.html Thu Mar 17 19:29:34 2011 +0200 @@ -196,9 +196,9 @@ <p> In online mode, link checker's output should not be buffered to avoid - browser timeouts. The link checker itself does not buffer its output, + browser timeouts. The link checker itself does not buffer its output, but in some cases output buffering needs to be explicitly disabled for - it in the web server running it. One such case is Apache's mod_deflate + it in the web server running it. One such case is Apache's mod_deflate compression module which as a side effect results in output buffering; one way to disable it for the link checker (while leaving it enabled for other resources if configured so elsewhere) is to add the following @@ -220,7 +220,7 @@ <p> The link checker honors proxy settings from the - <code><em>scheme</em>_proxy</code> environment variables. See + <code><em>scheme</em>_proxy</code> environment variables. See <a href="http://search.cpan.org/dist/libwww-perl/lib/LWP.pm#ENVIRONMENT">LWP(3)</a> and <a href="http://search.cpan.org/dist/libwww-perl/lib/LWP/UserAgent.pm#%24ua-%3Eenv_proxy">LWP::UserAgent(3)'s <code>env_proxy</code></a> method for more information. @@ -267,7 +267,7 @@ <a href="http://search.cpan.org/dist/libwww-perl/lib/LWP/RobotUA.pm">LWP::RobotUA</a> Perl module. It currently supports the "<a href="http://www.robotstxt.org/wc/norobots.html">original 1994 version</a>" - of the standard. The robots META tag, ie. + of the standard. The robots META tag, ie. <code><meta name="robots" content="..."></code>, is not supported. Other than that, the link checker's implementation goes all the way in trying to honor robots exclusion rules; if a @@ -309,7 +309,7 @@ If a link checker run in "summary only" mode takes a long time, some user agents may stop loading the results page due to a timeout. We have placed workarounds hoping to avoid this in the code, but have not - yet found one that would work reliably for all browsers. If you + yet found one that would work reliably for all browsers. If you experience these timeouts, try avoiding "summary only" mode, or try using the link checker with another browser. </p>
Received on Thursday, 17 March 2011 21:10:50 UTC