link-checker commit: Remove obsolete WWW::RobotRules comments.

changeset:   285:edc09632e805
user:        ville
date:        Tue Feb 10 19:21:01 2009 +0000
files:       bin/checklink
description:
Remove obsolete WWW::RobotRules comments.


diff -r 418e0549ff90 -r edc09632e805 bin/checklink
--- a/bin/checklink	Tue Feb 10 18:50:44 2009 +0000
+++ b/bin/checklink	Tue Feb 10 19:21:01 2009 +0000
@@ -5,7 +5,7 @@
 # (c) 1999-2009 World Wide Web Consortium
 # based on Renaud Bruyeron's checklink.pl
 #
-# $Id: checklink,v 4.143 2009-02-10 18:37:27 ville Exp $
+# $Id: checklink,v 4.144 2009-02-10 19:21:01 ville Exp $
 #
 # This program is licensed under the W3C(r) Software License:
 #       http://www.w3.org/Consortium/Legal/copyright-software
@@ -55,9 +55,7 @@
   # Cheat by defining something for the constructor, and resetting it later.
   my $from_ok = $from;
   $from ||= 'www-validator@w3.org';
-  # WWW::RobotRules <= 5.78 have bugs which cause suboptimal results with
-  # User-Agent substring matching against robots.txt files; "User-Agent: *"
-  # should work ok with all though, and "User-Agent: W3C-checklink" for >= 5.77
+
   my $self;
   if (USE_ROBOT_UA) {
     $self = $class->SUPER::new($name, $from, $rules);
@@ -267,7 +265,7 @@
   $PROGRAM     = 'W3C-checklink';
   $VERSION     = '4.4';
   $REVISION    = sprintf('version %s (c) 1999-2009 W3C', $VERSION);
-  my ($cvsver) = q$Revision: 4.143 $ =~ /(\d+[\d\.]*\.\d+)/;
+  my ($cvsver) = q$Revision: 4.144 $ =~ /(\d+[\d\.]*\.\d+)/;
   $AGENT       = sprintf('%s/%s [%s] %s',
                          $PROGRAM, $VERSION, $cvsver,
                          (W3C::UserAgent::USE_ROBOT_UA

Received on Thursday, 5 August 2010 14:47:31 UTC