- From: Jerry Peek <jpeek@jpeek.com>
- Date: Sun, 30 Jan 2022 11:48:26 -0700
- To: site-comments@w3.org
Received on Sunday, 30 January 2022 20:48:09 UTC
Thanks for the broken link finder https://validator.w3.org/checklink. I have a suggestion. Each of the files on the site I'm checking has a link to a particular file. That file is blocked in robots.txt. For every file the link checker reads, there's an error saying: Status: (N/A) Forbidden by robots.txt The link was not checked due to robots exclusion rules. Check the link manually. Checking manually has the same effect as if the link checker checks it: in both cases, the file is accessed. So I suggest changing the behavior to access the file anyway. (You might add an option that tells the link checker to behave as it does not: not check links blocked by robots.txt.) Jerry Peek
Received on Sunday, 30 January 2022 20:48:09 UTC