- From: Charles McCathieNevile <chaals@opera.com>
- Date: Mon, 04 Oct 2010 12:12:25 +0200
- To: w3c-wai-ig@w3.org, "Salinee Kuakiatwong" <salinee20@gmail.com>
On Mon, 04 Oct 2010 10:14:36 +0200, Salinee Kuakiatwong
<salinee20@gmail.com> wrote:
> Dear All,
>
> I'm writing a research paper to investigate the inter-reliability of
> automated evaluation tools. I used two automated web evaluation tools to
> scan the same web pages. The findings indicates there are highly
> discrepancies in the results between both tools although they're based on
> the same standard (WCAG 2.0).
>
> I'm new to the field. Any explanation for such a case?
Yes. Automated evaluation is pretty limited - each tool will use its own
set of algortihms and heuristics and therefore probably not even test
exactly the same things, let alone get the same results. You should do a
manual evaluation yourself as part of the research paper, which will give
you more insight into the particular issues that have arisen with the two
automatic evaluations.
cheers
Chaals
--
Charles McCathieNevile Opera Software, Standards Group
je parle français -- hablo español -- jeg lærer norsk
http://my.opera.com/chaals Try Opera: http://www.opera.com
Received on Monday, 4 October 2010 10:13:34 UTC