- From: <samuelm@dit.upm.es>
- Date: Wed, 13 Feb 2013 14:25:25 +0100
- To: "ERT WG" <public-wai-ert@w3.org>
Dear ERT,
I have supervised two M.Sc. theses which included a survey on
accessibility evaluation tools. For that, a set of descriptors were
defined, which were then applied to the different tools. I have quickly
compiled them and provide a summary below, in case they might be helpful
as an input for the Requirements Analaysis for Techniques for Automated
and Semi-Automated Evaluation Tools. Note this list is descriptive, not
prescriptive: it was just created as a framework to describe more easily
the different tools, but it does not imply any choice is superior above
others.
Regards,
Samuel.
Features of evaluation tools:
- Deployment:
· online service
· browser-triggered remote service (scriptlet, favelet, menu add-on)
· server-side module (i.e. web application)
· rich-client editor module (e.g. CMS, etc. maybe relying on remote
server support)
· browser plug-in
· installable desktop software
· stand-alone (no installation) desktop software
- Platform requirements: OS, environment, dependencies, etc.
- Retrieval of evaluated contents:
· capture rendered presentation directly from the browser
· access to public URI from a remote server
· access to a URI directly from the evaluator's equipment
· access to local file system: either accessing a file:/// URI, or
directly accessing the local file-system, or uploading a form-data
encoded file to a service
· direct user input.
- Analysis depth:
· single document,
· follow-links constrained to depth level
· follow-links constrained to path filter (i.e. set of directories,
subdirectories)
· follow-links constrained to domain filter (e.g. same domain, subdomains).
- Accessibility requirements tested:
· guideline families (here, usually WCAG 2.0)
· success criteria selection: one by one, by conformance level
· technique selection: automatic (depending on the content type),
partially manual.
· user-defined techniques (using formal languages, plugins, etc.)
- Reporting:
· summarized: scores -and specific metric used-, aggregated tables, radar
chart.
· detailed: table, tree-like, linear
· grouping: by criteria, level, result...
· visual annotation: on top of the original rendering of the content, on
top of the original source code
· output formats (e.g. HTML, PDF)
· EARL support, including any vocabulary extensions (e.g.
- Manual revision: manual annotation of the report, adding the results of
evaluation tests.
Apart from those features, other, more targeted tools were identified:
- Browser toolbars, characterized by their functionalities. These can
mainly be grouped in: content manipulation, content summarization, and
browser reconfiguration.
- Specific criteria: contrast analyzers, readability analyzers, formal
validators, etc.
- Emulators of specific user ability profiles ("disability simulators").
Received on Wednesday, 13 February 2013 13:30:03 UTC