Summary Report-Comments on WCAG Documentation- 10 Sep 03

Comments on Draft WCAG 2.0 Implementation Testing Framework:

Congratulations - This is a very good initiative towards categorizing some important ideas. However, as it currently stands the document seems to be a loose collection of ideas, without much coherent structure. A possible organization might be: (1) WCAG testing process information, (2) WCAG technical testing information, and (3) additional WCAG testing details. Comments in support of (1) and (2) are covered under "General Comments" below, and comments in support of (3) are covered under "more specific comments" below. Perhaps all that should be in this document ise WCAG-specific information, with references to other resources if they can be used (just a thought)?

General Comments (accessibility or W3C specific):

  1. What is the purpose and intent of this document? Is it designed to promote implementation testing, or to provide additional WCAG-specific conformance test information not mentioned elsewhere? Or both? This isn't made clear at the beginning of the document. Without a clear purpose and intent statement, as it is currently written, the draft document seems to have a mixed identity - a hodgepodge of test- related documents.
  2. The content of this document should follow existing QA documentation as much as possible. The two QA documents that may be applicable are: (1) stable parts of the Operational Guidelines (in particular, the QA Process Document), and (2) the Test Guidelines (under development).
  3. What happened to the WCAG test suite process document being developed by Phill Jenkins and Tim Boland at the March 2003 WCAG meeting? This document, developed from a QA template, may answer a number of questions raised. Why not use this W3C test suite process document, and if needed, the WCAG WG can have a "delta" on top of this document to address WCAG-specific issues, since the nature of WCAG testing may be different than previous testing efforts (e.g., testing guidelines rather than markup validity per se)?
  4. There is a generic accessibility testing technical document sent to the WCAG WG in March as well as to the AUWG. The AUWG WG felt it was a good starting but needed tailoring to the specific requirements of authoring tools testing. Perhaps the same is true for WCAG testing? In any case, much of the content of that document seems to overlap with content mentioned in this document.
More specific comments (things possibly WCAG-specific)
  1. Are the WCAG techniques documents mentioned normative in a testing sense (or informative)? This doesn't seem to be clearly addressed.
  2. Is it worthwhile dealing with different categories of web content and having different tests apply to each? In the latest ATAG working draft different categories of authoring tools are mentioned, with possibly different tests for each. Perhaps a resolution of this issue could be stated in this document if appropriate.
  3. It seems to me that WCAG implementation testing needs to be clearly defined in this document, in concert with any QA definitions, as opposed to testing conformance to the WCAG draft itself. Also the classification and organization of WCAG tests needs to be clearly explained, as well as how an offering would state its features for testing (by filling out a form, for example). WCAG testing may have particular requirements in these areas.
  4. It seems to be difficult for me to understand the purpose of the two tiers of implementation testers mentioned. What is the difference? Why not just one kind of implementation tester?
  5. I think that the importance of objective and testable conformance requirements (to the maximum extent possible) needs to be stated. It is recognized that certain testing activities of the W3C (such as voice browser, authoring tools, and possibly WCAG testing) may be somewhat subjective in nature, and thus it may be unavoidable to include some subjective tests, because of the nature of the WCAG draft, but such subjectivity should be kept to a minimum if possible.
  6. Usability testing principles (relating to accessibility) may play a role in certain aspects of WCAG testing (just like with authoring tools testing). It may be appropriate to draw upon usability testing resources in that event and mention in this document.
  7. I think that it is important to define test purpose, test environment, test procedure, and expected test results in a systematic fashion for each WCAG test. I think that it is also important to define which tests can be automated and which need to be "manually evaluated" NOTE: Much of this is covered in the above-mentioned testing technical document, but any WCAG-specific issues should be mentioned.
  8. Should tests be aimed at designers (proper design practices being followed) or at users (leading the user to do the "right" thing?) Perhaps there should be testing twice, at beginning and at end of "process" (that is, test both the user and the designer). It may be appropriate to mention a resolution/clarification of this issue in this document.
  9. For the CSS WG, for example, a "requirement" for exiting CR stage is that at least two implementations are able to provide support for mentioned functionality. Would a WCAG analogy be that at least two evalutation and repair tools (e.g., Bobby, A-Prompt) need to generate the same results (e.g., finding same problems with inaccessible code) for some particular content? If so, it may be appropriate to mention.
  10. An discussion item occurred previously in the AUWG as to whether it was appropriate to deliberately generate broken HTML, invalid CSS, non-well-formed XML, etc., as a part of accessibility testing, in content to see whether all errors could be caught by an authoring tool , or to keep the principle of requiring validated code as a prerequisite to any accessibility testing. Perhaps a resolution of this item could be mentioned in this document if appropriate.
  11. What is the relation of web content to a web resource, in the W3C sense (for example, to IRIs)? This is not mentioned in this document.
  12. The audience for the tests (users of the tests) need to be specified, as well as possible uses of the test results (legal, administrative (policy), etc.).
  13. The importance of "testing the tester" is not mentioned in this document. Perhaps it should be?
  14. Is there any implied extensibility of WCAG testing to content other than "web content" (assuming an agreed-upon definition exists). If so, it may be appropriate to mention.

Comments on the WCAG Working Draft of 24 June 2003:

Again, congratulations - This is an excellent step forward in a WCAG working draft. Comments here are divided into general comments on the entire document, and specific comments on each checkpoint.

General comments on entire draft (NOTE: The term "success criteria" is used throughout to refer to one criterion or multiple criteria.)

  1. Does the repeated use of the adjective "required" before each success criteria (which I like) imply that there will be "optional" success criteria?
  2. The kind of accessibility desired is not stated explicitly. Example: Success criteria for Checkpoint 4.1 "accessibility feature". What kind of accessibility? Visual? Auditory? Tactile?
  3. What is the difference between use of "non-normative" (#4 top layer of "how to read document", as well as Appendices A through C) and use of "informative" (as used in the checkpoint sections - for example, definitions for Checkpoint 2.1)? Perhaps one or the other term could be used, but not both, or at least indicate that the terms are equivalent, if they are meant to be? This is related to item below, but maybe on a higher level, relating to how the entire draft is viewed.
  4. Terminology within the draft seems not to be entirely consistent. EXAMPLE: "site", "resource", "application" are all used in section 3, and it is unclear whether the context and semantics of each use are entirely different. I think that it is important to come up with common definitions and use common terms that mean the same thing for the same concept.
  5. For the definitions under each checkpoint, are they consistent with terminology in the WAI glossary? W3C glossary? Other glossaries? I know there is W3C glossary work currently. The checkpoints are just goals, right?
  6. Is there a logical ordering to the checkpoints as presented? In the ATAG WD the checkpoints are presented in the natural order in which one would step through them - is that true here? How should one "process" the checkpoints (perhaps this could go into the "how to read this document" section)
  7. It might be good to insert quantifiers like "all" and use "MUST" where appropriate in every success criteria for testability (EXAMPLE: most success criteria language in the draft does not have such quantifiers). Thus, Words like "must do" and "required" should be used if appropriate to specify minimum requirements. NOTE: It is possible to test "should" statements but they should be considered as options.
  8. The relation between checkpoint definitions (goals?), success criteria, and text in boxes for each checkpoint seems not to be adequately explained; this may hinder usability of draft. Also, the format of the checkpoints seems not to be entirely consistent (EXAMPLE: Checkpoint 1.2 has a "best practices" section, whereas Checkpoint 1.3 does not) In general, for the checkpoint formats, it seems like implementation techniques, informational statements, and conformance requirements are mixed in together, which makes the objectivity of the entire success criteria statement difficult to determine. It may be good to only use conformance requirement language (or specify options where applicable) in success criteria language (EXAMPLE: For Success Criteria for Checkpoint 4.2, second sentence seems informational).
  9. I think a goal is that the number of success criteria per checkpoint should be minimized (EXAMPLE: Success criteria for checkpoint 1.2 has six items)
  10. Some of the success criteria seem to be duplicates of one another, and some statements in the success criteria appear contradictory or don't appear to belong in the same success criteria. EXAMPLES: Success criteria for 4.1 and 4.3 seem to overlap, and six items in success criteria for Checkpoint 1.2 seem to have little relationship with each other or are partial duplicates (#5 points to Checkpoint 1.1 for conformance). It may be desirable to ensure that semantics of each success criteria be unique as much as possible, and that each sentence in a success criteria statement contributes to the meaning of a success criteria statement in a consistent fashion, as well as being necessary for that meaning. Also, I think that hhe success criteria formats for each Checkpoint should be similar to the maximum extent possible (EXAMPLE: In almost any two success criteria from different checkpoints, the sentence structure is different).
  11. It seemed to be difficult for me to determine what is required (minimum requirements) by reading each success criteria (see item above). Thus there is a general concern with objectivity and testabiity of success criteria as stated. EXAMPLE: "easy to find" in success criteria for Checkpoint 3,4, and "significant" in success criteria for Checkpoint 1.2). It may be good to use words like "must do" and "required" to specify minimum requirements.

Specific Comments On Particular Parts of Document (NOTE: These are very detailed comments from my perspective, and mainly apply to evaluating objectivity and testability of Section 3 and specific success criteria language.)

  1. Section 3 - In the "conformance" subsection of 3, it may need to be defined what it means to "meet" or "satisfy" a checkpoint. For the "Core" group under "Conformance", "all" may need to be included; should "all" be included for the "Extended" subheading as well? What if not all Core checkpoints are passed but all Extended Checkpoints are passed (Case 5 under "Conformance Claims")? Is that possible? Is "Core" a strict subset of "Extended"? Are there dependencies between the checkpoints? Do the checkpoints overlap in any way? These questions may need to be explicitly answered in the draft.
  2. Section 3 - For the "Editorial Note" questions/issues, there may be some helpful AUWG information to address these (for example, in the first two bullets addressing "conformance claims"). QA documentation may help here as well?
  3. Section 3 - There seems to be confusion in the terminology of the four guidelines under "Overview of Design Principles" (example: "perceivable" vs. "understandable"); also, for Guidelines 1,2,3, there are circular definitions (example: "operable" is defined in terms of itself). Any overlap or ambiguity here?
  4. Success Criteria for Checkpoint 1.1 - Please insert "all" before 1 and 2, and replace "has" by "must have" if appropriate. Is there a third success criteria (empty circle after the first two items), or is the third item informational?
  5. Success Criteria for Checkpoint 1.2 - There may be a question as to objective testability of "perceived" and "to the extent possible" - #1, as well as "significant" (#2), and "is primarily" - #4. What is relationship/ordering among the six points of 1.2 success criteria?
  6. Success Criteria for Checkpoint 1.3 - It may be difficult to objectively test a "without" or "not" (testing a negation). The meaning of "emphasis" may be unclear. Is it being said here that semantics does not depend upon presentation? This criteria may need more definitions of terms. There may be some implementation techniques mixed in here.
  7. Success Criteria for Checkpoint 1.4 - what version of Unicode is being referenced? There is no reference to any Unicode conformance requirements. "All" should be placed in front in this criteria language if appropriate.
  8. Success Criteria for Checkpoint 1.5 - The word "all" may be needed in this success criteria, as well as more definitions of terms. There may be a question as to objective testability of "from each other"; does this phrase refer to position in document or kind of element?
  9. Success Criteria for Checkpoint 1.6 - what happens if background image is important to meaning of text?
  10. Success Criteria for Checkpoint 2.1 - It may be difficult to objectively test the word "concisely".
  11. Success Criteria for Checkpoint 2.2 - Please replace "is" by "must be" if appropriate. Also the last two bullets read differently (have a different format) from the first three bullets.
  12. Success Criteria for Checkpoint 2.3 - It may be difficult to objectively test "not designed to", because the intent of the designer is being queried. It also may be difficult to objectively test a negative, as well as to objectively test "as close as". The relation between a and b seems to be unclear. For b, in spite of the warning, does the user go to the page anyway with content that does not flicker substituted for the flickering content? Or is the user warned and thus does not go to the desired page, but is directed to a different (alternate) page with the same content without flicker? There seems to be an inconsistency between the first part of b, which says flicker is unavoidable, and the second part of b, which says that content without flicker is provided. The definition of "page" may need to be clarified and may be subjective; is "site" or "resource meant in this success criteria?
  13. Success Criteria for Checkpoint 2.4 - The term "words" may be English-centric, and testing "perceived" pages may be subjective. There may need to be a definition of "display orders". Are all of a,b,c required to be provided simultaneously, or just some of the items at different times?
  14. Success Criteria for Checkpoint 2.5 - What kind of error is meant in this success criteria? What constitutes an error? There are a lot of different kinds of errors and what is an error to one might not be (or be a warning) to another. Is every occurrence of every type of "error" meant? Please substitute "must be" for "is" if appropriate. What kind of feedback (notification or system shut down or aborted operation or warning) must be provided? Must the feedback be provided in every instance?
  15. Success Criteria for Checkpoint 3.1 - Seemingly inconsistent terms (passages, content, fragment) with the appearance of some overlap in meaning are used; also "document" (yet another different term) is used in "best practices" sentence. I think that there needs to be a consistent way of stating terms that mean the same thing in a success criteria. Please insert "all" in front of "passages" if appropriate. Please provide a definition of a "passage" if appropriate. Please replace "are" with "must be" if appropriate. It seems to be unclear as to whom the identification is directed (to the user, perhaps?).
  16. Success Criteria for Checkpoint 3.2 - It may be difficult to test "do not appear unambiguously" objectively. Please use "all" and "must be" if appropriate. Is there one unabridged dictionary for a language? It seems there may be several dictionaries for a language with "full (definition?)" content. This may have been a subject discussion at the W3C glossary meeting in March 2003. Also, "site" is used here, but "page" is used elsewhere; are the two terms being used interchangeably to mean the same thing? Would "resource" be a better term? How does one objectively measure the availability of a glossary on a site (depends upon definition of "site" and "availability")?
  17. Success Criteria for Checkpoint 3.3 - Is all content on a site or resource meant by "the content"? Which content? There may be an issue with the objective testability of "has been reviewed". Reviewed for what? There may be an issue with objective testability of "taking into account" and "apply as appropriate". Must all of a through g be applied simultaneously? If not, which ones apply to which content in an objective sense? There may be needed in this success criteria a definition of terms in a through g. Who is "they" in g?
  18. Success Criteria for Checkpoint 3.4 - A definition of "key" may be needed. Objective testability of "are generally found" may be an issue. Please use "all" and "must" where appropriate. What is the context of the locations? The relationship between 1, 2, and 3 seems unclear; either 1 or 2 or 3 is true at a particular time, but not all three simultaneously? Do 1,2,3 represent a partitioning of predictability? There may be questions as to the objective testability of "extreme" in 3, and objective testability of "easy to find" in 3a.
  19. Success Criteria for Checkpoint 4.1 - #1- The objective testability of "are avoided" may be an issue (avoided in what sense?). There may need to be unambiguous definitions of "feature" and "accessibility feature". The objective testability of "are used" (in several places) may be of concern. Please use "All markup" and "must" as appropriate. What is the relationship between "markup", "element", "content", and "document" (all terms used in this draft - do they mean the same thing)? What is the definition of "markup" vs. "element"? There seems to be an overlap with success criteria of checkpoint 4.3 of #2 and #3 of this checkpoint's success criteria.
  20. Success Criteria for Checkpoint 4.2 - What is the definition of "web resource" vs definitions of "site" or "page"? Please use "must include" if appropriate. What does "its" refer to? The objective testability of "as intended" may be an issue. There seems to be an inconsistency between sentences #1 and #2; also sentence #2 seems informational, not normative.
  21. Success Criteria for Checkpoint 4.3 - Please replace "the.." with all technologies or combinations of technologies" if appropriate. In what context are the technologies chosen (by site, resource, application, page)? For a, there may need to be a definition of W3C device independence (DI) support re: W3C DI activity? For b, there may need to be a definition of "accessibility feature". Is it required to do all of a thru e simultaneously? For c, a definition may be needed of "publicly", "interfaces", "interoperability"? A distinction concerning interoperability testing vs. conformance testing may need to be made. Is an interface a technique in this success criteria? For d, is this a requirement for one or all operating system accessibility features (definition?)? what if there are none? There may be an issue with testability of "suported by assistive technologies" (definition?); what if there is no support? The relation between 1 and 2 of this success criteria seems unclear. Why is 2 associated with this success criteria? There may need to be a definition of "custom user interface". Does a custom user interface represent a technique in this success criteria? The items under 1 seem ambiguous and not necessarily accessibility-related.