JW: The first criteria at level 1 should be divided into 2 clauses.
Either provide the characters in Unicode, or provide the mapping explicitly
in a standard way. W3C character model is Unicode with certain restrictions
on use; we probably don't want those restrictions on content. For example,
there are combining characters in Unicode that can either be represented
as a single Unicode or a sequence. The W3C character model says which must
be used.
CS: Using Unicode instead of W3C doesn't have an effect on whether
text is accessible.
GV: Jason is saying it is unambiguous. It does mean your screen reader
would need to support interpreting both representations.
JW: It is possible to do an automatic translation between the representations
without any loss.
CS: That might be a particular alphabet's code page mapping, or Unicode,
or W3C character model, or... All these
work in the real world and could be handled by assistive technology.
WC: The W3C character model is Unicode.
GV: Jason, you are saying that it should say "text in the content must
be in Unicode or an explicit mapping to Unicode must be provided"
JW: If there is an encoding that is a standard that has a mapping to
Unicode, that should satisfy.
GV: Unicode is double byte and Ascii is single byte. Would it satisfy?
Can you use single-byte characters and call it Unicode?
CS: There are existing mappings.
GV: How will the screen reader know what mapping to use?
CS: The OS usually provides this information.
JW: There is an issue, a potential ambiguity. If the content has to
be provided in a particular character set , that is one thing. But if there
is mapping, it has to be provided. The content requirement would be that
the mapping be there or that the User Agent can do the mapping.
CS: Need a meta-tag for declaring what code page you use. Then the
user agent knows what to do with it.
LGR: Need the encoding explicitly identified in the content.
CS: Why is declaring the natural language a level 3? It is trivially
easy and solves this problem.
GV: There are sites that are otherwise accessible but don't have that
info, and that are archival. Also, there are Word documents that don't
necessarily have this. But what does this really help? Usually you can
tell what language by looking at the page.
JW: Language and character sets are separate issues. But I can rework
success criterion 1.
WC: The Character Model has a lot of this in there. Why don't we just
use it?
??: And the Character Model will be required of W3C standards.
CS: There may be situations where it is not appropriate.
GV: But this is a level 1, so we have to be careful we don't wipe out
all PDF, for example.
LGR: Are all code pages Unicode subsets?
GV: Microsoft is setting things up to map to Unicode. Because it is
a level 1, it worries me.
CS: Why don't we put it in as it currently stands, with an open question
to refine this.
GW: Text in the content must be Unicode or automatically mapped back
to Unicode. Then define what automatically means. Net effect is screen
reader sees Unicode. User Agent can do this, OS can do this , etc.
JW: Would a person writing an authoring tool read that as meaning they
need to provide the character map? They should.
GV: If the Reader doesn't already do the mapping.
LGR: It depends on the encoding used for a font. Some standard encodings
can be converted to Unicode automatically. For other encodings, a ToUnicode
table is required.
JW: Just be sure that the author understands he is responsible to provide
the information, whether it is the mapping or the information necessary
to generate the mapping automatically. I don't think the clause after the
"or" makes it clear that the author needs to provide the necessary information.
GV: "information is provided so that it is automatically mapped."
CS: it will vary by technology and language what that information is.
In HTML, if the page is in ASCII, nothing additional needs to be provided.
LS: another question: are we asking for unambiguous decoding or unambiguous
meaning? Does meaning belong here?
JW: That is guideline 4.
GV: That's understanding. What the person is trying to convey with
the word (as opposed to understanding what word he is using). But have
we just ruled out English, e.g., read vs read.
CS: Perhaps we need to put something in about whether the language
supports it. English is full of different words that are spelled the same
way.
GV: Add the word "standard". "Standard symbols". "Standard usage symbols".
Then we need to say what standard usage is. But then Israel will say standard
usage is not to use them.
LS: "Commonly understood"? Everyone in Israel knows what those marks
mean.
GV: "Common usage" symbols
??: "Common symbols"
GV: "Symbols such as diacritic marks that are commonly used"
CS: adding diacritic marks to distinguish read and read will reduce
readability for most people.
LS: "Common" instead of "common usage"? Screen readers are very expensive
because they do include the language analysis tools, but they are only
80% accurate and very slow and very expensive. If you want a country like
Israel to be accessible, this is a requirement.
GV: Israel could require level 2 on this item. But putting it in level
1 imposes it on every country. Let's see whether language translation work
covers this.
JW: "Which is common for the language"
??: "commonly used and necessary"
JW: "commonly used in the language of the content and necessary"
LS: "commonly understood"
??: "that are found in common usage"
??: "found in standard usage"
GV: will update the checkpoint in the document
JW: Anyone not at the meeting who has comments on 2.1 should post them
to the mailing list.
$Date: 2002/06/20 21:45:12 $ Loretta Guarino Reid