- From: Gavin Nicol <gtn@ebt.com>
- Date: Wed, 18 Dec 1996 00:04:16 -0500
- To: papresco@calum.csclub.uwaterloo.ca
- CC: w3c-sgml-wg@w3.org
>>that having all whitespace be significant still seems a reasonable
>>way to go.
>
>Can you please describe *exactly* what that means?
....
>
>At other points, there has been discussion of having a DTD-reading "filter"
>remove the whitespace. Which seems to imply that the former would be *valid*
>as long as the filter is applied before the validation takes place. In this
>case, the grove which is being validated is different from the grove that a
>DTD-less parser would use.
I repeat my viewpoint:
1) The *parser* does not use a DTD, and so creates a pGrove (to use
Elliot's term) in which *all* non-markup charaters occur (lot's
of psuedo-elements).
[pGrove -> pGrove]
2) For pure XML *validators* of the pGrove, the following:
<LIST>
<ITEM>foo</ITEM>
</LIST>
would cause an error if LIST couldn't contain #PCDATA.
[pGrove -> validator]
3) For XML *validators* of the pGrove that are built to support
legacy SGML systems, the following:
<LIST>
<ITEM>foo</ITEM>
</LIST>
would not cause an error (ie. "normal" SGML behaviour because
they would perform some transformation of the pGrove).
[pGrove -> validator -> epGrove].
I expect to see most new applications built around (1), and many
others to use (3) to obtain the semantics they desire.
A "parser" is something that tokenises the stream, and checks only
the syntactic constraints imposed by the XML grammar.
A "validator" is something that takes a pGrove, and checks that it
comforms to the constraints imposed by the grammar as defined by a
DTD.
Received on Wednesday, 18 December 1996 00:05:41 UTC