W3C home > Mailing lists > Public > w3c-ietf-xmldsig@w3.org > April to June 2000

Re: No Character Normalization?

From: Khaja E. Ahmed <khajaa@valicert.com>
Date: Fri, 23 Jun 2000 15:26:47 -0700
To: John Boyer <jboyer@PureEdge.com>
Cc: Kevin Regan <kevinr@valicert.com>, w3c-ietf-xmldsig@w3.org
Message-id: <3953E427.6BDB0835@valicert.com>
I did not get this informative and authoritative response.  Not sure if the
group was not copied on it or if the mail gremlin got to my inbox.  Could this
possibly be mailed out to the group.   Thanks.


John Boyer wrote:

> Hi Kevin,
> I appreciated your email, however, I did not get it (due to a meeting) until
> after it was already answered.  Full credit for the informative and
> authoritative answer belongs to John Cowan.
> Sincerely,
> ***************************************
> John Boyer,
> Software Development Manager
> PureEdge Solutions (formerly UWI.Com)
> Creating Binding E-Commerce
> v:250-479-8334, ext. 143 f:250-479-3772
> 1-888-517-2675  http://www.PureEdge.com
> ***************************************
> -----Original Message-----
> From: Kevin Regan [mailto:kevinr@valicert.com]
> Sent: Friday, June 23, 2000 2:58 PM
> To: jboyer@PureEdge.com; w3c-ietf-xmldsig@w3.org
> Subject: RE: No Character Normalization?
> John,
> Thanks for the information.
> My greatest concern is to not have to tell my customer that "No, I
> can't sign that.  How did you create that document anyway?"
> If it is the usual case that documents are created in the normalized
> form, then it does not seem like a big issue.  What would happen
> in the case of an editor or application written in Java (Unicode)?
> It seems that this is the most important case given the close
> coupling of Java and XML.
> Another concern is whether a document can become "de-normalized" during
> transmission.  My previous question was not specific enough. I understand
> that documents can be converted to other character formats. However, I'm
> wondering if a document can leave one application in a normalized form, go
> through various character encodings, and enter another application
> with the characters no longer normalized (e.g.  A Java application to Java
> application might go from Unicode, to UTF-8 for transmission, and then
> back to Unicode in the other application).
> Finally, you mention that the detection of a non-normalized document
> would aid in the discovery of forgery.  My question is: should similar
> documents with different character models be equivalent?
> What would most people expect?  I don't really understand the usage
> enough to have an opinion on this...
> --Kevin
Received on Friday, 23 June 2000 18:25:00 UTC

This archive was generated by hypermail 2.3.1 : Tuesday, 6 January 2015 21:21:33 UTC