- From: Poul-Henning Kamp <phk@phk.freebsd.dk>
- Date: Sat, 09 Feb 2013 14:04:30 +0000
- To: Willy Tarreau <w@1wt.eu>
- cc: Martin J. Dürst <duerst@it.aoyama.ac.jp>, Mark Nottingham <mnot@mnot.net>, James M Snell <jasnell@gmail.com>, "ietf-http-wg@w3.org" <ietf-http-wg@w3.org>
Content-Type: text/plain; charset=ISO-8859-1 -------- In message <20130209133341.GA8712@1wt.eu>, Willy Tarreau writes: >I'm not saying I'm totally against UTF-8 in HTTP/2 [...] What and where do you mean when you say "UTF-8" In HTTP/2 ? I think we need to be more precise, to avoid misunderstandings. In HTTP/1, there is a peculiar mix between protocol-mechanics, and metadata: If I add a custom bit of metadata, it must follow certain rules, since otherwise it will break the protocol mechanics. For instance, I cannot define a custom header called: "FOO" CRNL CRNL ": " [8 zero bytes] If we define HTTP/2 as a "binary" protocol in some sensible way, this restriction could go away, and we'd just move something like: <HDR nlen=7,blen=8> "FOO" CRNL CRNL \0\0\0\0\0\0\0\0 down the wire, and not care about what it is, what it means or what character set, if any, it is encoded in. It is only the metadata that needs inspection along the way where we need to decide about UTF-8, and it really isn't that much. Host: Why would we care about the character set ? We're just going to pass it to DNS anyway. URI: At least the query strings, possibly all of it ? But do we really care ? Provided we take the Host part out, as proposed, we treat this as a unit. Cache-Control: And what good would UTF-8 do here in the first place ? So where is it you want UTF-8, and what difference will it make ? -- Poul-Henning Kamp | UNIX since Zilog Zeus 3.20 phk@FreeBSD.ORG | TCP/IP since RFC 956 FreeBSD committer | BSD since 4.3-tahoe Never attribute to malice what can adequately be explained by incompetence.
Received on Saturday, 9 February 2013 14:04:54 UTC