W3C home > Mailing lists > Public > public-i18n-geo@w3.org > June 2003

RE: backburner question: Checking HTTP Headers (was: Sites to see headers)

From: Martin Duerst <duerst@w3.org>
Date: Tue, 17 Jun 2003 12:37:11 -0400
Message-Id: <4.2.0.58.J.20030617114239.02c9d620@localhost>
To: <ishida@w3.org>, <public-i18n-geo@w3.org>
Cc: "'Olivier Thereaux'" <ot@w3.org>

At 15:35 03/06/17 +0100, Richard Ishida wrote:
>Martin,
>
>Thanks for proposing this.
>
>I have just edited your file as follows:
>-       standardised the structure
>-       added a 'content created' date at the bottom (as per Tex's
>suggestion)
>-       added a charset declaration
>-       removed trailing dots from headers except last header
>(consistent with current approach), and changed 'Notes' header to 'By
>the way...' (hope that's ok)

Thanks a lot, also for your other comments.


>I have the following suggestions:
>
>-       the background and answer suggest to me that the question should
>rather be "How can I check the character encoding information sent in
>the HTTP header of a web document?"

Done.


>-       I'd suggest using h3 elements rather than bulleted list (for
>readability, but also consistency - see eg
>http://www.w3.org/International/questions/qa-date-format.html)

Done.


>-       the ((X)HTML) source =>  the (X)HTML source  OR   (and the
>(X)HTML source)

I took away the outer parentheses.


>-       Content-Type: text/html; charset=utf-8  - I'm not sure the utf-8
>needs to be italicised

I put it in italics to show that that's the place where people
have to look to. Any alternatives welcome.


>-       the actual markup =>  the markup

done.


>-       e.g. a document claiming to be iso-8859-1 may not actually be
>iso-8859-2. => for example, a document claiming to be iso-8859-1 is
>actually encoded using iso-8859-2 or some other encoding.

done.

>Hope that helps.

Thanks!


>See also comments below.

I read them. I'll try to follow. But I'd prefer markup
conventions that are easy to deal in a visual editor such
as Amaya.


Regards,    Martin.


>RI
>
>============
>Richard Ishida
>W3C
>
>tel: +44 1753 480 292
>http://www.w3.org/International/
>http://www.w3.org/People/Ishida/
>
>
>
> > -----Original Message-----
> > From: public-i18n-geo-request@w3.org
> > [mailto:public-i18n-geo-request@w3.org] On Behalf Of Martin Duerst
> > Sent: 16 June 2003 22:47
> > To: public-i18n-geo@w3.org
> > Cc: Olivier Thereaux
> > Subject: backburner question: Checking HTTP Headers (was:
> > Sites to see headers)
> >
> >
> >
> > I just made a question out of the list of sites to see
> > HTTP headers. Thanks to Andrew and Tex for their help.
> >
> > This is for my next round, so we don't have to
> > discuss this on this Wednesday. Please see
> > http://www.w3.org/International/questions/qa-headers-charset.h
>tml
>
>
> >Olivier, I have copied you because this mentions the validator.
>
> >Richard, I'm not sure I got the
> ><div class="content"> markup right. It's not clear
> >what it is for,
>
>I should have used a class="section" div that surrounded both h2 and
>section content.  At the time I was not sure whether the negative
>indents that that would require for the headings would work properly
>everywhere, and didn't have time to explore it.
>
>It makes it much easier to process the page with XSLT if additional
>structure such as sections are added, but it also allows me to add a
>consistent css margin to the section on one line rather than having to
>add margins to any element that might appear in the section.
>
>
> >but it seems to affect styling in
> >somewhat strange ways.
>
>Amaya certainly seems to turn this into a dog's breakfast.  It looks
>fine in NN7, NN4, IE and Opera.
>
>
> >In general, to reduce overhead,
> >it is easiest to mark up all the other parts and leave
> >the unmarked parts as simple content, but maybe I got
> >something wrong.
>
>
>Note that there is one <div class="content"> for each section (I should
>have called it 'section-content' - or implemented properly as
>'section').  That was the only problem - I fixed it for you.
>
>
> >regards,     Martin.
>
>
>
>At 09:52 03/06/12 +1000, Andrew Cunningham wrote:
>
> >Martin Duerst wrote:
> >>There are the sites I know to check
> >>http://webtools.mozilla.org/web-sniffer/
> >>http://www.delorie.com/web/headers.html
> >
> >http://www.rexswain.com/httpview.html
> >
> >
> >likewise if your interested in the http request
> >
> >http://www.delorie.com:81/some/url.html
> >http://www.i18ngurus.com/cgi-bin/TestLang.pl
> >
> >
> >
> >--
> >Andrew Cunningham
> >Multilingual Technical Officer
> >Online Projects Team, Vicnet
> >State Library of Victoria
> >328 Swanston Street
> >Melbourne  VIC  3000
> >Australia
> >
> >andrewc@vicnet.net.au
> >
> >Ph. +61-3-8664-7430
> >Fax: +61-3-9639-2175
> >
> >http://www.openroad.net.au/
> >http://www.libraries.vic.gov.au/
> >http://www.vicnet.net.au/
Received on Tuesday, 17 June 2003 12:37:19 GMT

This archive was generated by hypermail 2.2.0+W3C-0.50 : Tuesday, 8 January 2008 14:12:37 GMT