W3C home > Mailing lists > Public > www-html@w3.org > November 2006

Re: [OT] Suggestion: http request bundle

From: David Woolley <david@djwhome.demon.co.uk>
Date: Sun, 5 Nov 2006 14:02:29 +0000 (GMT)
Message-Id: <200611051402.kA5E2Tm09540@djwhome.demon.co.uk>
To: www-html@w3.org

> Since browsers are currently configured to load HTML pages in a rather
> slow and piecemeal manner, one can obtain performance improvements by
> reducing the number of JS and CSS files required on initial load.
> So my comment was simply that Doug's suggestion about reducing the
> number of individual files an HTML file is dependent, actually has
> performance benefits.

His proposal for doing so was to do so at the HTTP level, in a way
which would require every browser, proxy and server to change.  I was
pointing out that if the developers thought it important enough, they
could achieve this with existing servers and clients, although HTTP 1.1
proxies are relatively thin on the ground.

> These benefits to HTML page loading seem both on-topic, and don't
> require reconfiguring every server, proxy and browser...which you seem
> to be suggesting is "all you need to do"! :)

As proposed, it required not only configuring, but modifying the code.

One could conceive of a <link rel="resourcelibrary" element that provided
an archive of the static resources for the page, and meta data to allow
the browser to determine when they became out of date.  If you used .zip
type formats, where the metadata is uncompressed, that wouldn't offer much
performance advantage over proper implementation of HTTP/1.1, and it would
significantly increase the the overhead on proxies, as they would have
to transfer the complete library when any component changed, and, unless
they were modified to understand the format (which would be a layering
violation of HTTP) would also have to store copies of the individual
resources, many of which would be images, and therefore already

The ZIP format, itself, isn't particularly satisfactory, as the main
location for the metadata is at the very end of the file, so one would
have to rely on the recovery information if one wanted to display 
images as soon as possible (which, given the general failure of authors
to use alt attributes properly, tends to be quite important).

Also, my impression is that images and out of line scripts do often
considerably exceed the main content page in size.  About the only reason
for this not being true is that very few authors actually separate 
behaviour and styling from content, so content pages are often huge.

Generally, though, given the endemic, and often pointless, uses, by web
authors, of techniques that frustrate caching, and other inefficient
practices, I don't believe that authors are concerned about download

Until authors stop serving cookies with every image, embedding large
amounts of styling in every element open tag, starting every page with a
huge EcmaScript library, and adding meta lines to every HTML file with the
intention, and probably effect, although often without any standardised
meaning, of frustrating caching, I don't think it is worth considering
other browser and server changes to optimise downloads.

(There is also nothing to stop pages being in .mhtml format, although
I've not checked to see whether this works for remote resources.)
Received on Sunday, 5 November 2006 14:02:43 UTC

This archive was generated by hypermail 2.4.0 : Thursday, 30 April 2020 16:21:01 UTC