W3C home > Mailing lists > Public > whatwg@whatwg.org > February 2012

[whatwg] Caching of identical files from different URLs using checksums

From: Ashley Sheridan <ash@ashleysheridan.co.uk>
Date: Fri, 17 Feb 2012 18:39:42 +0000
Message-ID: <1329503983.2640.2.camel@localhost.localdomain>
On Fri, 2012-02-17 at 19:05 +0100, Julian Reschke wrote:

> On 2012-02-17 09:42, Sven Neuhaus wrote:
> > Hello,
> >
> > as of 2012, some websites are including popular javascript libraries from CDNs, like
> > Google's. The benefits are:
> >
> > * Traffic savings for the site operator because the javascript libraries are downloaded from
> >    the CDN and not from the site that uses them
> > * If enough sites refer to the same external file, the browser will cache the file and even if
> >    it's a first visit, the (potentially large) javascript file will not have to be downloaded.
> >
> > There are however some drawbacks to this approach:
> >
> > * Security: The site operator is trusting an external site.  If the CDN serves a malicious file
> >    it will directly lead to code execution in browsers under the domain settings of the site
> >    including it (a form of cross site scripting).
> > * Availability: The site depends on the CDN to be available. If the CDN is down the site may not
> >    be available at all.
> > * Privacy: The CDN will see requests for the file with HTTP referer headers for every visitor
> >    of the site.
> > * Extra DNS lookup if file is not already cached
> > * Extra HTTP connection (can't use persistent connection because it's a different site) if file is not cached
> >
> > I am proposing a solution that will solve all these problems, keep the benefits and offers
> > some extra advantages:
> >
> > 1. The site stores a copy of the library file(s) on its own site.
> > 2. The web page includes the library from the site itself instead of from the CDN
> > 3. The script tag specifies a checksum calculated using a cryptographic hash function.
> >
> > With this solution, whenever a browser downloads a file and stores it in the local cache, it calculates
> > its checksum. The browser can check its cache for an (identical) file with the same checksum
> > (no matter what URL it was retrieved from) and use it instead of downloading the file again.
> >
> > This suggestion has previously been discussed here ( http://lists.whatwg.org/pipermail/whatwg-whatwg.org/2006-November/thread.html#7825 ), however for a different purpose (file integrity instead of caching identical files from different sites) and I don't feel the points raised back then apply.
> >
> > If a library is popular, chances are that many sites are including the identical file and it will
> > already be in the browser's cache. No network access is necessary to use it, improving the users'
> > privacy. It doesn't matter if the sites store the library file at a different URL. It will always
> > be identified by its checksum. The cached file can be used more often.
> >
> > The syntax used to specify the checksum is using the fragment identifier component of a URI
> > (RFC 3986 section 3.5).
> > ...
> 
> Stop here. That's not what the fragment identifier is for.
> 
> Instead, you could specify the hash as a separate attribute on the 
> containing element.
> 
> Best regards, Julian


But wouldn't this all break one of the key reasons why shared libraries
are useful? If I'm using a shared Javascript library hosted on Google
Code, for example, and there's an urgent security update, I'd want it to
be applied and not left out. Also, plenty of the hosted libraries are at
URLs that will grab the latest stable release of something. I don't want
my site to break overnight just because the latest stable release for a
library came out whilst I was asleep.

-- 
Thanks,
Ash
http://www.ashleysheridan.co.uk
Received on Friday, 17 February 2012 10:39:42 UTC

This archive was generated by hypermail 2.3.1 : Monday, 13 April 2015 23:09:11 UTC