- From: Henry S. Thompson <ht@inf.ed.ac.uk>
- Date: Sun, 06 Nov 2011 17:52:44 +0000
- To: www-tag@w3.org
Now available online at http://www.w3.org/2001/tag/2011/11/04-minutes.html and below in text form. ht ------------ [1]W3C - DRAFT - TAG f2f 04 Nov 2011 [2]Agenda See also: [3]IRC log Attendees Present TAG members: Dan Appelquist, Tim Berners-Lee, Yves Lafon (for item 1), Peter Linss, Ashok Malhotra, Larry Masinter, Noah Mendelsohn, Jonathan Rees (for items 2 & 3), Henry S. Thompson; Guests: Mike Belshe (for item 1), Alexandre Bertails (for items 1 & 3), Chris Lilley (for item 3), Peter St. Andre (for item 1), Rigo Wenning (for item 2) Regrets Jeni Tennison Chair Noah Mendelsohn Scribe Henry S. Thompson Contents * [4]Topics 1. [5]SPDY 2. [6]Publishing and Linking on the Web 3. [7]3023bis -- Media type registration for the XML family 4. [8]Web Storage 5. [9]Admin * [10]Summary of Action Items __________________________________________________________________ SPDY <stpeter_> [11]http://www.ietf.org/proceedings/80/slides/httpbis-7.pdf (SPDY slides from IETF 80) NM: [introduces the TAG to MB] MB: I was at Google until 2 months ago, worked on the Chrome team, we started SPDY around 2008 ... Google focus on performance, so interested in protocol speedup ... Using the existing mechanisms in HTTP was just gnarly ... So we started experimenting in the lab with doing something of our own ... but based on a lot of prior work in a lot of areas ... SPDY is beginning to spread -- Firefox have started some work ... to date we've owned it, published an informal spec., some unit tests ... Firefox adoption possibility has pushed us towards standardization ... Interop guarantee is necessary before we can move forward MB: Roberto Peon is the person at Google who is on point for SPDY now ... We looked at taking this to the IETF, and that took me to PStA NM: We got in to this also in part because of our contact with Jim Gettys over the buffer bloat issue MB: State of play -- Google Chrome now using SPDY for all SSL traffic, see [12]about:net-internals ... Firefox is implementing it ... Amazon Kindle Fire recently announced that they will be using SPDY ... A number of other less big names involved, implementations in Python and Ruby, etc. ... Main parts of SPDY: 1. Multiplexing; 2. Compression; 3. Prioritization; 4. [Maybe] Server push MB:Primary focus is on improving page-load latency MB: Browser use of HTTP and TCP involves various attempts to game the TCP expected behaviour, particularly in the area of multiple connections ... SPDY tries to avoid this by addressing multiplexing, with prioritization, directly <Yves> problem in doing multiplexing at the spdy level (or httpmux earlier) is the bad interaction that might happen between the tcp window size and the chunk size at l7. MB: Google research found that when there are packet losses, having two connections is a real win ... Multiplexing uses fewer connections which overall simplifies things ... Performance == minimal latency NM: Trying to get this via multiple connections did seem to make things worse MB: Wins with NAT as well ... Two connections is already recommended in HTTP1.1 <Yves> note that the number of // connections in http has been removed in httpbis (as it was not relfecting the reality of things) MB: When that got multiplied by separate hosts for e.g. js and jpg, and then 2 per went to 6 per, suddenly we were up to 12 --- 18 connections for a single page TBL: TCP will back off and reroute if things get stuck MB: We took a serious approach to this ... The average hosts hit per page is 8, rising 9 ... and the size of things retrieved has grown too ... Browsers are trading off resource fetching against page load performance ... SPDY was trying to take that optimization off of browsers' backs <Yves> having mux helps against multiple tcp conn and badly implemented http pipelining. MB: Important because the difficulty of modelling web pages has grown enormously ... JS, CSS have execution times, so building benchmarks for web page latency is very tricky ... Didn't want to just use Chrome, because it already has a set of decisions built in ... But building a platform from scratch for benchmarking was too big a job ... So we did in the end build plugins for Chrome to benchmark SPDY and HTTP side-by-side ... Very glad that we have Firefox now doing their own similar work ... First three goals above are parallel to HTTP -- GET etc. ... Server push is different/new, which requires client- and server-level rework ... We tried some experimental services using push ... The cache is an issue ... You have to understand the service/application detail, and you only get rid of one round-trip ... Doesn't look like there's enough energy HT: There are apps that make sense to build if you don't have to busy wait. MB: which is kind of what we have today HT: In HTTP 1.1 you can hang on a get. NM: Yes, comet. HT: But there were too many glitches. MB: Yes -- GoogleDocs does mutual refresh by hanging gets ... So if you have 30 GoogleDocs open, there's a problem with the limit of 6 connections per host ... So GoogleDocs fakes it with multiple hosts :-( ... SPDY multiplexing is enough to fix this TBL: Shared editing experience everywhere would be really good MB: Hanging get over SPDY is cheap, and does the job, so Server Push doesn't seem so urgent ... Server push is server-initiated MB: Why SSL? Well, SPDY doesn't require SSL ... First choice we had to make was TCP or UDP -- TCP, to save hassle ... So, what port? 80 or 443 -- pipelining not really there for 80, pblms with proxies, slowly getting sorted out NM: Problem would be that new stuff would confuse proxies if it went through port 80, right? MB: Right ... So we went to 443 ... And in any case, we came to feel that securing the Web was independently a good thing [Various]: Certificates are broken, how to secure the web is a huge issue, maybe not in scope today LM: What about IPv6? TBL: For stuff that doesn't need to be secured, it's very tempting to get to a P2P solution. ... But some things, say NYT front page or a TED video, which are truly public ... must that be destined for P2P, or some other architecture? MB: You're right, we recognize that not everything should go this way ... But operators are not very good at recognizing the difference ... Even the front-page of the NYT is not a simple case, if it's personalized to you on the basis of personal (private) info, then that is important to keep secure <Yves> encrypting everything is a double edged sword... MB: The Google China experience made us all very sensitive to the fact that personal data can in fact be a matter of life and death, and you never know where it's going to turn up <timbl> (ages ago .. TBL: Peter, you said when 1 TCP connection has losses and slows down it is faster to add a second connection. Of course, when there is congestion adding more connections adds to the congestion, which on a large scale when everyone does it once, could have overall very detrimental effect on latency for everyone else. So one should simulate or measure the effect of doing this to everyone on the net ) MB: The whole security layer needs improvement ... both in terms of security as such, and in terms of speed ... What about proxies? <noah> I think this will drive ubiquitous inclusion of SSL acceleration in hardware, something I've thought for a long time would be a good enabling step MB: On the back end, inside firewall, use SPDY w/o SSL MB: Proxies are a good thing, and SPDY doesn't have a story about how to play nice with proxies ... SPDY does not address cacheable secured content ... But everyone is using Content Distribution Networks [CDNs], which have largely overtaken proxies for many large operators ... But this lack is a weakness for SPDY <noah> I think it's really large organizations that are deploying CDNs...you prejudice against the long tail when you assume that everything accessed from distant locations is sourced by a large organization like CNN or Google LM: We need an analysis of what the impact of not being able to cache actually is MB: Right -- what's the impact in aggregate -- even though there are clear cases where it loses on an individual basis NM: For the original pre-CDN world, your ISP got you pages that started a long way away quickly ... That won't work with SPDY LM: Right, so that's why some more global measurement and analysis is required MB: And of course any SSL use today is already not proxied ... With SNI, you can see the hostname, but not in vanilla SSL, which makes virtual hosting difficult <Yves> java doesn't have SNI as well :( LM: Corporations may not be happy with the loss of filtering capability that follows from ubiquitous SSL usage PL: Moving to signed content is another important avenue to look at, necessary for peer to peer failover for http MB: There are a lot of horror stories out there from big sites about proxy badnesses ... SSL removes that vulnerability ... The mobile operators have this lose-lose tradeoff between idiosyncratic compression (fast but potentially bad on the device) vs. not (slow but reliable on the device) ... Patrick McManus of Firefox has looked at some numbers ... 83 connections for the NYT home page puts really bad pressure on NAT ... But the NAT things cuts both ways -- dependence on a single channel makes NAT dropout more noticeable/serious MB: Speculation about Kindle Fire -- you could push the multiplexing out to the Amazon connection point at EC2 ... So that all traffic goes via a single connection (over 443) from the Kindle customer ... This appears to contradict the end-to-end story SSL demands ... Requires the notion of trusted proxy -- SSL man-in-the-middle ... So, and explicit proxy: Kindle to EC2, or anyone to their corporate firewall ... Yes, there is a potential for head-of-line blocking, which can amplify in certain cases ... But overall we are still winning ... It's difficult to model this, you have to collect empirical data ... No doubt that with multiple streams, you are more vulnerable ... I think there are some TCP tweaks that can help, we're working on it ... [Stuff about 'slow startup' which scribe didn't get] NM: Adding another stream to SPDY doesn't allow cross-stream fixup, right? MB: Yes ... SPDY does in general fix the head-of-line blocking problem ... Firefox guys have been trying to get pipelining working better, but it's really hard ... They presented at IETF last year [ref?] ... We were pushed to start all the way down at the packet protocol level, but resisted ... We think there's a lot of room to optimize on top of TCP, and we'll only look downwards after that's worked through DKA: So standardization -- What does "take this to IETF in 2012" mean in detail? Should the TAG stay involved, and if so why -- in what way does it impact on Web Arch? NM: Yes, I think TAG should stay involved, but we should discuss this on a telcon <noah> ACTION: Noah to schedule discussion of how, if at all, TAG should continue to be involved with SPDY [recorded in [13]http://www.w3.org/2001/tag/2011/11/04-minutes.html#action01] <trackbot> Created ACTION-626 - Schedule discussion of how, if at all, TAG should continue to be involved with SPDY [on Noah Mendelsohn - due 2011-11-11]. TBL: At the Web level, things such as Content type, the Link header, things like that ... Does SPDY change that? Where are HTTP headers? MB: Almost entirely untouched ... So the framing layer is pretty much all that changes ... We looked at 'improving' some aspects of HTTP -- absolutizing all URIs ... but there are some servers which don't support it TBL: Absolute URIs can indeed be problematic LM: Does HTTP 1.1 really require support? [Various]: Yes MB: Net-net on that -- we backed off doing anything like that ... Not yet sure how we go to IETF, exploring that with PStA LM: SPDY sounds extremely immature to me -- the impact of this on a wide scale, outside Chrome<->Google servers, is just unknown ... W3C used to have resources in this area, it would be good to have W3C involved in taking something like this forward ... Without any guarantee that the outcome will be much like SPDY ... But I think a prerequisite for standardization is more exploration of the requirements and consequences NM: W3C should do this? My sense is that we've been happy with IETF taking the lead LM: Well, comparing two protocols wrt 'page load latency' is a W3C area ... 'Doing this' will involve a lot of different tasks - "What does the Web require in terms of optimization" is a W3C issue, almost by definition TBL: Yes NM: So, push back on taking it to IETF? TBL: No, makes sense to do the protocol there, as LM said ... Low-level question -- Can the client change its mind about priority? MB: Change-priority is not supported, but tab-change in the browser might provoke us to rethink that ... Note that the priorities are advisory, the server can do whatever it likes <timbl> chrome://net-internals/#spdy MB: Wrt standardization and changes -- putting it out there means Google understands that other perspectives are now needed, and will lead to changes NM: Need to stop, so thanks to Mike ... the TAG will come back to this -- who do we feed back to? <noah> [14]http://groups.google.com/group/spdy-dev <noah> Send email to this group: spdy-dev 'at' googlegroups.com HST: I think the W3C IETF liaison needs to be aware of this, and help us decide where the W3C needs to be involved NM: Suspended until 1050 Publishing and Linking on the Web NM: Resuming <noah> [15]http://www.w3.org/2001/tag/doc/publishingAndLinkingOnTheWeb-2011-10 -27.html DKA: [introduces the above doc] ... Questions for Rigo -- Is the overall goal sensible/useful; Are there other terms we should add, e.g. 'performance'; Are there other regulatory/policy issues we should add to the framing? LM: Great start, ready to figure out what the next steps are ... To be useful, this has to be published in a way that gets community consensus ... How do we get there <masinter> the TAG could publish it as a NOTE and start a community group? <jar> but we wanted this to go rec track. <jar> Thinh [Nguyen] says effectiveness much enhanced by rec status NM: Maybe we should schedule detailed review of this at some length at the January F2F <noah> ACTION: Noah to schedule very detailed line-by-line review of Pub&Linking draft at January F2F [recorded in [16]http://www.w3.org/2001/tag/2011/11/04-minutes.html#action02] <trackbot> Created ACTION-627 - Schedule very detailed line-by-line review of Pub&Linking draft at January F2F [on Noah Mendelsohn - due 2011-11-11]. LM: Can we get the TAG out of the critical path? LM: Community group, which might have some lawyers in it RW: The scope of this document is too large <noah> I would prefer to focus on linking RW: Publishing is a nightmare, and linking even more so ... Legal tactic is to partition as much as possible ... Reduce this, make two documents <jar> jar to rw: this is not a legal document DKA: Start with linking? RW: Yes RW: Look at www.linksandlaw.com <jar> linksandlaw.com - I've given out the pointer a few times <noah> [17]http://www.linksandlaw.com/linkingcases.htm RW: there are a very large number of relevant cases LM: Would you come to the TAG for help if you were tasked to do so? RW: Your first document, to get the story clear, is not for the world TBL: Why not for the world? RW: If you talk too much, your central message goes away [tip of the hat to TLR] NM: Advice could focus on "if ... then ..." heads-up kinds of observations RW: There is a long-standing conflict between the technical and legal communities RW: In particular a conflict over ownership of terms ... There are already laws, with terminology definitions <masinter> I think we should take what we have and see how we can make it useful.... LM: I just want to make this useful enough to publish ... Suppose this is just an outline of what the TAG understands in this space ... And accepting that we can't resolve the conflicts ... Could we point out where the conflicts are? RW: If you asked me, I would try to come up with a concise statement of the fact that publication implies the possibility of linking to ... See e.g. the KPMG example from linksandlaw: [18]http://www.linksandlaw.com/linkingcases-linkingpolicies.htm#KPMG <jar> oops, I had promised to write something about "don't link" terms of use <jar> look at the american airlines site (whose URL I can't give you) RW: Not stop this document, to get our own understanding clear ... but also publish simple short observations that are at the core DKA: But that's a legal statement NM: We set out to avoid making policy statements ... We can't state that publishing gives a right to others to link RW: Not a right, but an expectation TBL: You're preaching to the choir, but how do we say this? RW: We can distinguish between linking itself and the existence of access control NM: [Hypothetical KPMG example] ... How do we make it a technical observation, not a policy one RW: The fact that you include a pointer to something on the Web in your document has no meaning for the content over there, and is completely unrelated to the thing identified TBL: If you can from the KPMG home page browse to another page, I should be able to pass that link to someone else ... The UK position appears to be that publishing a link collection to pirate music sites is to be an accessory to copyright theft ... We could try reciprocal banning, by notifying KPMG that they are not allowed to read the W3C site :-) JAR: Wrt what RW and LM said, the original idea, from Thinh Nguyen, was that it would be useful, to forestall bad decisions, if there was a document that simply stated what the technical community thinks these terms mean. ... Thinh went on to say that to get the necessary impact, it needs to go out as a REC ... So it doesn't try to argue with the law, it just says what the technical understanding of these terms is <jar> session with Thinh: [19]http://www.w3.org/2001/tag/2010/12/02-minutes.html#item01 ... DKA: So we started from there, and the fact that a URI is public identifier, to get to the parallel between speech acts and URI use, and that brings in the 'right to link', parallel to free speech ... from which we got drawn in to the distinction between linking and embedding ... And getting embedding clear requires us to get the publishing/hosting distinction clear ... Because it isn't clear that a page with text and video involves multiple sources RW: The legal side knows about this DKA: So, maybe we don't need all of that? NM: We're getting different advice from different legal sources. . . AM: Alternatively, maybe we should not publish a TAG doc't, but something in the popular space, e.g. the NYT over Tim's byline <masinter> so we should make a short statement around which we can get consensus, focused on one small issue around linking and copyright <masinter> at least that's what Rigo is advocating RW: The TAG endorsement of a short statement about the passive linking case which is that URIs are constitutive of the Web and publishing a page with a URI creates a legitimate expectation of linkability AM: If we did a short statement, how do we get it out? To make an impact? <masinter> the current document is useful for us, but way too broad. RW: The publication channel of the TAG itself should be ordinary, which has our opinion LM: It could be a finding RW: And then you go to the NYT and say "The TAG has said. . ." NM: If the W3C/the membership/TBL want to say [a quasi-hortatory claim about the right to link], that's fine, but it's not for the TAG to do so ... The TAG does architecture TBL: But the architecture has a fundamental social component <masinter> i think this can be a finding, making a statement about the architectural assumption of the web and underlying many of the TAG's other activities is that linking is fundamental, that providing URIs for content is recommended, exactly to promote linking TBL: Spam is a social violation of the architecture ... So I think the TAG can speak on this subject in social terms NM: Yes, I'm comfortable with talking about social value, and the impact of policy on value TBL: Forbiding incoming links breaks the [social] system NM: But there are clearly (jurisdictional bounded) cases where linking to unacceptable material is itself unacceptable LM: We haven't said anything about laws TBL: But we are close to that in saying KPMG are doing something wrong RW: Wrt DKA's point, using the free speech analogy is going further that I would <masinter> we have designed the system such that linkability is a benefit. Attempts to restrict linkability is counter to the effective use of the system as designed. RW: I would focus on the passive case: it's wrong for sites to pubish rules which forbid linking to them <masinter> we don't have to say "the law is a bad idea", we have to point out the negative consequences of such laws <masinter> i would still like to see Jeni/Dan's document published as a note after review, even if we also have a finding about common use of linking and the TAG's position on it <jar> Thinh Nguyen: default rules are important in court... may be supplied by technical standards (that was Thinh last December) LM: Trying to avoid value judgements, but just say what the consequences of doing so would be <jar> The consequences speak for themselves. Generally speaking. NM: I want to focus on the consequences of laws which enable behaviours HST: That's much narrower than I hear from JAR and LM, who are happy to discuss consequences of actions, not just laws (or not laws at all) NM: Look at the KPMG case -- it's going to get legal very quickly ... Immediately the question will arise as to whether their statement is enforceable by law RW: As Thinh Nguyen said, the legal interpretation will be based on usage, on community habits RW: Is the expectation in practice that certain things hold? ... So for example the attempt to smuggle in obligations via shrink-wrap was surprising, contrary to expectation, and so not supported by the courts RW: Similarly the expectation is that publishing something on the Web gives the possibility of linking <masinter> it gives the possibility of linking, and we've encouraged that JAR: I actually do think it makes sense for a technical person to say that "operating under such-and-such a restriction would have the following [bad] consequences", so I'm not ruling out laws <noah> +1 to what Jonathan just said <masinter> i'm trying to distinguish between "such restrictions are bad" and "negative consequences of such restrictions are X, Y, Z; such restrictions are bad if they do not have clear redeeming value" <masinter> the web was designed for X, it was built with this assumption HST: OK, I see a continuum, was a mistake to try to push for two distinct positions DKA: Saying "restricting linking will have a detrimental effect on the Web" is weak from a legal perspective -- The lawyer will say "Not my problem" <noah> What I want to rule out, mostly, is: "law XXX should not be passed". I would rather say: "if you pass law XXX, you should understand that the consequences to the operation of the system, and to its positive social value will be YYY" <masinter> restrictions on linking are impossible to accomplish because of web architecture? search, robots.txt, harvesting, ...? <masinter> we should be clearer about who the primary audience is for this finding <masinter> maybe a blog post? DKA: We're not asking for unrestricted freedom to link, but no less freedom than the freedom of speech NM: The speech parallel is weak, because URIs don't point to consistent things necessarily ... I prefer the address analogy NM: Consider the address of [a prohibited organization] in a public works document about street repairs versus in a list of recommended destinations NM: So consider a log of URIs versus a Web page which references one LM: Summarizing RW -- the scope of this document is too broad, you should find a few one-page extracts DKA: I'll take this back to Jeni and consider all of the input we've gotten ... And decide whether to take this forward broadly but internally ... or whether we can pare it down effectively NM: There are clearly different opinions about: ... 1) What the goal of the document is; ... 2) What its scope is. ... We should acknowledge the lack of consensus, and maybe the divergence of advice we're getting RW: Pursuing (w/o publishing) this document, will improve the value of TAG utterances in the future ... That the passive case: I'm a site, and I forbid linking, is wrong is what you should say ... not the active "I have a right to link" ... The latter gets quickly extremely messy HST: I'm not sure sending the editors off to work harder when the TAG hasn't agreed on scope is an invitation to waste time DKA: I wasn't going to dive right in to cut the document down -- I want to work with all the feedback we've gotten this week, particularly on Wednesday, and that's where I want to focus LM: So a new draft is worth working on if it yields something we can publish NM: Remember Goals and Success Criteria -- we should keep these ([20]http://www.w3.org/2001/tag/products/PublishingLinking.html) in mind ... And consider revising them too <noah> ACTION: Appelquist with help from Jeni to propose changes to goals, success criteria etc. for publishing/linking product page [recorded in [21]http://www.w3.org/2001/tag/2011/11/04-minutes.html#action03] <trackbot> Created ACTION-629 - With help from Jeni to propose changes to goals, success criteria etc. for publishing/linking product page [on Daniel Appelquist - due 2011-11-11]. LM: Right -- for example split the work between a small 'official' publication and a larger background unofficial 'white paper' NM: Thank you Rigo 3023bis -- Media type registration for the XML family CL: I've mailed a summary of recent progress on 3023bis to www-tag ... Previous draft deprecated text/xml ... Implementors pushed back, as we have to support it even if new authors don't use it ... I took this to IETF80, got lots of interaction ... HTTPbis then removed the default charset handling rules, which is consistent with applications today ... But that left email out of sync with HTTP ... But but it now appears there is willingness to fix the text/... story to allow individual text/ media types to declare their own charset rules ... So text/xml can say there is no default charset and the XML spec rules determine [which will apply to email as well as HTTP] LM: So, what is in the way of publication? CL: On this front, nothing, but the fragid situation is still pending JAR: xhtml+xml can have RDFa in it HST: What's the problem with that? JAR: The case is this -- you have a URI with a fragid, and you want to follow your nose ... If you look at the 3023bis registration, it says XPointer tells you the semantics ... so you look for an ID, and don't find one ... So there's an error (per XPOinter) HST: The no new syntax move, for SVG: ... if syntax is not a valid xpointer, then defer to the specific media type reg ... the xhtml media type has to say something about RDFa ... and 3023 needs an override policy jar: the new piece for me was realizing that we need to also talk about application/xhtml+xml HST: 3023bis says it is the definitive spec for fragids ... what it needs to do is to say under what circumstances it defers to the subsidiary media type reg <masinter> why is this on critical path for 3023bis ? ACTION Henry to work with Chris Lilley to bring forward prose for 3023bis wrt generic processing of fragment identifiers which addresses the rdf+xml and xhtml+xml issues <trackbot> Created ACTION-628 - Work with Chris Lilley to bring forward prose for 3023bis wrt generic processing of fragment identifiers which addresses the rdf+xml and xhtml+xml issues [on Henry Thompson - due 2011-11-11]. ACTION-628 due 2011-12-31 <trackbot> ACTION-628 Work with Chris Lilley to bring forward prose for 3023bis wrt generic processing of fragment identifiers which addresses the rdf+xml and xhtml+xml issues due date now 2011-12-31 <noah> That's fine, Henry, assuming you'll also put in the cover email a bit of framing to remind something about the history and context for whatever is proposed LM: Isn't this a general point about mixins <masinter> use of RDFa as a mixin applies to more things, like JSON LM: Can't we get some cleaner layering? NM: In the interest of getting 3023bis out, can't we do this locally first, in 3023bis, on the assumption that it will be consistent with any cleaner general solution? LM: Getting mixins right is likely to be a huge problem, you're going to get stuck in a tarpit CL: You may be right, we'll see LM: The way to get follow-your-nose for mixins requires you to modify the host-language media type definitions systematically HST: In my recent email I distinguish between "witting" and "unwitting" host languages ... The unwitting case is when an XML language (e.g. XML Schema, XSLT, XML Signatures) allows any attribute on any element, as long as it's in a namespace that's not theirs HST: The witting cases work easily, but the unwitting are important. They allow mixins without explicitly naming them in their specs, so they can't be expected to change the media type registration <masinter> are there other mixins other than RDFa that add fragment identifier possibilities? <masinter> can i mix-in SVG into something and use SVG visible objects as fragment targets? HST: 'Unwitting' embedding of RDFa can't trigger a change to a media type registration LM: Would unwitting embedding of e.g. SVG introduce the possibility of using SVG-style fragids which only work in the SVG sub-part? NM: There is at least some discussion of that case in the Self-describing Web finding TBL: This comes under the XML functions story as well LM: So, go ahead, but be aware that there be dragons Web Storage AM: There is a Web Storage Last Call WD just out ... We need to decide whether we want to comment Admin NM: RDFa vs. Microdata will require our attention wrt HTML WG process by mid-January, we will return to this Summary of Action Items [NEW] ACTION: Appelquist with help from Jeni to propose changes to goals, success criteria etc. for publishing/linking product page [recorded in [22]http://www.w3.org/2001/tag/2011/11/04-minutes.html#action03] [NEW] ACTION: Noah to schedule discussion of how, if at all, TAG should continue to be involved with SPDY [recorded in [23]http://www.w3.org/2001/tag/2011/11/04-minutes.html#action01] [NEW] ACTION: Noah to schedule very detailed line-by-line review of Pub&Linking draft at January F2F [recorded in [24]http://www.w3.org/2001/tag/2011/11/04-minutes.html#action02] [End of minutes] __________________________________________________________________ Minutes formatted by David Booth's [25]scribe.perl version 1.135 ([26]CVS log) $Date: 2011/11/06 17:48:54 $ References 1. http://www.w3.org/ 2. http://lists.w3.org/Archives/Public/www-tag/2011Oct/0119.html 3. http://www.w3.org/2011/11/04-tagmem-irc 4. http://www.w3.org/2001/tag/2011/11/04-minutes.html#agenda 5. http://www.w3.org/2001/tag/2011/11/04-minutes.html#item01 6. http://www.w3.org/2001/tag/2011/11/04-minutes.html#item02 7. http://www.w3.org/2001/tag/2011/11/04-minutes.html#item03 8. http://www.w3.org/2001/tag/2011/11/04-minutes.html#item04 9. http://www.w3.org/2001/tag/2011/11/04-minutes.html#item05 10. http://www.w3.org/2001/tag/2011/11/04-minutes.html#ActionSummary 11. http://www.ietf.org/proceedings/80/slides/httpbis-7.pdf 12. about:net-internals 13. http://www.w3.org/2001/tag/2011/11/04-minutes.html#action01 14. http://groups.google.com/group/spdy-dev 15. http://www.w3.org/2001/tag/doc/publishingAndLinkingOnTheWeb-2011-10-27.html 16. http://www.w3.org/2001/tag/2011/11/04-minutes.html#action02 17. http://www.linksandlaw.com/linkingcases.htm 18. http://www.linksandlaw.com/linkingcases-linkingpolicies.htm#KPMG 19. http://www.w3.org/2001/tag/2010/12/02-minutes.html#item01 20. http://www.w3.org/2001/tag/products/PublishingLinking.html) 21. http://www.w3.org/2001/tag/2011/11/04-minutes.html#action03 22. http://www.w3.org/2001/tag/2011/11/04-minutes.html#action03 23. http://www.w3.org/2001/tag/2011/11/04-minutes.html#action01 24. http://www.w3.org/2001/tag/2011/11/04-minutes.html#action02 25. http://dev.w3.org/cvsweb/~checkout~/2002/scribe/scribedoc.htm 26. http://dev.w3.org/cvsweb/2002/scribe/ -- Henry S. Thompson, School of Informatics, University of Edinburgh 10 Crichton Street, Edinburgh EH8 9AB, SCOTLAND -- (44) 131 650-4440 Fax: (44) 131 651-1426, e-mail: ht@inf.ed.ac.uk URL: http://www.ltg.ed.ac.uk/~ht/ [mail from me _always_ has a .sig like this -- mail without it is forged spam]
Received on Sunday, 6 November 2011 17:55:38 UTC