W3C home > Mailing lists > Public > whatwg@whatwg.org > February 2007

[whatwg] several messages about HTML5 -- authors' tools

From: ddailey <ddailey@zoominternet.net>
Date: Thu, 22 Feb 2007 08:14:25 -0500
Message-ID: <003601c75683$60778aa0$6401a8c0@disxgdg31szkx7>
Interesting thread (including various sub-ravels thereof).

Suppose in a semantically charged, but markup-impoverished medium such as 
the textual narrative (constituting the majority of the content of the web 
as we know it), we seek to build the word processor that generates not only 
the surface structure (the sentences and paragraphs) but the semantic 
structure as well. How do we minimize the author's effort? Authors will not 
want to write both their utterances and the translation of those utterances 
into semantic tags -- it is simply too labor intensive (unless we care more 
about form than substance and chose to purge ill-formed ideas from the human 
corpus).

Rather we may seek a word-processor that deduces semantics from authors' 
expressions. Yeah, without the existence of full-blown AI (which has been a 
while in coming, now), 40% (or so) of such deductions will be incorrect. But 
suppose following the creation of a sentence, or a paragraph, or a larger 
chunk of text, semantically enabled software were to pose to the author a 
finite (and hopefully small) number of "deductive disambiguators"?

"Dear author, did you mean to imply that AI will indeed, be arriving soon?"

"Dear author, who exactly does 'we' refer to in the above paragraph -- I'm 
sorry, but I see no people mentioned before?"

Using a relatively simply inference engine (SFOL + set theory + predicate 
calculus + arithmetic + time + causality + modal logic) coupled with 
thesauri and parsers (all available client-side these days), and (most 
importantly) the author's expert intervention,  I rather suspect that the 
40% (incorrect deductions) could be brought down to 8% with an additional 
cost of 20% in authorial time investment. With current software that most 
folks use, and requiring authors to generate their own semantics, I think we 
might expect to achieve 5% spurious deduction with 400% additional 
investment of authors' time. The cost-benefit ratio is just too high with 
current desktop tools.

In semantically impoverished (not in the evocative space it engenders, but 
in the surface expression of its utterance) but markup-rich environments 
such as SVG, the generation of a parallel semantic substrate is going to be 
a lot more difficult, but maybe that's why we have things like sXBL: to 
allow semantics to be imported from other disciplines.

That's one approach. Another is to build a semantic expression system for 
which we abandon our native languages and agree to write in a semantic 
shorthand (with lots of parentheses, by the way). For even one language, the 
task of finding a minimal set of semantic primitives (from its monolingual 
dictionary) is NP-complete, but if we seek such a shorthand to span the 
space of human semantics, it may take longer to bring into existence than AI 
itself . The different language families I have looked at probably share a 
core semantics of only about 20% of the expressive space of any one language 
by itself. The nice thing about such languages is that people from different 
linguistic backgrounds can all read the same text; the hassle is that it's 
hard to translate ordinary expressions into such languages.

cheers,
David Dailey

----- Original Message ----- 
From: "Elliotte Harold" <elharo@metalab.unc.edu>
To: "Ian Hickson" <ian at hixie.ch>
Cc: <whatwg at lists.whatwg.org>; "Vlad Alexander (xhtml.com)" 
<vlad.alexander at xhtml.com>
Sent: Wednesday, February 21, 2007 4:34 PM
Subject: Re: [whatwg] several messages about HTML5


> Ian Hickson wrote:
>
>> The original reason I got involved in this work is that I realised that 
>> the human race has written literally billions of electronic documents, 
>> but without ever actually saying how they should be processed.
>
> That's a feature, not a bug.
>
>> If, in a thousand years, someone found a trove of HTML documents and 
>> decided they
>> would right an HTML browser to view them, they couldn't do it! Even with 
>> the existing HTML specs -- HTML4, SGML, DOM2 HTML, etc -- a perfect 
>> implementation couldn't render the vast majority of documents as they 
>> were originally intended.
>>
>
> Authorial intent is a myth. Documents don't have to be rendered like the 
> author intended, nor should we expect them to be.  We don't read Homer 
> like Homer intended, but we still read him, well more than a thousand 
> years later. (For one thing Homer actually intended that people listen to 
> the poems, not read them.)
>
> This is not to say that I don't think it's useful to define a standard 
> tree structure for documents. It is useful. However the benefit of this 
> exercise is not in maintaining authorial intent. That's tilting at 
> windmills, and will never succeed no matter what we do.
>
> -- 
> ?Elliotte Rusty Harold  elharo at metalab.unc.edu
> Java I/O 2nd Edition Just Published!
> http://www.cafeaulait.org/books/javaio2/
> http://www.amazon.com/exec/obidos/ISBN=0596527500/ref=nosim/cafeaulaitA/
>
> 
Received on Thursday, 22 February 2007 05:14:25 UTC

This archive was generated by hypermail 2.3.1 : Monday, 13 April 2015 23:08:32 UTC