W3C home > Mailing lists > Public > whatwg@whatwg.org > January 2009

[whatwg] Trying to work out the problems solved by RDFa

From: Calogero Alex Baldacchino <alex.baldacchino@email.it>
Date: Fri, 09 Jan 2009 02:54:08 +0100
Message-ID: <4966AE40.3040600@email.it>
Charles McCathieNevile ha scritto:
> On Sun, 04 Jan 2009 03:51:53 +1100, Calogero Alex Baldacchino 
> <alex.baldacchino at email.it> wrote:
>
>> Charles McCathieNevile ha scritto:
>> ... it shouldn't be too difficoult to create a custom parser, 
>> comforming to RDFa spec and availing of data-* attributes...
>>
>> That is, since RDFa can be "emulated" somehow in HTML5 and tested 
>> without changing current specification, perhaps there isn't a strong 
>> need for an early adoption of the former, and instead an "emulated" 
>> mergence might be tested first within current timeline.
>
> In principle this is possible. But the data-* attributes are designed 
> for private usage, and introducing a public usage means creating a 
> risk of clashes that pollute RDFa data gathered this way. In other 
> words, this is indeed feasible, but one would expect it to show that 
> the data generated was unreliable (unless privately nobody is 
> interested in basic terms like about). 

This is why I was thinking about somewhat "data-rdfa-about", 
"data-rdfa-property", "data-rdfa-content" and so on, so that, for the 
purposes of an RDFa processor working on top of HTML5 UAs (perhaps in a 
test phase, if needed at all, of course), an element dataset would give 
access to "rdfa-about", instead of just "about", that is using the 
prefix "rdfa-" as acting as a namespace prefix in xml (hence, as if 
there were "rdfa:about" instead of "data-rdfa-about" in the markup).

This way, the public exposure of RDFa attributes on top of the generic 
and normally-private dataset feature might be enough circumscribed to 
avoid clashes. That is, if RDFa shows its best benefits when used to 
address small-scale needs involving trusted/reliable (meta-)data, it 
should be fair to assume all involved parties are aware that each one is 
using RDFa, and aren't just running an RDFa processor in the hope to 
gather enough informations.

 From this point of view, it should be quite unlike to find people using 
"data-rdfa-about" to express different semantics in the same page 
(whereas data-property might cause ambiguity, for instance), as well as 
it is (or should be) quite unlike to find namespaces using the very same 
prefix involved in the same xml document (that is, I think choosing a 
name including a namespace prefix for a data-* attribute (and also for a 
class in a generic container as a div or a span, to tell it represents 
an external element) can replicate quite safely the xml extensibility 
for custom uses, to some extent, without requiring a wide support for it 
in text/html document - since it seems that xhtml extensibility is not a 
major concern, at least not enough to be worth merging it into html).

Just an idea, though.

However, AIUI, actual xml serialization (xhtml5) allows the use of 
namespaces and prefixed attributes, thus couldn't a proper namespace be 
introduced for RDFa attributes, so they can be used, if needed, in 
xhtml5 documents? I think such might be a valuable choice, because it 
seems to me RDFa attributes can be used to address such cases where 
metadata must stay as close as possible to correspondent data, but a 
mistake in a piece of markup may trigger the adoption agency or foster 
parenting algorithms, eventually causing a separation between metadata 
and content, thus possibly breaking reliability of gathered 
informations. From this perspective, a parser stopping on the very first 
error might give a quicker feedback than one rearranging misnested 
elements as far as it is reasonably possible (not affecting, and instead 
improving, content presentation and users' "direct" experience, but 
possibly causing side-effects with metadata).

Also, if the above is true, using namespaced and prefixed attributes 
instead of ones laying in the same namespace shared both by html5 and by 
xhtml5 (in theory) might prevent the use of such metadata in a document 
whose parsing rules might lead to possible side-effects.

> Such results have been used to suggest that poorly implemented 
> features should be dropped, but this hypothetical case suggests to me 
> that the argument is wrong, and that if in the face of reasons why the 
> data would be bad people use them, one might expect better usage by 
> formalising the status of such features and getting decent 
> implementations.
>

Generally speaking, I think reasoning in terms of "poor implementation" 
vs "rare usage" is quite like moving as a dog biting his own tail, 
because poorly implemented features are forcedly rarely used, and rarely 
used features can't convince UAs developers to implement them (in 
general). But, if a feature is widely needed, several hacks may born, 
thus providing an evidence of a global problem to be solved in a certain 
manner by implementing a certain, well-conceived feature.

As far as I've understood it, that's the main guideline to change actual 
specification, which is moving on the base of a bullet-tracing evolution 
(perhaps weighted on the need for completely new features, as a balance 
between the need for innovation and that for backward compatibility), 
rather than a "cathedral-wise" definition of what is or can be a useful 
feature to be implemented. For this reason, I think that mapping RDFa 
attributes on data-rdfa-* attributes to experiment a convergence between 
RDFa attributes and html5 specific features might be a start point to 
get RDFa attributes both specified and widely supported by 
implementations (either as they're defined in W3C Recommendation, or in 
the form of data-rdfa-*, hence dealt with differently from data-* 
attributes, for backward compatibility with such early implementations - 
a slightly different (or somehow prefixed) name shouldn't be much of a 
problem, as far as the name is not a problem per se (e.g. it is not 
prone to clashes) and allows a one-to-one correspondence).

However, if a custom/small scale solution met a wide support and a deep 
integration into major browsers, maybe misuses and abuses (which a 
proper formalisation couldn't prevent) might become widespread, thus 
making disadvantages (appear or be) greater than advantages, if measured 
on a wider scale (the same as the implementation). Therefore, I think a 
good start point can consist of partly introducing support on top of 
existing features (in the case of RDFa, either through well-groomed, 
custom data-* attributes in html5, or by defining a proper namespace 
with a proper prefix for xhtml5), without requiring a deep integration 
of a processor for the new feature, but instead letting it be a (custom) 
plugin/extension, or an api for a (custom) web application needing it - 
since a person just wishing to get access to some content without caring 
of metadata and metadata reliability could just visit a page, while an 
organisation wishing to interchange RDFa modelled data with another one 
can run a separate processor (eventually a webapp based on a browser 
built-in API, or a plugin, to create a suitable interface for queries) 
to extract and merge informations.

>>>> What is the cost of having different data use specialised formats?
>>>
>>> If the data model, or a part of it, is not explicit as in RDF but is 
>>> implicit in code made to treat it (as is the case with using scripts 
>>> to process things stored in arbitrarily named data-* attributes, and 
>>> is also the case in using undocumented or semi-documented XML 
>>> formats, it requires people to understand the code as well as the 
>>> data model in order to use the data. In a corporate situation where 
>>> hundreds or tens of thousands of people are required to work with 
>>> the same data, this makes the data model very fragile.
>>>
>>
>> I'm not sure RDF(a) solves such a problem. AIUI, RDFa just binds 
>> (xml) properties and attributes (in the form of curies) to RDF 
>> concepts, modelling a certain kind of relationships, whereas it 
>> relies on external schemata to define such properties. Any 
>> undocumented or semi-documented XML formats may lead to misuses and, 
>> thus, to unreliably modelled data,
> ...
>
>> I think the same applies to data-* attributes, because _they_ 
>> describe data (and data semantics) in a custom model and thus _they_ 
>> need to be documented for others to be able to manipulate them; the 
>> use of a custom script rather than a built-in parser does not change 
>> much from this point of view.
>
> RDFa binds data to RDF. RDF provides a well-known schema language with 
> machine-processable definition of vocabularies, and how to merge 
> information between them. In other words, if you get the underlying 
> model for your data right enough, people will be able to use it 
> without needing to know what you do.
>
> Naturally not everyone will get their data model right, and naturally 
> not all information will be reliable anyway. However, it would seem to 
> me that making it harder to merge the data in the first place does not 
> assist in determining whether it is useful. On the other hand, certain 
> forms of RDF data such as POWDER, FOAF, Dublin Core and the like have 
> been very carefully modelled, and are relatively well-known and 
> re-used in other data models. Making it easy to parse this data and 
> merge it, according to the existing well-developed models seems valuable.
>

I admit I'm not very expert in RDF use, thus I have a few questions. 
Specifically, maybe I can guess the advantages when using the same 
(carefully modelled, and well-known) vocabulary/ies; but when two 
organizations develop their own vocabularies, similar yet different, to 
model the same kind of informations, is merging of data enough? Can a 
processor give more than a collection of triples, to be then interpreted 
basing on knowledge on the used vocabulary/ies?

I mean, I assume my tools can extract RDF(a) data from whatever 
document, but my query interface is based on my own vocabulary: when I 
merge informations from an external vocabulary, do I need to translate 
one vocabulary to the other (or at least to modify the query backend, so 
that certain curies are recognized as representing the same concepts - 
e.g. to tell my software that 'foaf:name' and 'ex:someone' are 
equivalent, for my purposes)? If so, merging data might be the minor 
part of the work I need to do, with respect to non-RDF(a) metadata (that 
is, I'd have tools to extract and merge data anyway, and once I 
translated external metadata to my format, I could use my own tools to 
merge data), specially if the same model is used both by mine and an 
external organization (therefore requiring an easier translation).

Thus, I'm thinking the most valuable benefit of using RDF/RDFa is the 
sureness that both parties are using the very same data model, despite 
the possible use of different vocabularies -- it seems to me that the 
concept of triples consisting of a subject, a predicate and an object is 
somehow similar to a many-to-many association in a database, whereas one 
might prefer a one-to-many approach - though, the former might be a 
natural choice to model data which are usually sparse, as in a document 
prose.

>
>>> Ian wrote:
>>>> For search engines, I am not convinced. Google's experience is that
>>>> natural language processing of the actual information seen by the 
>>>> actual end user is far, far more reliable than any source of metadata.
>>>> Thus from Google's perspective, investing in RDFa seems like a poorer
>>>> investment than investing in natural language processing.
>>>
>>> Indeed. But Google is something of an edge case, since they can 
>>> afford to run a huge organisation with massive computer power and 
>>> many engineers to address a problem where a "near-enough" solution 
>>> brings themn the users who are in turn the product they sell to 
>>> advertisers. There are many other use cases where a small group of 
>>> people want a way to reliably search trusted data.
>>>
>>
>> I think the point with general purpose search engines is another one: 
>> natural language processing, whereas being expensive, grants a far 
>> more accurate solution than RDFa and/or any other kind of metadata 
>> can bring to a problem requiring data must never need to be trusted 
>> (and, instead, a data processor must be able to determine data's 
>> level of trust without any external aid).
>
> No, I don't think so. Google searches based on analysis of the open 
> web are *not* generally more reliable than faceted searches over a 
> reliable dataset, and in some instances are less reliable.
>
> The point is that only a few people can afford to invest in being a 
> general-purpose search engine, whereas many can afford to run a 
> metadata-based search system over a chosen dataset, that responds to 
> their needs (and doesn't require either publishing their data, or 
> paying Google to index it).
>

My point is that possible assumptions over datasets reliability is the 
borderline between wide-scale data extraction/classification, which is 
the main problem solved by a general purpose search engine, and implies 
the best assumption by default is datasets reliability is uncertain, and 
(very) small-scale data modelling, were a direct and immediate 
evaluation over datasets reliability is possible and easy to do, so that 
a custom search engine could reliably be based on such metadata. I think 
no comparison is possible between the two scales, thus no generalization 
is possible when trying to guess whether metadata can do more good than 
harm, but instead each case should be analysed separately, and everyone 
should agree which one is the best context (eventually both) where RDFa 
should be used, to understand what's the best way to implement it and if 
it's worth to be introduced in html5 -- as far as I can tell, both of us 
agree that small-scale is the main context.

But perhaps some edge-side case should be considered to draw a better 
picture. For instance, one such case might be a browser availing of 
metadata to search a resource in its local history, or within a web page 
and related/linked pages (to a certain digree and level of depth), 
because its scale would be small with respect to the effective number of 
scanned resources, but wide with respect to the potential number of 
sources for those resources, that is, because a browser implementing a 
metadata extraction and merging engine and a query interface to look for 
gleaned informations would deal with a | limited number | of | 
etherogeneous sources | at a given time.

Once major browsers provided (and exposed by default) such a 
functionality, a growing number of users would (try and) use it, thus a 
growing number of sites would experiment metadata. At the beginning 
everything might work fine, since only honest sites would experiment 
honest metadata (such as wikis, for instance), but once the number of 
sites and users availing of metadata reached a threshold point spammers 
would start including spam metadata in their sites (with otherwise 
trustful content) and in other sites through advertisements. Such a 
scenario might lead to a bad balance between benefits and disadvantages 
for the average user, thus pushing (some) browser vendors to limit or 
even to wholly drop native support, and I guess this is not a wishable 
solution for the Semantic Web Industry.

That is, choosing a proper level of integration for RDF(a) support into 
a web browser might divide success from failure. I don't know what's the 
best possible level, but I guess the deepest may be the worst, thus 
starting from an external support through out plugins, or scripts to be 
embedded in a webbapp, and working on top of other feature might work 
fine and lead to a better, native support by all vendors, yet limited to 
an API for custom applications -- whereas any changes to html to include 
RDFa attributes would be fully meaningful if leading to a full support 
and exposed features to avail of metadata, which I don't think is much 
of a benefit for the great majority of (home) users.

Everything, IMHO

WBR, Alex

 
 
 --
 Caselle da 1GB, trasmetti allegati fino a 3GB e in piu' IMAP, POP3 e SMTP autenticato? GRATIS solo con Email.it http://www.email.it/f
 
 Sponsor:
 Meetic: il leader italiano ed europeo per trovare l'anima gemella online. Provalo ora
 Clicca qui: http://adv.email.it/cgi-bin/foclick.cgi?mid=8291&d=9-1
Received on Thursday, 8 January 2009 17:54:08 UTC

This archive was generated by hypermail 2.4.0 : Wednesday, 22 January 2020 16:59:09 UTC