- From: <Stephan.Engberg@Priway.com>
- Date: Wed, 27 Mar 2013 18:29:06 +0100
- To: pierre.andrews@gmail.com
- Cc: public-gld-comments@w3.org
- Message-ID: <OFC2FBA932.A4F9AD0E-ONC1257B39.007EC51E@obivision.com>
Hi Pierre,
I may be missing some perspective or mechanism as I only recently started
looking into the work. Notiving the deadline, I decided it was better to
voice my understanding.
> If you check the GLD recommendations that are currently being created,
the standard that is recommended to use is RDF and vocabularies based on
this standard, which I think is quite flexible and able
> to adapt freely with the changing "world reality". If you think this is
not the case, it would be interesting to hear your point of view, even if
I think it's an issue wider than the GLD group.
RDF is indeed very flexible. So flexibile that it is non-operational so
instead we get a lot of low-level one-size-fits-nothing standards.
I see this work as the later. From what I see, the chosen level of
abstraction is too restricting as in ONE standard failling on both
supporting the reality of changes in many different directions in
parallell and especially on security.
In general, I see government creating tomorrows legacy and problems today
making one-size-fits-nothing standards and central structure. We see the
mistakes in reality as in UK, EU, Denmark, Estonia eGovernment etc. where
the shortterm focus on just getting some connection create unflexible,
unsecure and centralised legacy structures. It turns into building
"virtual mainframes" on open structures instead of extending the open
structures to the application layers. Whether it is bad craftsmanship or
ideological regulation is irrelevant as the consequences are real.
>As Richard mentioned, it would be good to understand how these comments
specifically relate to the current GLD work.
What we need are semantic structures that
a) Support control at endpoints instead of centraslisation
b) Support heterogenous resolution and asynchroneus change at runtime.
>From one aspect we need non-ambiguity in order not to kill patients, from
another aspect we need flexiblity to crasp the world wihtout assuming we
can define everything for everyone wihtout ending up in dys-functional and
ineffective structures.
I think that an inpur is the suggestion of a ore dynamic structure based
on gradual standardisation, i.e. a standard for gradually creaing
interoperaiblity between existing systems rather than forcing everything
into one standard.
> If you think there are recommendations that are missing, please be
specific on where and what could be rewritten.
Sorry, I dont have a sponsor. It would require a lot more work to go into
that level of detail. I know it is not particularly helpfull, but that is
the reality.
> I have checked the linksmart sourceforge page, and it's not clear what it
offers and how it solves the issues you have raised.
My point was an analogy.
A number of aspects.
1) Run-time resolution fully data-driven and each "device" able to be
instantiated as different logical devices simultanously
2) Semantic devices can be combined and nested in such a way that any
combination of physcial and logical devices can be created. I dont see
mechanisms to link logics with data to create new semantic data?
3) The implementation do not contain the full structure. My focus were on
semantic resolution of security at runtime mapped towards an overall
security onthology. In security one critical element is that evalutaions
of security propertios can change and thus the need for runtime 3rd party
assertion or status verficiation as part of a semantic link. I dont see
such mechanisms.
4) The structure hide properties, e.g. IPV6 addresses are contained as
they leak information.
5) My real point is the need for heteorogenous semantic definitions. I am
not certain that it is benefiscial to lock structures against ONE
defintion, but rather or in parallelt have mechanisms to link data elemens
or a more loosely coupled manor including the ablut to have multiple
definitions semantically with various degree of formalism. E.g. I should
be able to make smenatic data deinfiions that can change and are resolved
at runtime as we get wiser and the smenatic linkage grow as we dig into
the mechanisms.
E.g. in the HYDRA project we worked to create parameterized security
resolved against an onoltogy of disjunkt security objektives and with
ontology-support to compare various technologies and cryptio-proofs
semantic contribution to security.
> Now, for security of the opened data, it's an interesting issue, and there
is a section about it in the GLD recommendation document. Maybe you want
to make specific comments on this.
I am not sure which section, you refer to.
But e.g. the defintion of a person is clearly hopeless as it assume
identification which means that the structure is incompatible with clould
(as that would require pseudonyms only)
http://xmlns.com/foaf/spec/#term_Person
Similar I cannot find anything dealing with distributed key control and
management. Perimeter security is clearly unable to cope with the
challanges, we need to build security into the data structures and means
of encryption
Try for instance having a look on this for intra-contry interoperability
and security in cloud (where security is not even realistic in theory).
http://digitaliser.dk/resource/896495
Hope the above is usefull.
Semantic interoperability is important but as Einsteain said - make it as
simple as possible, but not more.
Regards
Stephan Engberg
- - - - - - - - - - - - - - - - -
Stephan J. Engberg
Priway - Security in Context
Pierre Andrews <pierre.andrews@gmail.com>
25-03-2013 20:16
To: Stephan.Engberg@priway.com
cc: public-gld-comments@w3.org
Fax to:
Subject: Re: When are "open" data open?
Hi Stephan,
I have checked the linksmart sourceforge page, and it's not clear what it
offers and how it solves the issues you have raised.
As Richard mentioned, it would be good to understand how these comments
specifically relate to the current GLD work. If you think there are
recommendations that are missing, please be specific on where and what
could be rewritten.
If you check the GLD recommendations that are currently being created, the
standard that is recommended to use is RDF and vocabularies based on this
standard, which I think is quite flexible and able to adapt freely with
the changing "world reality". If you think this is not the case, it would
be interesting to hear your point of view, even if I think it's an issue
wider than the GLD group.
Now, for security of the opened data, it's an interesting issue, and there
is a section about it in the GLD recommendation document. Maybe you want
to make specific comments on this.
Thanks,
Pierre
--
Pierre Andrews, Ph.D.
Research Fellow
On Mon, Mar 25, 2013 at 12:02 PM, <Stephan.Engberg@priway.com> wrote:
Dear Sir,
Creating semantic interoperability represent huge possibilities for
cost-redcution, improving quality and enabling new kinds of previously
unseen solutions.
However, when studying the available work on linked data, 2 vital aspects
not incorporated jumps to my mind - one about innovation or continous
change and one about Empowerment or the assurances that control rests with
the entity at risk and defining the demand (mostly the citizen)
a) The approach assume standardisation around a single univeral definition
b) The approach fail to separate between data that are safe to share and
data that represent a risk to someone.
Ad a) Making strucgtures arund a single univesal standard would make
everything stalemate by legacy.
We need structures that are much more resilient to continous change in
many
directions. And yes this means that we must accept that we cannot FORCE
the
world into a standard bucket unless such as bucket is able to crasp the
world reality.
I sugest a nested approach without any assumptions on outcome. We applied
such an approach in the EU HYDRA project which is partly implemented
http://sourceforge.net/projects/linksmart/
Ad b) Even more important is the need to respect fundamental rights and
society needs.
Buracurats and cynical corporate interests wants to ecxhange data ABOUT
someone as that increase their power and ability to profit. However such a
structure represent a failure by design. EVEN if "anonymised" or
"pseudonymised" such an approach represent a certain failure as it drives
linkage in sources without security.
I kindly refer you to this presentation that are in essence stating the
key
elements.
https://ec.europa.eu/digital-agenda/sites/digital-agenda/files/Stephan.pdf
As can be seen the definition of what can constitute "open" data and how
data must be incapsulated to maintain or eliminate linkage to context is
not a simple question.
We should be extremely carefull NOT to see this from a system-centric or
bureaucrat perspective for WHATEVER excuse, e.g. assuming researchers or
even security administrators CAN access and link data on individuals for
research perspectives.
I kindly suggest to you that failure to incorporate the two above issues
represents a failure to the economy not smaller than that of former
Eastern
European Communism as it leads to legacy-based ineffectiveness and massive
centraslisation of power and control at the expense of citizens and
society.
Sincerely,
Stephan Engberg
Priway - Security in Context
.. because the alternative is not an option
=======================================================
Stephan Engberg | Stephan.Engberg@priway.com
Priway - VAT/SE DK 25 77 53 76
Stengaards Alle 33D - 2800 Kgs. Lyngby - Denmark
Tel.: (+45) 2834 0404 - Internet: www.priway.com
Received on Wednesday, 27 March 2013 17:27:41 UTC