- From: M005994 <Jiang.Guoqian@mayo.edu>
- Date: Sat, 17 Sep 2011 14:20:17 -0500
- To: "Hau, Dave (NIH/NCI) [E]" <haudt@mail.nih.gov>, Jim McCusker <james.mccusker@yale.edu>, conor dowling <conor-dowling@caregraf.com>
- CC: "public-semweb-lifesci@w3.org" <public-semweb-lifesci@w3.org>
- Message-ID: <CA9A5D21.4B82%Jiang.Guoqian@mayo.edu>
Hi, Dave, >>>Is there a REST mechanism to expose RDF data? There is a Linked Data API specification available which supports the creation of simple RESTful APIs over RDF triple stores. http://code.google.com/p/linked-data-api/wiki/Specification -Guoqian Guoqian Jiang, Ph.D. =========================================== Assistant Professor of Medical Informatics Division of Biomedical Statistics & Informatics, Mayo Clinic College of Medicine 200 First Street, SW, Rochester, MN, 55905 Tel: 1-507-266-1327 Fax: 1-507-284-0360 Email: jiang.guoqian@mayo.edu =========================================== On 9/17/11 1:32 PM, "Hau, Dave (NIH/NCI) [E]" <haudt@mail.nih.gov> wrote: > Back to "forget the plumbing" and focus on the content, how would you define a > unit of content for a particular purpose (e.g. a blood pressure reading) in > OWL and/or RDF? This would correspond to an archetype or Detailed Clinical > Model (DCM), and would be a subset of the domain ontology(s). > > I also came across the hData effort which seems very promising: > > http://www.projecthdata.org > > They are proposing a REST mechanism for transport (with some basic HTTP based > security as well), and a generic content format (hData Record Format) that's > primarily XML based currently, but potentially could be adapted to carry RDF > payload. (The REST mechanism claims conformance to the OMG RLUS profile, with > a semantic signifier linking the data to the information model. IMHO this > could potentially be adapted to use RDF instead, that's linked to concepts in > an OWL model.) Is there a REST mechanism to expose RDF data? > > - Dave > > > > > From: Jim McCusker [mailto:james.mccusker@yale.edu] > Sent: Monday, August 22, 2011 9:40 AM > To: conor dowling > Cc: Hau, Dave (NIH/NCI) [E]; public-semweb-lifesci@w3.org > Subject: Re: FW: A Fresh Look Proposal (HL7) > > I was just crafting a mail about how our investment in XML technologies hasn't > paid off when this came in. What he said. :-) > > On Mon, Aug 22, 2011 at 9:33 AM, conor dowling <conor-dowling@caregraf.com> > wrote: > >>> >> The content matters, the format does not. > > > > should be front and center. Talk of XML that or JSON this, of RDF as XML in a > chain is a distraction - it's just plumbing. There are many tool-chains and > implementors are big boys - they can graze the buffet themselves. > > > > Central to any patient model rework (I hope) would be the interplay of formal > specifications for terminologies like SNOMED along with any patient data > information model. What should go in the terminology concept (the "object" in > RDF terms) - what is left in the model (the "predicate"). Right now, this > interplay is woefully under specified and implementors throw just about any > old concept into "appropriate" slots in RIM (I know this from doing meaningful > use tests: http://www.caregraf.com/blog/being-allergic-to-allergies, > http://www.caregraf.com/blog/there-once-was-a-strawberry-allergy ) BTW, if > SNOMED is the terminology of choice (for most) then the dance of it and any > RIM-2 should drive much of RIM-2's form. > > > > This is a chance to get away from a fixation on formats/plumbing/"the trucks > for data" and focus on content and in that focus to consider every aspect of > expression, not just the verbs (RIM) or the objects (SNOMED) but both. > > > > Back to "forget the plumbing": if you want to publish a patient's data as an > RDF graph or relational tables or you want a "document" to send on a wire, if > you want to query with a custom protocol or use SPARQL or SQL, you should be > able to and not be seen as an outlier. Each can be reduced to equivalents in > other formats for particular interoperability. The problem right now is that > so much time is spent talking about these containers and working between them > and too little time is given over to what they contain, > > > > Conor > > > > On Mon, Aug 22, 2011 at 6:01 AM, Hau, Dave (NIH/NCI) [E] <haudt@mail.nih.gov> > wrote: > > I see what you're saying and I agree. > > The appeal of XML (i.e. XML used with an XSD representing model syntactics, > not XML used as a serialization as in RDF/XML) is due in part to: > > - XML schema validation API is available on virtually all platforms e.g. Java, > Javascript, Google Web Toolkit, Android etc. > - XML schema validation is relatively lightweight computationally. Pellet ICV > and similar mechanisms are more complete in their validation with the model, > but much more computationally expensive unless you restrict yourself to a > small subset of OWL which then limits the expressiveness of the modeling > language. > - XML provides a convenient bridge from models such as OWL to relational > databases e.g. via JAXB or Castor to Java objects to Hibernate to any RDB. > - Relational querying and XML manipulation skills are much more plentiful in > the market than SPARQL skills currently. > - Some of the current HL7 artifacts are expressed in XSD format, such as their > datatypes (ISO 21090 ; although there are alternative representations such as > UML, and there is an abstract spec too from HL7). If we operate with OWL and > RDF exclusively, would need to convert these datatypes into OWL. > > Maybe it'd be worthwhile to get a few of us who are interested in this topic > together, with some of the HL7 folks interested, and have a few calls to flush > this out and maybe write something up? > > - Dave > > > > > > From: Jim McCusker [mailto:james.mccusker@yale.edu] > Sent: Sunday, August 21, 2011 6:12 PM > To: Hau, Dave (NIH/NCI) [E] > Cc: public-semweb-lifesci@w3.org > Subject: Re: FW: A Fresh Look Proposal (HL7) > > > I feel I need to cut to the chase with this one: XML schema cannot validate > semantic correctness. > > > > It can validate that XML conforms to a particular schema, but that is > syntactic. The OWL validator is nothing like a schema validator, first it > produces a closure of all statements that can be inferred from the asserted > information. This means that if a secondary ontology is used to describe some > data, and that ontology integrates with the ontology that you're attempting to > validate against, you will get a valid result. An XML schema can only work > with what's in front of it. > > > > Two, there are many different representations of information that go beyond > XML, and it should be possible to validate that information without anything > other than a mechanical, universal translation. For instance, there are a few > mappings of RDF into JSON, including JSON-LD, which looks the most promising > at the moment. Since RDF/XML and JSON-LD both parse to the same abstract > graph, there is a mechanical transformation between them. When dealing with > semantic validity, you want to check the graph that is parsed from the > document, not the document itself. > > > > The content matters, the format does not. For instance, let me define a new > RDF format called RDF/CSV: > > > > First column is the subject. First row is the predicate. All other cell values > are objects. URIs that are relative are relative to the document, as in > RDF/XML. > > > > I can write a parser for that in 1 hour and publish it. It's genuinely useful, > and all you would have to do to read and write it is to use my parser or write > one yourself. I can then use the parser, paired with Pellet ICV, and validate > the information in the file without any additional work from anyone. > > > > Maybe we need a simplified XML representation for RDF that looks more like > regular XML. But to make a schema for an OWL ontology is too much work for too > little payoff. > > > > Jim > > On Sun, Aug 21, 2011 at 5:45 PM, Hau, Dave (NIH/NCI) [E] <haudt@mail.nih.gov> > wrote: > > Hi all, > > As some of you may have read, HL7 is rethinking their v3 and doing some > brainstorming on what would be a good replacement for a data exchange paradigm > grounded in robust semantic modeling. > > On the following email exchange, I was wondering, if OWL is used for semantic > modeling, are there good ways to accomplish the following: > > 1. Generate a wire format schema (for a subset of the model, the subset they > call a "resource"), e.g. XSD > > 2. Validate XML instances for conformance to the semantic model. (Here I'm > reminded of Clark and Parsia's work on their Integrity Constraint Validator: > http://clarkparsia.com/pellet/icv ) > > 3. Map an XML instance conformant to an earlier version of the "resource" to > the current version of the "resource" via the OWL semantic model > > I think it'd be great to get a semantic web perspective on this fresh look > effort. > > Cheers, > Dave > > > > Dave Hau > National Cancer Institute > Tel: 301-443-2545 <tel:301-443-2545> > Dave.Hau@nih.gov > > > > > From: owner-its@lists.hl7.org [mailto:owner-its@lists.hl7.org] On Behalf Of > Lloyd McKenzie > Sent: Sunday, August 21, 2011 12:07 PM > To: Andrew McIntyre > Cc: Grahame Grieve; Eliot Muir; Zel, M van der; HL7-MnM; RIMBAA; HL7 ITS > Subject: Re: A Fresh Look Proposal > > Hi Andrew, > > > > Tacking stuff on the end simply doesn't work if you're planning to use XML > Schema for validation. (Putting new stuff in the middle or the beginning has > the same effect - it's an unrecognized element.) The only alternative is to > say that all changes after "version 1" of the specification will be done using > the extension mechanism. That will create tremendous analysis paralysis as we > try to get things "right" for that first version, and will result in > increasing clunkiness in future versions. Furthermore, the extension > mechanism only works for the wire format. For the RIM-based description, we > still need proper modeling, and that can't work with "stick it on the end" no > matter what. > > > > That said, I'm not advocating for the nightmare we currently have with v3 > right now. > > > > I think the problem has three parts - how to manage changes to the wire > format, how to version resource definitions and how to manage changes to the > semantic model. > > > > Wire format: > > If we're using schema for validation, we really can't change anything without > breaking validation. Even making an existing non-repeating element repeat is > going to cause schema validation issues. That leaves us with two options (if > we discount the previously discussed option of "get it right the first time > and be locked there forever": > > 1. Don't use schema > > - Using Schematron or something else could easily allow validation of the > elements that are present, but ignore all "unexpected" elements > > - This would cause significant pain for implementers who like to use schema to > help generate code though > > > > 2. Add some sort of a version indicator on new content that allows a > pre-processor to remove all "new" content if processing using an "old" handler > > - Unpleasant in that it involves a pre-processing step and adds extra "bulk" > to the instances, but other than that, quite workable > > > > I think we're going to have to go with option #2. It's not ideal, but is > still relatively painless for implementers. The biggest thing is that we can > insist on "no breaking x-path changes". We don't move stuff between levels in > a resource wire format definition or rename elements in a resource wire format > definition. In the unlikely event we have to deprecate the entire resource > and create a new version. > > > > Resource versioning: > > At some point, HL7 is going to find at least one resource where we blew it > with the original design and the only way to create a coherent wire format is > to break compatibility with the old one. This will then require definition of > a new resource, with a new name that occupies the same semantic space as the > original. I.e. We'll end up introducing "overlap". Because overlap will > happen, we need to figure out how we're going to deal with it. I actually > think we may want to introduce overlap in some places from the beginning. > Otherwise we're going to force a wire format on implementers of simple > community EMRs that can handle prescriptions for fully-encoded chemo-therapy > protocols. (They can ignore some of the data elements, but they'd still have > to support the full complexity of the nested data structures.) > > > > I don't have a clear answer here, but I think we need to have a serious > discussion about how we'll handle overlap in those cases where it's necessary, > because at some point it'll be necessary. If we don't figure out the approach > before we start, we can't allow for it in the design. > > > > All that said, I agree with the approach of avoiding overlap as much as > humanly possible. For that reason, I don't advocate calling the Person > resource "Person_v1" or something that telegraphs we're going to have new > versions of each resource eventually (let alone frequent changes). > Introduction of a new version of a resource should only be done when the pain > of doing so is outweighed by the pain of trying to fit new content in an old > version, or requiring implementers of the simple to support the structural > complexity of our most complex use-cases. > > > > > > Semantic model versioning: > > This is the space where "getting it right" the first time is the most > challenging. (I think we've done that with fewer than half of the normative > specifications we've published so far.) V3 modeling is hard. The positive > thing about the RFH approach is that very few people need to care. We could > totally refactor every single resource's RIM-based model (or even remove them > entirely), and the bulk of implementers would go on merrily exchanging wire > syntax instances. However, that doesn't mean the RIM-based representations > aren't important. They're the foundation for the meaning of what's being > shared. And if you want to start sharing at a deeper level such as > RIMBAA-based designs, they're critical. This is the level where OWL would > come in. If you have one RIM-based model structure, and then need to refactor > and move to a different RIM-based model structure, you're going to want to map > the semantics between the two structures so that anyone who was using the old > structure can manage instances that come in with the new structure (or vice > versa). OWL can do that. And anyone who's got a complex enough > implementation to parse the wire format and trace the elements through the > their underlying RIM semantic model will likely be capable of managing the OWL > mapping component as well. > > > > > > In short, I think we're in agreement that separation of wire syntax and > semantic model are needed. That will make model refactoring much easier. > However we do have to address how we're going to handle wire-side and resource > refactoring too. > > > > > > Lloyd > > -------------------------------------- > Lloyd McKenzie > > +1-780-993-9501 <tel:%2B1-780-993-9501> > > > > Note: Unless explicitly stated otherwise, the opinions and positions expressed > in this e-mail do not necessarily reflect those of my clients nor those of the > organizations with whom I hold governance positions. > > On Sun, Aug 21, 2011 at 7:53 AM, Andrew McIntyre > <andrew@medical-objects.com.au> wrote: > > Hello Lloyd, > > While "tacking stuff on the end" in V2 may not at first glance seem like an > elegant solution I wonder if it isn't actually the best solution, and one that > has stood the test of time. The parsing rules in V2 do make version updates > quite robust wrt backward and forward inter-operability. > > I am sure it could be done with OWL but I doubt we can switch the world to > using OWL in any reasonable time frame and we probably need a less abstract > representation for commonly used things. In V2 OBX segments, used in a > hierarchy can create an OWL like object-attribute structure for information > that is not modeled by the standard itself. > > I do think the wire format and any overlying model should be distinct entities > so that the model can be evolved and the wire format be changed in a backward > compatible way, at least for close versions. > > I also think that the concept of templates/archetypes to extend the model > should not invalidate the wire format, but be a metadata layer over the wire > format. This is what we have done in Australia with an ISO 13606 Archetypes in > V2 projects. I think we do need a mechanism for people to develop templates to > describe hierarchical data and encode that in the wire format in a way that > does not invalidate its vanilla semantics (ie non templated V2 semantics) when > the template mechanism is unknown or not implemented. > > In a way the V2 specification does hit at underlying objects/Interfaces, and > there is a V2 model, but it is not prescriptive and there is no requirement > for systems to use the same internal model as long as they use the bare bones > V2 model in the same way. Obviously this does not always work as well as we > would like, even in V2, but it does work well enough to use it for quite > complex data when there are good implementation guides. > > If we could separate the wire format from the clinical models then the 2 can > evolve in their own way. We have done several trial implementations of Virtual > Medical Record Models (vMR) which used V3 datatypes and RIM like classes and > could build those models from V2 messages, or in some cases non standard Web > Services, although for specific clinical classes did use ISO 13606 archetypes > to structure the data in V2 messages. > > I think the dream of having direct model serializations as messages is flawed > for all the reasons that have made V3 impossible to implement in the wider > world. While the tack it on the end, lots of optionality rationale might seem > clunky, maybe its the best solution to a difficult problem. If we define tight > SOAP web services for everything we will end up with thousands of slightly > different SOAP calls for every minor variation and I am not sure this is the > path to enlightenment either. > > I am looking a Grahams proposal now, but I do wonder if the start again from > scratch mentality is not part of the problem. Perhaps that is a lesson to be > learned from the V3 process. Maybe the problem is 2 complex to solve from > scratch and like nature we have to evolve and accept there is lots of junk > DNA, but maintaining a working standard at all times is the only way to avoid > extinction. > > I do like the idea of a cohesive model for use in decision support, and that's > what the vMR/GELLO is about, but I doubt there will ever be a one size fits > all model and any model will need to evolve. Disconnecting the model from the > messaging, with all the pain that involves, might create a layered approach > that might allow the HL7 organism to evolve gracefully. I do think part of the > fresh look should be education on what V2 actually offers, and can offer, and > I suspect many people in HL7 have never seriously looked at it in any depth. > > Andrew McIntyre > > > > Saturday, August 20, 2011, 4:37:37 AM, you wrote: > > Hi Grahame, > > Going to throw some things into the mix from our previous discussions because > I don't see them addressed yet. (Though I admit I haven't reread the whole > thing, so if you've addressed and I haven't seen, just point me at the proper > location.) > > One of the challenges that has bogged down much of the v3 work at the > international level (and which causes a great deal of pain at the > project/implementation level) is the issue of refactoring. The pain at the UV > level comes from the fact that we have a real/perceived obligation to meet all > known and conceivable use-cases for a particular domain. For example, the > pharmacy domain model needs to meet the needs of clinics, hospitals, > veterinarians, and chemotherapy protocols and must support the needs of the > U.S., Soviet union and Botswana. To make matters more interesting, > participation from the USSR and Botswana is a tad light. However the fear is > that if all of these needs aren't taken into account, then when someone with > those needs shows up at the door, the model will need to undergo substantive > change, and that will break all of the existing systems. > > The result is a great deal of time spent gathering requirements and > refactoring and re-refactoring the model as part of the design process, > together with a tendency to make most, if not all data elements optional at > the UV level. A corollary is that the UV specs are totally unimplementable in > an interoperable fashion. The evil of optionality that manifested in v2 that > v3 was going to banish turned out to not be an issue of the standard, but > rather of the issues with creating a generic specification that satisfies > global needs and a variety of use-cases. > > The problem at the implementer/project level is that when you take the UV > model and tightly constrain it to fit your exact requirements, you discover 6 > months down the road that one or more of your constraints was wrong and you > need to loosen it, or you have a new requirement that wasn't thought of, and > this too requires refactoring and often results in wire-level > incompatibilities. > > One of the things that needs to be addressed if we're really going to > eliminate one of the major issues with v3 is a way to reduce the fear of > refactoring. Specifically, it should be possible to totally refactor the > model and have implementations and designs work seemlessly across versions. > > I think putting OWL under the covers should allows for this. If we can assert > equivalencies between data elements in old and new models, or even just map > the wire syntaxes of old versions to new versions of the definition models, > then this issue would be significantly addressed: > - Committees wouldn't have to worry about satisfying absolutely every use-case > to get something useful out because they know they can make changes later > without breaking everything. (They wouldn't even necessarily have to meet all > the use-cases of the people in the room! :>) > - Realms and other implementers would be able to have an interoperability path > that allowed old wire formats to interoperate with new wireformats through the > aid of appropriate tooling that could leverage the OWL under the covers. (I > think creating such tooling is *really* important because version management > is a significant issue with v3. And with XML and schemas, the whole "ignore > everything on the end you don't recognize" from v2 isn't a terribly reasonable > way forward. > > I think it's important to figure out exactly how refactoring and version > management will work in this new approach. The currently proposed approach of > "you can add stuff, but you can't change what's there" only scales so far. > > > I think we *will* need to significantly increase the number of Resources (from > 30 odd to a couple of hundred). V3 supports things like invoices, clinical > study design, outbreak tracking and a whole bunch of other healthcare-related > topics that may not be primary-care centric but are still healthcare centric. > That doesn't mean all (or even most) systems will need to deal with them, but > the systems that care will definitely need them. The good news is that most > of these more esoteric areas have responsible committees that can manage the > definition of these resources, and as you mention, we can leverage the RMIMs > and DMIMs we already have in defining these structures. > > > The specification talks about robust capturing of requirements and > traceability to them, but gives no insight into how this will occur. It's > something we've done a lousy job of with v3, but part of the reason for that > is it's not exactly an easy thing to do. The solution needs to flesh out > exactly how this will happen. > > > We need a mapping that explains exactly what's changed in the datatypes (and > for stuff that's been removed, how to handle that use-case). > > There could still be a challenge around granularity of text. As I understand > it, you can have a text representation for an attribute, or for any XML > element. However, what happens if you have a text blob in your interface that > covers 3 of 7 attributes inside a given XML element. You can't use the text > property of the element, because the text only covers 3 of 7. You can't use > the text property of one of the attributes because it covers 3 separate > attributes. You could put the same text in each of the 3 attributes, but > that's somewhat redundant and is going to result in rendering issues. One > solution might be to allow the text specified at the element level to identify > which of the attributes the text covers. A rendering system could then use > that text for those attributes, and then render the discrete values of the > remaining specified attributes. What this would mean is that an attribute > might be marked as "text" but not have text content directly if the parent > element had a text blob that covered that attribute. > > > > New (to Grahame) comments: > > I didn't see anything in the HTML section or the transaction section on how > collisions are managed for updates. A simple requirement (possibly optional) > to include the version id of the resource being updated or deleted should > work. > > To my knowledge, v3 (and possibly v2) has never supported true "deletes". At > best, we do an update and change the status to nullified. Is that the > intention of the "Delete" transaction, or do we really mean a true "Delete"? > Do we have any use-cases for true deletes? > > I wasn't totally clear on the context for uniqueness of ids. Is it within a > given resource or within a given base URL? What is the mechanism for > referencing resources from other base URLs? (We're likely to have networks of > systems that play together.) > > Nitpick: I think "id" might better be named "resourceId" to avoid any possible > confusion with "identifier". I recognize that from a coding perspective, > shorter is better. However, I think that's outweightd by the importance of > avoiding confusion. > > In the resource definitions, you repeated definitions for resources inherited > from parent resources. E.g. Person.created inherited from > Resource.Base.created. Why? That's a lot of extra maintenance and potential > for inconsistency. It also adds unnecessary volume. > > Suggest adding a caveat to the draft that the definitions are placeholders and > will need significant work. (Many are tautological and none meet the Vocab > WG's guidelines for quality definitions.) > > Why is Person.identifier mandatory? > > You've copied "an element from Resource.Base.???" to all of the Person > attributes, including those that don't come from Resource.Base. > > Obviously the workflow piece and the conformance rules that go along with it > need some fleshing out. (Looks like this may be as much fun in v4 as it has > been in v3 :>) > > The list of identifier types makes me queasy. It looks like we're > reintroducing the mess that was in v2. Why? Trying to maintain an ontology > of identifier types is a lost cause. There will be a wide range of > granularity requirements and at fine granularity, there will be 10s of > thousands. The starter list is pretty incoherent. If you're going to have > types at all, the vocabulary should be constrained to a set of codes based on > the context in which the real-world identifier is present. If there's no > vocabulary defined for the property in that context, then you can use text for > a label and that's it. > > I didn't see anything on conformance around datatypes. Are we going to have > datatype flavors? How is conformance stated for datatype properties? > > I didn't see templateId or flavorId or any equivalent. How do instances (or > portions there-of) declare conformance to "additional" constraint > specifications/conformance profiles than the base one for that particular > server? > > We need to beef up the RIM mapping portion considerably. Mapping to a single > RIM class or attribute isn't sufficient. Most of the time, we're going to > need to map to a full context model that talks about the classCodes, moodCodes > and relationships. Also, you need to relate attributes to the context of the > RIM location of your parent. > > There's no talk about context conduction, which from an implementation > perspective is a good thing. However, I think it's still needed behind the > scenes. Presumably this would be covered as part of the RIM semantics layer? > > In terms of the "validate" transaction, we do a pseudo-validate in pharmacy, > but a 200 response isn't sufficient. We can submit a draft prescription and > say "is this ok?". The response might be as simple as "yes" (i.e. a 200). > However, it could also be a "no" or "maybe" with a list of possible > contraindications, dosage issues, allergy alerts and other detected issues. > How would such a use-case be met in this paradigm? > > At the risk of over-complicating things, it might be useful to think about > data properties as being identifying or not to aid in exposing resources in a > de-identified way. (Not critical, just wanted to plant the seed in your head > about if or how this might be done.) > > > All questions and comments aside, I definitely in favour of fleshing out this > approach and looking seriously at moving to it. To that end, I think we need > a few things: > - A list of the open issues that need to be resolved in the new approach. > (You have "todo"s scattered throughout. A consolidated list of the "big" > things would be useful.) > - An analysis of how we move from existing v3 to the new approach, both in > terms of leveraging existing artifacts and providing a migration path for > existing solutions as well as what tools, etc. we need. > - A plan for how to engage the broader community for review. (Should ideally > do this earlier rather than later.) > > Thanks to you, Rene and others for all the work you've done. > > > Lloyd > > -------------------------------------- > Lloyd McKenzie > > +1-780-993-9501 <tel:%2B1-780-993-9501> > > > > Note: Unless explicitly stated otherwise, the opinions and positions expressed > in this e-mail do not necessarily reflect those of my clients nor those of the > organizations with whom I hold governance positions. > > > On Fri, Aug 19, 2011 at 9:08 AM, Grahame Grieve <grahame@kestral.com.au >> > wrote: > hi All > > Responses to comments > > #Michael > >> > 1. I would expect more functional interface to use these resources. > > as you noted in later, this is there, but I definitely needed to make > more of it. That's where I ran out of steam > >> > 2. One of the things that was mentioned (e.g. at the Orlando >> > WGM RIMBAA Fresh Look discussion) is that we want to use >> > industry standard tooling, right? Are there enough libraries that >> > implement REST? > > this doesn't need tooling. There's schemas if you want to bind to them > >> > 2b. A lot of vendors now implement WebServices. I think we should >> > go for something vendors already have or will easilly adopt. Is that the >> case with REST? > > Speaking as a vendor/programmer/writer of an open source web services > toolkit, I prefer REST. Way prefer REST > >> > Keep up the good work! > > ta > > #Mark > >> > I very much like the direction of this discussion towards web services >> > and in particular RESTful web services. > > yes, though note that REST is a place to start, not a place to finish. > >> > At MITRE we have been advocating this approach for some time with our hData >> initiative. > > yes. you'll note my to do: how does this relate to hData, which is a > higher level > specification than the CRUD stuff here. > > #Eliot > >> > Hats off - I think it's an excellent piece of work and definitely a step in >> right direction. > > thanks. > >> > I didn't know other people in the HL7 world other than me were talking >> about >> > (highrise). Who are they? > > not in Hl7. you were one. it came up in some other purely IT places that I > play > >> > 5) Build it up by hand with a wiki - it is more scalable really since you > > wiki's have their problems, though I'm not against them. > >> > 1) I think it would be better not to use inheritance to define a patient as >> > a sub type of a person. The trouble with that approach is that people can > > On the wire, a patient is not a sub type of person. The relationship > between the two is defined in the definitions. > >> > A simpler approach is associate additional data with a person if and when >> > they become a patient. > > in one way, this is exactly what RFH does. On the other hand, it creates a > new identity for the notion of patient (for integrity). We can discuss > whether that's good or bad. > >> > 2) I'd avoid language that speaks down to 'implementers'. It's enterprise > > really? Because I'm one. down the bottom of your enterprise pole. And > I'm happy to be one of those stinking implementers down in the mud. > I wrote it first for me. But obviously we wouldn't want to cause offense. > I'm sure I haven't caused any of that this week ;-) > >> > 3) If you want to reach a broader audience, then simplify the language. > > argh, and I thought I had. how can we not use the right terms? But I > agree that the introduction is not yet direct enough - and that's after > 4 rewrites to try and make it so.... > > Grahame > > > ************************************************ > To access the Archives of this or other lists or change your list settings and > information, go to: > http://www.hl7.org/listservice > > ************************************************ > To access the Archives of this or other lists or change your list settings and > information, go to: http://www.hl7.org/listservice > > >
Received on Sunday, 18 September 2011 22:09:47 UTC