Re: [galeon] [WCS-2.0.swg] CF-netCDF standards initiatives

NOTE: The galeon mailing list is no longer active. The list archives are made available for historical reasons.

How so?

To write a parser for a binary file one needs to know the detailed structure of 
the file - and if that changes your code is now broken.  With GML or any 
XML-based encoding the file structure is not known to your application.  Where 
do you see the problems in parsing and understanding GML?

R

-----Original Message-----
From: Gerry Creager [mailto:gerry.creager@xxxxxxxx] 
Sent: August 21, 2009 1:17 PM
To: Ron Lake
Cc: John Caron; Unidata GALEON; wcs-2.0.swg
Subject: Re: [galeon] [WCS-2.0.swg] CF-netCDF standards initiatives

I'm not so sure that's true.

Ron Lake wrote:
> Hi John:
> 
> Surely the GML encoding is going to be simpler to parse and "understand" than 
> any equivalent binary encoding.
> 
> R
> 
> -----Original Message-----
> From: galeon-bounces@xxxxxxxxxxxxxxxx 
> [mailto:galeon-bounces@xxxxxxxxxxxxxxxx] On Behalf Of John Caron
> Sent: August 21, 2009 12:52 PM
> Cc: Unidata GALEON; wcs-2.0.swg
> Subject: Re: [galeon] [WCS-2.0.swg] CF-netCDF standards initiatives
> 
> A few thoughts on these issues:
> 
> The WCS/SWE/OGC are making major contributions in elaborating conceptual 
> models and defining remote access protocols based on them. More 
> problematic is managing complexity and adoption paths of existing 
> communities. One of the main weaknesses of WCS is the likely 
> proliferation of binary encoding formats. WMS has been lucky to be able 
> to use existing, universal formats like JPEG, and partly because of 
> that, has been more widely adopted.
> 
> A central issue is the relation of the WCS data encoding format to its 
> GML envelope. Should the data be "just bits" and let the GML encode all 
> the semantics, or should a binary format also encode some/all of the 
> semantics? In the first case, any client must understand GML, which is a 
> big barrier to use by existing applications, plus the possibility exists 
> that some essential information is offline or unavailable, as Tom W 
> points out. In the second case, we have a proliferation of data encoding 
> formats and the possibility that the semantics are encoded differently 
> between the GML and the data.
> 
>  From Unidata's POV, namely that of supporting a large community of 
> users of netcdf and other scientific data formats, using netCDF-CF to 
> encode the data is obvious. However, this is not just a matter of one 
> more group advocating for their favorite bespoke format. IMHO, netCDF is 
> close to being a "maximally simple implementation" of a binary data 
> encoding that allows data to be more than "just numbers". Its BNF 
> grammer fits on one side of a page of paper. If one chooses to encode 
> semantics into the data format, and not require a client to know GML, 
> then you need something like netCDF, and you wont find an encoding 
> significantly simpler. The CF Conventions group has been slowly adding 
> the necessary semantics for 10 years or so. From that motivation we 
> decided to recommend netCDF as a WCS encoding format.
> 
> The hard part is describing in a formal way the relationship of the 
> netCDF-CF binary response to the actual request and to any GML part of 
> the response. Stefano et al are doing the heroic work of describing the 
> mapping between the two models, but there is and perhaps should be a 
> loose coupling between the two. There are always edge cases: what 
> happens when the requested lat/lon bounding box crosses a discontinuity 
> of the projection plane? One might get different, but reasonable, files 
> back from different servers of the same dataset. So what can we 
> standardize? None of this is unique to netcdf-CF or binary encodings.
> 
> Taking the first path of letting GML carry all the semantics has many 
> advantages also, especially for new development and GIS-centric clients, 
> and as new libraries are developed that can mitigate some of the 
> complexity of GML. In (our) scientific community, these are new 
> possibilities, to be experimented with and prototypes tested. Once there 
> are real services delivering "gotta have" data we will do whatever we 
> need to do to "get me some of that" ;^)
> 
> John Caron
> Unidata
> 
> Steve Hankin wrote:
>>
>> Robin, Alexandre wrote:
>>> Hi Steve,
>>>
>>> Just to clarify when I said NetCDF was a "NEW standard" I meant a new 
>>> standard in OGC.
>>>
>>> As I was telling Ben in an offline email, I am totally conscious of 
>>> its penetration and usefulness in certain communities.
>>>
>>> However, _I am not convinced that /having two standards doing the 
>>> same thing in OGC /is sending the right message and is the best way 
>>> to go for a standardization organization_.
>>>
>> Hi Robin,
>>
>> I confess that I was aware of using a cheap rhetorical device when I 
>> twisted your intended meaning of "NEW". (Begging your tolerance.) It 
>> was helpful in order to raise more fundamental questions. You have 
>> alluded to a key question just above. Is it really best to think of 
>> the target of OGC as a the development of a single, definitive 
>> standard? one that is more general and more powerful than all existing 
>> standards? Or is it better to think of OGC as a process, through which 
>> the forces of divergence in geospatial IT systems can be weakened 
>> leading towards convergence over time? The notion that there can be a 
>> single OGC solution is already patently an illusion. Which one would 
>> you pick? WFS? WCS? SOS with SWE Common? SOS with its many other XML 
>> schema? (Lets not even look into the profusion of WFS application 
>> schema.) I trust that we are not pinning our hopes on a future 
>> consolidation of all of these. There is little evidence to indicate 
>> that we can sustain the focus necessary to traverse that path. The 
>> underlying technology is not standing still.
>>
>> What Ben (and David Arctur and others) have proposed through seeking 
>> to put an OGC stamp of approval on netCDF-CF technology is similar to 
>> what OGC has achieved through putting its stamp on KML ("There are 
>> sound business and policy reasons for doing so.") It is to create a 
>> process -- a technical conversation if you will -- which will lead to 
>> interoperability pathways that bridge technologies and communities. 
>> Real-world interoperability.
>>> There has been a lot of experimentation with SWE technologies as well 
>>> that you may not know about and in many communities, especially in 
>>> earth science.
>>>
>>> What I'm saying is that perhaps it is worth testing bridging NetCDF 
>>> to SWE before we go the way of stamping two 100% overlapping 
>>> standards as OGC compliant.
>>>
>> Complete agreement that this sort of testing ought to occur. And 
>> interest to hear more about what has been achieved. But great 
>> skepticism that there is this degree of overlap between the 
>> approaches. And disagreement that this testing ought to be a 
>> precondition to OGC recognition of a significant ,community-proven 
>> interoperability mechanism like netCDF. OGC standardization of netCDF 
>> will provide a forum for testing and experimentation to occur much 
>> more rapidly and for a 2-way transfer of the best ideas between 
>> approaches. NetCDF & co. (its API, data model, CF, DAP) have a great 
>> deal to offer to OGC.
>>
>> - Steve
>>> Regards,
>>>
>>> -------------------------------------------------
>>>
>>> **Alexandre Robin**
>>>
>>> Spot Image, Web and E-Business
>>>
>>> Tel: +33 (0)5 62 19 43 62
>>>
>>> Fax: +33 (0)5 62 19 43 43
>>>
>>> http://www.spotimage.com
>>>
>>> Before printing, think about the environment
>>>
>>> ------------------------------------------------------------------------
>>>
>>> *De :* Steve Hankin [mailto:Steven.C.Hankin@xxxxxxxx]
>>> *Envoyé :* jeudi 20 août 2009 20:58
>>> *À :* Tom Whittaker
>>> *Cc :* Robin, Alexandre; Ben Domenico; Unidata GALEON; wcs-2.0.swg
>>> *Objet :* Re: [galeon] [WCS-2.0.swg] CF-netCDF standards initiatives
>>>
>>> Hi Tom,
>>>
>>> I am grateful to you for opening the door to comments "from 10 
>>> thousand feet" -- fundamental truths that we know from many years of 
>>> experience, but that we fear may be getting short shrift in 
>>> discussions of a new technology. I'd like to offer a comment of that 
>>> sort regarding the interplay of ideas today between Robin ("/I hope 
>>> we don't have to define a NEW standard .../") and Carl Reed ("/there 
>>> are other organizations interested in bringing legacy spatial 
>>> encodings into the OGC. There are sound business and policy reasons 
>>> for doing so./").
>>>
>>> The NEW standard in this discussion is arguably SWE, rather than 
>>> netCDF. NetCDF has decades of practice behind it; huge bodies of data 
>>> based upon it; a wide range of applications capable of accessing it 
>>> (both locally and remotely); and communities that depend vitally upon 
>>> it. As Ben points out, netCDF also has its own de jure pedigree.
>>>
>>> A key peril shared by most IT standards committees -- a lesson that 
>>> has been learned, forgotten, relearned and forgotten again so many 
>>> times that it is clearly an issue of basic human behavior -- is that 
>>> they will try to innovate. Too-common committee behavior is to 
>>> propose, discuss and document new and intriguing technologies, and 
>>> then advance those documents through a de jure standards process, 
>>> despite an insufficient level of testing. The OGC testbed process 
>>> exists to address this, but we see continually how large the gap is 
>>> between the testbed process and the pace and complexity of 
>>> innovations emerging from committees.
>>>
>>> Excellent reading on this subject is the essay by Michi Henning, /The 
>>> Rise and Fall of CORBA/ (2006 -- 
>>> http://queue.acm.org/detail.cfm?id=1142044). Among the many insights 
>>> he offers is
>>>
>>> **'Standards consortia need iron-clad rules to ensure that they 
>>> standardize existing best practice.** There is no room for innovation 
>>> in standards_._ Throwing in "just that extra little feature" 
>>> inevitably causes unforeseen technical problems, despite the best 
>>> intentions.'
>>>
>>> While it adds weight to an argument to be able to quote from an 
>>> in-print source, this is a self-evident truth. We need only reflect 
>>> on the recent history of IT. What we need is to work together to find 
>>> ways to prevent ourselves from continually forgetting it.
>>>
>>> There is little question in my mind that putting an OGC stamp of 
>>> approval on netCDF is a win-win process -- for the met/ocean/climate 
>>> community and for the broader geospatial community. It will be a path 
>>> to greater interoperability in the long run and it deserves to go 
>>> forward. The merits of SWE (or GML) as an alternative approach to the 
>>> same functionality also deserve to be explored and tested in 
>>> situations of realistic complexity. But this exploration should be 
>>> understood initially as a process of R&D -- a required step before a 
>>> "standards process" is considered. If that exploration has already 
>>> been done it should be widely disseminated, discussed and evaluated.
>>>
>>> - Steve
>>>
>>> ==================================
>>>
>>> Tom Whittaker wrote:
>>>
>>> I may be ignorant about these issues, so please forgive me if I am
>>> completely out-of-line....but when I looked at the examples, I got
>>> very concerned since the metadata needed to interpret the data values
>>> in the "data files" is apparently not actually in the file, but
>>> somewhere else.  We've been here before:  One of the single biggest
>>> mistakes that the meteorological community made in defining a
>>> distribution format for realtime, streaming data was BUFR -- because
>>> the "tables" needed to interpret the contents of the files are
>>> somewhere else....and sometimes, end users cannot find them!
>>>  
>>> NetCDF and ncML maintain the essential metadata within the files:
>>> types, units, coordinates -- and I strongly urge you (or whomever) not
>>> to make the  "BUFR mistake" again -- put the metadata into the files!
>>> Do not require the end user to have to have an internet connection to
>>> simply "read" the data....many people download the files and then
>>> "take them along" when traveling, for example.
>>>  
>>> If I simply downloaded the file at
>>> <http://schemas.opengis.net/om/1.0.0/examples/weatherObservation.xml>
>>> I would not be able to read it.  In fact, it looks like even if I also
>>> got the "metadata" file at:
>>> <http://schemas.opengis.net/om/1.0.0/examples/weatherRecord1.xml>
>>> I would still not be able to read it, since it also refers to other
>>> servers in the universe to obtain essential metadata.
>>>  
>>> That is my 2 cents worth....and I hope I am wrong about what I saw in
>>> the examples....
>>>  
>>> tom
>>>  
>>>   
>> ------------------------------------------------------------------------
>>
>> _______________________________________________
>> galeon mailing list
>> galeon@xxxxxxxxxxxxxxxx
>> For list information, to unsubscribe, visit: 
>> http://www.unidata.ucar.edu/mailing_lists/ 
> 
> _______________________________________________
> galeon mailing list
> galeon@xxxxxxxxxxxxxxxx
> For list information, to unsubscribe, visit: 
> http://www.unidata.ucar.edu/mailing_lists/ 
> 
> _______________________________________________
> galeon mailing list
> galeon@xxxxxxxxxxxxxxxx
> For list information, to unsubscribe, visit: 
> http://www.unidata.ucar.edu/mailing_lists/ 

-- 
Gerry Creager -- gerry.creager@xxxxxxxx
Texas Mesonet -- AATLT, Texas A&M University
Cell: 979.229.5301 Office: 979.458.4020 FAX: 979.862.3983
Office: 1700 Research Parkway Ste 160, TAMU, College Station, TX 77843



  • 2009 messages navigation, sorted by:
    1. Thread
    2. Subject
    3. Author
    4. Date
    5. ↑ Table Of Contents
  • Search the galeon archives: