This archive contains answers to questions sent to Unidata support through mid-2025. Note that the archive is no longer being updated. We provide the archive for reference; many of the answers presented here remain technically correct, even if somewhat outdated. For the most up-to-date information on the use of NSF Unidata software and data services, please consult the Software Documentation first.
Roy Mendelssohn <address@hidden> writes: > Thanks Russ! The first link in particular is the type of information > we are looking for. There are now quite a few ways to represent he > same types of data, but it would be nice if there were consistent > ways for doing so, plus some of the ways will be more efficient in > terms of access, compression, ability to use chunking etc. etc. > > Also, based on the python interface that Jeff Whitaker wrote and his > test programs (while the calls are in python the bulk of the dirty > work is done by the C libs), read access to a file using compression > on a large array was roughly 8-10 times slower than the same array > uncompressed, and the savings in size was roughly one order of > magnitude. Does this sound like what you would expect? The > compression feature is very attractive given the size of our files, > but that is a reasonably large speed hit. Yes, this sounds correct. I am working on documenting this a bit more and then asking the HDF5 programmers if this is a bug or expected. Certainly I expected compressed data to read faster, I am not seeing that result. Thanks, Ed -- Ed Hartnett -- address@hidden