This archive contains answers to questions sent to Unidata support through mid-2025. Note that the archive is no longer being updated. We provide the archive for reference; many of the answers presented here remain technically correct, even if somewhat outdated. For the most up-to-date information on the use of NSF Unidata software and data services, please consult the Software Documentation first.
Hi Rostislav, Sorry for the delayed response. This is indeed a problem in the hdf-java code. Specifically, it fails when a variable has a very large chunk size. Thanks for the report! At this time, the only way to work around the problem is to rechunk the file to use smaller chunks, which is something you might want to do anyway. Currently, you have: cnc_POLLEN_BIRCH_m22:_ChunkSizes = 97, 8, 315, 467 ; Or, in other words, the chunk is as large as the entire variable. Such a large chunk size blunts the potential performance gains that chunking was meant to provide. Furthermore, HDF clients must read the entire chunk into memory before they can access even a small piece of it. That's actually what's causing netCDF-Java to fail when chunks are very large. By rechunking your dataset to use smaller chunks, I was able to read it successfully: nccopy -k nc4 -d 1 -c time/13,y/40,x/59,height/1 SILAM-POLLEN-europe20160402.nc4 SILAM-POLLEN-europe20160402_rechunked.nc4 Again, there is still a definite problem in netCDF-Java that needs to be addressed, but rechunking may be something you want to consider anyway. [1] and [2] are good primers on the matter. Cheers, Christian [1] http://www.unidata.ucar.edu/blogs/developer/entry/chunking_data_why_it_matters [2] http://www.unidata.ucar.edu/blogs/developer/en/entry/chunking_data_choosing_shapes Ticket Details =================== Ticket ID: KXL-349288 Department: Support THREDDS Priority: High Status: Closed