[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]
[THREDDS #KXL-349288]: Problem subsetting netcdf files
- Subject: [THREDDS #KXL-349288]: Problem subsetting netcdf files
- Date: Sun, 10 Apr 2016 02:36:36 -0600
Hi Rostislav,
Sorry for the delayed response. This is indeed a problem in the hdf-java code.
Specifically, it fails when a variable has a very large chunk size. Thanks for
the report!
At this time, the only way to work around the problem is to rechunk the file to
use smaller chunks, which is something you might want to do anyway. Currently,
you have:
cnc_POLLEN_BIRCH_m22:_ChunkSizes = 97, 8, 315, 467 ;
Or, in other words, the chunk is as large as the entire variable. Such a large
chunk size blunts the potential performance gains that chunking was meant to
provide.
Furthermore, HDF clients must read the entire chunk into memory before they can
access even a small piece of it. That's actually what's causing netCDF-Java to
fail when chunks are very large.
By rechunking your dataset to use smaller chunks, I was able to read it
successfully:
nccopy -k nc4 -d 1 -c time/13,y/40,x/59,height/1
SILAM-POLLEN-europe20160402.nc4 SILAM-POLLEN-europe20160402_rechunked.nc4
Again, there is still a definite problem in netCDF-Java that needs to be
addressed, but rechunking may be something you want to consider anyway. [1] and
[2] are
good primers on the matter.
Cheers,
Christian
[1]
http://www.unidata.ucar.edu/blogs/developer/entry/chunking_data_why_it_matters
[2]
http://www.unidata.ucar.edu/blogs/developer/en/entry/chunking_data_choosing_shapes
Ticket Details
===================
Ticket ID: KXL-349288
Department: Support THREDDS
Priority: High
Status: Closed