[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]
[Support #CUV-251255]: Nccopy extremly slow / hangs
- Subject: [Support #CUV-251255]: Nccopy extremly slow / hangs
- Date: Thu, 14 Jul 2011 09:06:18 -0600
Hi Mark,
At your convenience, build the current snapshot of netCDF from the
source at:
ftp://ftp.unidata.ucar.edu/pub/netcdf/snapshot/netcdf-4-daily.tar.gz
Then see if the new nccopy does the job for you. I've done some
timings, which appear below. As input, I used a netCDF-3 classic file
of about 17.5 GB that contains the 1698x1617x1596 variable last (which
is required in classic format for variables to exceed 4 GB).
Using nccopy to convert the file to a netCDF-4 contiguous (unchunked)
file took about 15 minutes on my desktop Linux system:
$ /usr/bin/time ./nccopy -k4 cvx3.nc cvx4.nc; ls -l cvx[34].nc
13.40user 52.60system 14:54.69elapsed 7%CPU (0avgtext+0avgdata
11120maxresident)k
34284824inputs+68470264outputs (60major+6283minor)pagefaults 0swaps
-rw-rw-r-- 1 russ ustaff 17528362416 Jul 13 13:43 cvx3.nc
-rw-rw-r-- 1 russ ustaff 17528380036 Jul 13 14:10 cvx4.nc
To copy and compress at level 1, using the default 100x96x94 chunks took about
14 minutes and shows the resulting file compressed to about 6.1 GB:
$ /usr/bin/time ./nccopy -k4 -d1 cvx3.nc cvx4-d1.nc; ls -l cvx4-d1.nc
506.78user 20.34system 14:05.81elapsed 62%CPU (0avgtext+0avgdata
1359236maxresident)k
34252976inputs+11967392outputs (74major+1043484minor)pagefaults 0swaps
-rw-rw-r-- 1 russ ustaff 6126519401 Jul 13 14:29 cvx4-d1.nc
To instead do what you need, reshaping to chunks oriented along the
time dimension of shape 1698x25x24, took about the same amount of time
by specifying a chunk cache of 18 GB and lots of items in the chunk
cache. You can also see that with these chunks, the data did not
compress quite as well, but your mileage may vary.
$ /usr/bin/time ./nccopy -k4 -d1 -h 18G -e10001 -m 1G -c
time/1698,latitude/25,longitude/24 cvx3.nc cvx4-time.nc; ls -l cvx4-time.nc
567.68user 44.88system 14:28.54elapsed 70%CPU (0avgtext+0avgdata
18348304maxresident)k
32663816inputs+12791808outputs (57major+10639640minor)pagefaults 0swaps
-rw-rw-r-- 1 russ ustaff 6543504177 Jul 14 06:40 cvx4-time.nc
If I instead left off the new -h and -e arguments to nccopy and just
used the nccopy defaults of a 4M cache and 1009 cache elements, the
above nccopy ran for more than 12 hours and still hadn't even
completed half of the output.
An alternative for reshaping chunks in large data files is to use the
HDF5 tool h5repack. However, I can't recommend that currently, as an
nccopy test similar to your case but with a smaller number of times
rechunked a 3GB file in 2.5 minutes but took 6.25 hours using
h5repack. There are no options for controlling the chunk cache with
h5repack, which may explain why it took 150x longer than nccopy in
this case.
--Russ
Russ Rew UCAR Unidata Program
address@hidden http://www.unidata.ucar.edu
Ticket Details
===================
Ticket ID: CUV-251255
Department: Support netCDF
Priority: Normal
Status: Closed