Since version 3.3, support for NetCDF4 chunking and (loss-less) compression has been included. These options build on the standard NetCDF output and allow the user control over the size of the chunks via namelist settings. Chunking and compression can lead to significant reductions in file sizes for a small runtime overhead. For a fuller discussion on chunking and other performance issues the reader is referred to the NetCDF4 documentation found here.
The new features are only available when the code has been linked with a NetCDF4 library (version 4.1 onwards, recommended) which has been built with HDF5 support (version 1.8.4 onwards, recommended). Datasets created with chunking and compression are not backwards compatible with NetCDF3 "classic" format but most analysis codes can be relinked simply with the new libraries and will then read both NetCDF3 and NetCDF4 files. NEMO executables linked with NetCDF4 libraries can be made to produce NetCDF3 files by setting the ln_nc4zip logical to false in the namnc4 namelist:
!----------------------------------------------------------------------- &namnc4 ! netcdf4 chunking and compression settings ("key_netcdf4") !----------------------------------------------------------------------- nn_nchunks_i= 4 ! number of chunks in i-dimension nn_nchunks_j= 4 ! number of chunks in j-dimension nn_nchunks_k= 31 ! number of chunks in k-dimension ! setting nn_nchunks_k = jpk will give a chunk size of 1 in the vertical which ! is optimal for postprocessing which works exclusively with horizontal slabs ln_nc4zip = .true. ! (T) use netcdf4 chunking and compression ! (F) ignore chunking information and produce netcdf3-compatible files /
If key_ netcdf4 has not been defined, these namelist parameters are not read. In this case, ln_nc4zip is set false and dummy routines for a few NetCDF4-specific functions are defined. These functions will not be used but need to be included so that compilation is possible with NetCDF3 libraries.
When using NetCDF4 libraries, key_ netcdf4 should be defined even if the intention is to create only NetCDF3-compatible files. This is necessary to avoid duplication between the dummy routines and the actual routines present in the library. Most compilers will fail at compile time when faced with such duplication. Thus when linking with NetCDF4 libraries the user must define key_ netcdf4 and control the type of NetCDF file produced via the namelist parameter.
Chunking and compression is applied only to 4D fields and there is no advantage in chunking across more than one time dimension since previously written chunks would have to be read back and decompressed before being added to. Therefore, user control over chunk sizes is provided only for the three space dimensions. The user sets an approximate number of chunks along each spatial axis. The actual size of the chunks will depend on global domain size for mono-processors or, more likely, the local processor domain size for distributed processing. The derived values are subject to practical minimum values (to avoid wastefully small chunk sizes) and cannot be greater than the domain size in any dimension. The algorithm used is:
ichunksz(1) = MIN( idomain_size,MAX( (idomain_size-1)/nn_nchunks_i + 1 ,16 ) )
ichunksz(2) = MIN( jdomain_size,MAX( (jdomain_size-1)/nn_nchunks_j + 1 ,16 ) )
ichunksz(3) = MIN( kdomain_size,MAX( (kdomain_size-1)/nn_nchunks_k + 1 , 1 ) )
ichunksz(4) = 1
As an example, setting:
nn_nchunks_i=4, nn_nchunks_j=4 and nn_nchunks_k=31
for a standard ORCA2_LIM configuration gives chunksizes of 46x38x1 respectively in the mono-processor case (i.e. global domain of 182x149x31). An illustration of the potential space savings that NetCDF4 chunking and compression provides is given in table 11.3 which compares the results of two short runs of the ORCA2_LIM reference configuration with a 4x2 mpi partitioning. Note the variation in the compression ratio achieved which reflects chiefly the dry to wet volume ratio of each processing region.
|
When key_ iomput is activated with key_ netcdf4 chunking and compression parameters for fields produced via iom_put calls are set via an equivalent and identically named namelist to namnc4 in xmlio_server.def. Typically this namelist serves the mean files whilst the namnc4 in the main namelist file continues to serve the restart files. This duplication is unfortunate but appropriate since, if using io_servers, the domain sizes of the individual files produced by the io_server processes may be different to those produced by the invidual processing regions and different chunking choices may be desired.
Gurvan Madec and the NEMO Team
NEMO European Consortium2017-02-17