HDF5 is a unique technology suite that makes possible the management of extremely large and complex data collections.
The HDF5 technology suite includes:
- A versatile data model that can represent very complex data objects and a wide variety of metadata.
- A completely portable file format with no limit on the number or size of data objects in the collection.
- A software library that runs on a range of computational platforms, from laptops to massively parallel systems, and implements a high-level API with C, C++, Fortran 90, and Java interfaces.
- A rich set of integrated performance features that allow for access time and storage space optimizations.
- Tools and applications for managing, manipulating, viewing, and analyzing the data in the collection.
The following version are currently available on the cluster:
... on envirnoment hpc-uniol-env:
HDF/4.2.11-intel-2016b HDF/4.2.12-gimkl-7.2017.3 HDF/4.2.12-goolf-7.2.11 HDF5/1.8.17-intel-2016b
... on environment hpc-env/6.4:
HDF/4.2.14-GCCcore-6.4.0 HDF5/1.10.1-foss-2017b HDF5/1.10.1-intel-2018a HDF5/1.10.2-gimkl-2018a HDF5/1.10.2-intel-2018a HDF5/1.10.5-intel-2018a
... on environment hpc-env/8.3:
HDF5/1.10.5-gompi-2019b HDF5/1.10.5-gompic-2019b HDF5/1.10.5-iimpi-2019b HDF5/1.12.0-iomkl-2019b HDF5/1.12.1-gcccuda-2021.06 HDF5/1.12.1-iomkl-2019b HDF5/1.12.1-iimpi-2019b
Using HDF on the HPC cluster
If you want to use HDF on the cluster, you will need to load the corresponding module first. You can do that with the command:
module load HDF
This will load HDF4 in version 4.2.11. If you need a different version, you will need to specify that like so:
module load HDF5/1.8.17-intel-2016b
The full documentation and further informations can be found here.