Which compression algorithms are typically used when compressing HDF5 file that were created without applying any HDF5 compression filters?
My HDF5 files are created with h5py
in Python 3.8 and contain N-dimensional numpy arrays of 32-bit floats ranging between -1.0 and 1.0 shapes similar to (1000000,10,200)
. Data is being read as (1,10,200)
arrays in a random pattern. Chunking the HDF5 datasets appear to make non-continuous/random reading/selection significantly slower, thus chunking was disabled which prevented the use of HDF5 compression filters.