I would like to save on my disk huge amount (millions) of small numpy matrices (shape=(224, 224, 1), dtype=np.float32).
I would like to have efficient access, and some kind of transparent compression.
One solution which works for me are hadoop files (and h5py library). There is only one drawback: I would like to have also multithreading access. I would like to write to files from few threads. I would like to care about conflicts in writing on my own.
Is there any suggested technology/database I should use?