I am going to work with very large spatio-temporal dataset stored in PostgreSQL server with read-only access. The goal is to load these dataset to Python environment, do all processing and create some kind of learning model.
It wouldn't be a nice idea to have these dataset loaded in python from postgres every time I want work.
I am wondering if there is a way to have Postgres dump these dataset to disk in a format (e.g. .pkl
or .hdf
) that could support faster loading in the future. I really don't think \COPY
to csv
would be an option.