I'm working with a collection of big data sets https://catalogue.data.gov.bc.ca/dataset/vegetation-composite-polygons and am running into issues with st_read taking a long time to load in data. I'm running R Studio 1.2.1335 and R 3.6.1 on a 2018 MacBook Pro with 32 GB of RAM.
The data in the above link only represents 1/8 of the total data I have (I have data for 2008-2018).
Currently, loading in the unformatted data with st_read and .gdb requires most of the day (5-8 hours). After I format and trim the data to include only what I want, loading in the data only takes 5-10 minutes.
I am interested in solutions to avoid the process taking 10 days to load in each layer of the .gdb and formatting the data. Ideally, I'd be able to load in all the layers at once and format all 62,795,746 rows at the same time (taken from st_layers).
I've tried Loading data using .gdb, .shp, and .gpkg.
.gpkg is fast once I get the data formatted.