I have a ".csv.gz" file that is 100GB large in a remote linux. I definitely do not want to unzip it because the size would reach to 1T.
I am looking online for reading files. I saw on suggestion here
python: read lines from compressed text files
gzip? pandas? iterator?
My mentor suggested to pip the data after unzip it.
I would also need to consider the memory. So readlines() is definitely not my consideration.
I wonder if anyone has an optimal solution for this because the file is really large and it would take me a lot of time to just do anything.