I am reading a large csv file 25GB into pandas.DataFrame. My pc specifications are:
- Intel core i7-8700 3.2GHz
- RAM 16G
- windows 10
- DataFrame.shape =144,000,000 rows by 13 cols
- csv file size on disk says 24GB
reading this file takes a long time like 20 minutes sometimes. Is there any recommendation, code wise, that I can do better?
*note: This DF is needed in whole, since I am going to Join(Merge) with another one.