I have a huge (>1GB text file) that I want to parse and convert into smaller files.
My text file looks like this:
Iteration column1 column2 .... column 10k
1 data_1_1 data_1_2 data_1_10k
2 data_2_1 data_2_2 data_2_10k
...
10k data_10k_1 data_10k_2 data_10k_10k
I want to parse this text file and convert it into 10k CSV files where each CSV file will contain the following data:
Iteration, column
1, data_1
2, data_2
...,
10k, data_10k
I'm looking for the fastest method to do this in python. Would it be possible to parallelize this into 10k chunks?