What is the most efficient way of incrementally building a numpy array, e.g. one row at a time, without knowing the final size in advance?
My use case is as follows. I need to load a large file (10-100M lines) for which each line requires string processing and should form a row of a numpy array.
Is it better to load the data to a temporary Python list and convert to an array or is there some existing mechanism in numpy that would make it more efficient?