I use pandas to work with data. I like this approach because manipulation of data is very easy in pandas (selecting rows, adding row, removing columns, grouping by, joining tables and so on).
My question is if pandas is also a good way to go if the data are huge. In particular I worry about modifying and extracting data. Before I can modify data or extract something from the data I need to read (load) the data from a file and then, after I did what I wanted to do (selecting or modifying) I need to save the data back to the file. I am afraid that this "loading" and "saving" if data may be very slow for huge data. By huge data I understand several hundred millions rows.
In particular, my question is if pandas can be used as a replacement of databases (for example SQLite or MySQL). Alternatively, would it be faster to use a python interface for MySQL to find a particular row in a huge table (saved in a MySQL database) in comparison with finding the same row in a corresponding data frame that is saved as a file.