I'm trying to parse a huge JSON file (14GB approx.) on Python to do some data mining research I'm working on.
The problem is that when I use the built-in JSON module, it tries to load the full file on memory, until it runs out of it.
Of course, I could find a machine that can stand this file on RAM, in fact I have, but this is not a nice way of doing it.
What I have tried:
import json
with open('myfile.json', 'r'):
loaded_json = json.load(file)
# ...do stuff
What I would like is a way of using this file with the regular JSON interface of lists and dicts, but this way should processes the file directly from disk or by chunks on memory.
Thanks!