I have a huge file ~8 GB in JSON and I want to read it as stream with chunks of 1000 examples at a time. So I searched a lot and tried several packages but not of them really did the job.
The format of my file is as follows:
{
"Elem1": [
{
"orgs": [],
},
{
"people":[]
},
],
"Elem2"":[
{
"orgs": [],
},
{
"people":[]
},
],
...
}
As you can see, each element
is saved as a tuple with two dicts and reoccurring keys in it.
Is there a way how I could read/load/process the file above in chunks of elements i.e.
chunk_1 = [ Elem1, Elem2, ... ]
into the RAM and get the values for the keys out of them?
Any ideas how to do that? Would appreciate your help.
Best regards Chris