I have a dataset which is 70 GB. i already posted about how to read huge xml files. i tried iteration method but with this method the file reads fine upto 7-8 hours, after that the IDE closes. i tried with pycharm, anconda, and spider. i increased my RAM from 4 GB to 8 GB. is there another way to read this file full without any issue?
i increased my RAM from 4 to 8GB, and used iteration method, although somehow it works fine with this method but after reading dataset upto 7- 8 hours the IDE get closes and system get hangs.
here is my code i tried
import xml.etree.ElementTree as etree
for event, elem in etree.iterparse('Tags.xml', events=('start', 'end')):
for rows in elem:
count = count + 1
print(rows.attrib)
elem.clear()