I have data which don't fit in memory. So, I was reading in the followings links:
https://stackoverflow.com/a/32582416/9759150
https://stackoverflow.com/a/29518749/9759150
the previous ones related to this: https://spark.apache.org/faq.html
According with the reading Spark writes on disk if data don't fit in memory. But I want to avoid the writing on disk. So I want to know if I can determine how many times do I need to iterate over the data to process it only on memory. Can I do this? How?