I am new to spark and relatively new to Linux in general. I am running Spark on local Ubuntu in client mode. I have RAM of 16 GB. I installed apache spark following this link. And I am able to run and process large volume of data. The challenge is exporting the resulting data frames in csv. With even 100k rows of data I am getting all sorts of memory issues. On contrast I was able to process partitioned python files of totaling several millions of rows.
Based on lots of googling, I believe the problem lies with my spark.driver.memory. I need to change this but since I am running on client mode I should change it in some configuration file. How can I locate if I have an existing Spark configuration file or how do I create a new one and set spark.driver.memory to 2GB.