I am reading table data from sql server and storing it as a Dataframe in spark i want to write back the df to a parquet file in s3 as the table has around 30 columns and 204 million rows, right now with spark.write its taking around 25-30 minutes to write back to s3 as a parquet file is there any fastest way to do the same ?
Edit-1 Do suggest me if you have any better approach to the same as well instead of direct spark.write