I have a code in apache spark 1.6.3 running on qubole which writes data to multiple tables(parquet format) on s3. At the time of writing to tables I keep getting java.io.FileNotFound exception
.
I am even setting: spark.sql.parquet.output.committer.class=org.apache.spark.sql.parquet.DirectParquetOutputCommitter.
But this does not seem to solve my problem. Also while checking the logs I see that the exception is due to the _temporary location being missing
. I don't understand why the _temporary location even after using DirectParquetOutputCommitter
. This exception keeps occurring.
Please let me know if anyone know something to solve this in qubole. Thanks.