I have an aws glue python job which joins two Aurora tables and writes/sinks output to s3 bucket as json format. The job is working fine as expected. By default the output file is written to s3 bucket in this name format/pattern "run-123456789-part-r-00000" [Behind the scene its running pyspark code in a hadoop cluster, so the file name is hadoop-like]
Now, my question is how to write the file with a specific name like "Customer_Transaction.json" instead of "run-***-part****"
I tried converting to DataFrame and then write as json, like below, but did not work
customerDF.repartition(1).write.mode("overwrite").json("s3://bucket/aws-glue/Customer_Transaction.json")