maybe someone knows an easier way to do this. I am running an EMR cluster (6.x) (1 Master, 1 Slave) with Spark (3.x) on it. Trying to write some data to mysql RDS with a spark job.
spark-submit --jars s3://s3-bucket-jar-assignment/mysql-connector-java-8.0.25.jar s3://s3-bucket-scripts-assignment/scripts/pyspark_script.py
I get this error: I have to mention that I have not installed the jar on the master. How do I do that if I have the jar on an s3 bucket?
SLF4J: Class path contains multiple SLF4J bindings.
SLF4J: Found binding in [jar:file:/usr/lib/spark/jars/slf4j-log4j12-1.7.30.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: Found binding in [jar:file:/usr/share/aws/emr/emrfs/lib/slf4j-log4j12-1.7.12.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: Found binding in [jar:file:/usr/share/aws/redshift/jdbc/redshift-jdbc42-1.2.37.1061.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation.
SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]
File "/mnt/tmp/spark-079b5158-31f7-419b-9877-0e557b9aa612/pyspark_script.py", line 11
.config(conf=SparkConf()).getOrCreate()
^
IndentationError: unexpected indent
21/12/18 20:58:16 INFO ShutdownHookManager: Shutdown hook called
21/12/18 20:58:16 INFO ShutdownHookManager: Deleting directory /mnt/tmp/spark-079b5158-31f7-419b-9877-0e557b9aa612