My build.sbt
file has this:
scalaVersion := "2.10.3"
libraryDependencies += "com.databricks" % "spark-csv_2.10" % "1.1.0"
I am running Spark in standalone cluster mode and my SparkConf is SparkConf().setMaster("spark://ec2-[ip].compute-1.amazonaws.com:7077").setAppName("Simple Application")
(I am not using the method setJars
, not sure whether I need it).
I package the jar using the command sbt package
. Command I use to run the application is ./bin/spark-submit --master spark://ec2-[ip].compute-1.amazonaws.com:7077 --class "[classname]" target/scala-2.10/[jarname]_2.10-1.0.jar
.
On running this, I get this error:
java.lang.RuntimeException: Failed to load class for data source: com.databricks.spark.csv
What's the issue?