I'm running Hive 2.1.1, Spark 2.1.0 and Hadoop 2.7.3.
I tried to build Spark following the Hive on Spark: Getting Started:
./dev/make-distribution.sh --name "hadoop2-without-hive" --tgz "-Pyarn,hadoop-provided,hadoop-2.7,parquet-provided"
However, I couldn't find any spark-assembly jar files under the spark directory (find . -name "spark-assembly*.jar"
returns nothing back). Instead of linking the spark-assembly jar to HIVE_HOME/lib
, I tried export SPARK_HOME=/home/user/spark
.
I get the following Hive error in beeline:
0: jdbc:hive2://localhost:10000> set hive.execution.engine=spark;
0: jdbc:hive2://localhost:10000> insert into test (id, name) values (1, 'test1');
Error: Error running query: java.lang.NoClassDefFoundError: scala/collection/Iterable (state=,code=0)
I think the error is caused by missing spark-assembly jars.
How could I build / Where could I find those spark-assembly jar files?
How could I fix the above error?
Thank you!