1

, I have installed
1) scala-2.10.3

2) spark-1.0.0 Changed spark-env.sh with below variables

export SCALA_HOME=$HOME/scala-2.10.3

export SPARK_WORKER_MEMORY=16g

I can see Spark master.

3) shark-0.9.1-bin-hadoop1

Changed shark-env.sh with below variables

export SHARK_MASTER_MEM=1g

SPARK_JAVA_OPTS=" -Dspark.local.dir=/tmp "

SPARK_JAVA_OPTS+="-Dspark.kryoserializer.buffer.mb=10 "

SPARK_JAVA_OPTS+="-verbose:gc -XX:-PrintGCDetails -XX:+PrintGCTimeStamps "

    export SPARK_JAVA_OPTS

    export HIVE_HOME=/usr/share/dse/hive

export HIVE_CONF_DIR="/etc/dse/hive"

export SPARK_HOME=/home/ubuntu/spark-1.0.0

export SPARK_MEM=16g

source $SPARK_HOME/conf/spark-env.sh

4) In DSE, Hive version is Hive 0.11 Existing Hive-site.xml is

    <configuration>

     <!-- Hive Execution Parameters -->
         <property>
                <name>hive.exec.mode.local.auto</name>
                <value>false</value>
             <description>Let hive determine whether to run in local mode automatically</description>
        </property>

        <property>
                <name>hive.metastore.warehouse.dir</name>
              <value>cfs:///user/hive/warehouse</value>
             <description>location of default database for the warehouse</description>
        </property>

        <property>
             <name>hive.hwi.war.file</name>
              <value>lib/hive-hwi.war</value>
                <description>This sets the path to the HWI war file, relative to ${HIVE_HOME}</description>
          </property>

          <property>
                <name>hive.metastore.rawstore.impl</name>
                <value>com.datastax.bdp.hadoop.hive.metastore.CassandraHiveMetaStore</value>
                <description>Use the Apache Cassandra Hive RawStore implementation</description>
          </property>

         <property>
                <name>hadoop.bin.path</name>
                <value>${dse.bin}/dse hadoop</value>
          </property>

          <!-- Set this to true to enable auto-creation of Cassandra keyspaces as Hive Databases -->
          <property>
                <name>cassandra.autoCreateHiveSchema</name>
                <value>true</value>
          </property>
    </configuration>

5) while running Shark shell getting error:

 Unable to instantiate Org.apache.hadoop.hive.metastore.HiveMetaStoreClient

And

6) While running shark shell with -skipRddReload - I'm able to get Shark shell but not able to connect hive and not able execute any commands.

 shark> DESCRIVE mykeyspace;

and getting error message:

FAILED: Error in metastore: java.lang.RuntimeException: Unable to instantiate org.apache.haddop.hive.metastore.HiveMataStoreClient.
FAILED: Execution Error, return code 1 from org.apache.hadoop.hive.q1.exec.DDLTask.

Please provide details how to configure spark/shark on Datastax enterprise (Cassandra).

ChrisGPT was on strike
  • 127,765
  • 105
  • 273
  • 257
user3632180
  • 105
  • 2
  • 13
  • Have you looked at [this issue](http://stackoverflow.com/questions/22711364/java-lang-runtimeexceptionunable-to-instantiate-org-apache-hadoop-hive-metastor)? – rs_atl Jun 10 '14 at 18:39
  • Yes, i have seen that issue but my issue is on shark shell on cassandra (DSE). – user3632180 Jun 11 '14 at 05:15

0 Answers0