You are viewing a plain text version of this content. The canonical link for it is here.
Posted to users@zeppelin.apache.org by Ruslan Dautkhanov <da...@gmail.com> on 2016/11/14 08:07:39 UTC

Zeppelin Spark2 / Hive issue

Dear Apache Zeppelin User group,

Got Zeppelin running, but can't get %spark.sql interpreter running
correctly,
getting [1] in console output.

Running latest Zeppelin (0.6.2), Spark 2.0, Hive 1.1, Hadoop 2.6, Java 7.

My understanding is that Spark wants to initialize Hive Context, and Hive
isn't getting its config correclty so tries to create a Derby local database
in ZEPPELIN_HOME/bin/
Nope, we don't use Derby for HMS and it's using a remote RDBMS, so clearly
Hive/Spark isn't getting Hive settings correctly.

Btw, adding
export ZEPPELIN_SPARK_USEHIVECONTEXT=false
to zeppelin-env.sh didn't change anything.

See [2] for zeppelin-env.sh .


PS. I was trying to workaround this through setting
export SPARK_CLASSPATH=$HIVE_HOME/hive/lib/*:/etc/hive/conf
but it gives other problems - complains
"Caused by: org.apache.spark.SparkException: Found both
*spark.driver.extraClassPath
and* *SPARK_CLASSPATH*. *Use only the former*."
Sidenote: that's interesting that spark-shell with the same SPARK_CLASSPATH
gives opposite message -
" *SPARK_CLASSPATH *was detected (set to '...').
*This is deprecated in Spark 1.0*+. "



Thank you,
Ruslan Dautkhanov


[0]

export JAVA_HOME=/usr/java/java7
export ZEPPELIN_MEM="-Xms1024m -Xmx2048m -XX:MaxPermSize=512m"

export ZEPPELIN_LOG_DIR="/home/rdautkha/zeppelin/log"
export ZEPPELIN_PID_DIR="/home/rdautkha/zeppelin/run"
export ZEPPELIN_WAR_TEMPDIR="/home/rdautkha/zeppelin/tmp"
export ZEPPELIN_NOTEBOOK_DIR="/home/rdautkha/zeppelin/notebooks"

export SPARK_HOME=/opt/cloudera/parcels/SPARK2/lib/spark2
export SPARK_SUBMIT_OPTIONS="--principal rdautkhanov@CORP.EPSILON.COM
--keytab /home/rdautkha/.kt --conf spark.driver.memory=7g --conf
spark.executor.cores=2 --conf spark.executor.memory=8g"
export SPARK_APP_NAME="Zeppelin (test notebook 1)"

export HADOOP_CONF_DIR=/etc/hadoop/conf

export PYSPARK_PYTHON="/opt/cloudera/parcels/Anaconda/bin/python2"
export
PYTHONPATH="/opt/cloudera/parcels/SPARK2/lib/spark2/python:/opt/cloudera/parcels/SPARK2/lib/spark2/python/lib/py4j-0.10.3-src.zip"
export MASTER="yarn-client"

# these last three lines (and their combinations) were added in attempt to
resolve the the problem:
export
SPARK_CLASSPATH=/opt/cloudera/parcels/CDH/lib/hive/lib/*:/etc/hive/conf
export ZEPPELIN_SPARK_USEHIVECONTEXT=false
export HIVE_CONF_DIR=/etc/hive/conf



[1]

Caused by: ERROR XBM0H: Directory
/opt/zeppelin/zeppelin-0.6.2-bin-all/bin/metastore_db cannot be created.
        at
org.apache.derby.iapi.error.StandardException.newException(Unknown Source)
        at
org.apache.derby.iapi.error.StandardException.newException(Unknown Source)
        at
org.apache.derby.impl.services.monitor.StorageFactoryService$10.run(Unknown
Source)
        at java.security.AccessController.doPrivileged(Native Method)
        at
org.apache.derby.impl.services.monitor.StorageFactoryService.createServiceRoot(Unknown
Source)
        at
org.apache.derby.impl.services.monitor.BaseMonitor.bootService(Unknown
Source)
        at
org.apache.derby.impl.services.monitor.BaseMonitor.createPersistentService(Unknown
Source)
        at
org.apache.derby.impl.services.monitor.FileMonitor.createPersistentService(Unknown
Source)
        at
org.apache.derby.iapi.services.monitor.Monitor.createPersistentService(Unknown
Source)
        ... 104 more
============= end nested exception, level (1) ===========
============= begin nested exception, level (2) ===========
ERROR XBM0H: Directory
/opt/zeppelin/zeppelin-0.6.2-bin-all/bin/metastore_db cannot be created.
        at
org.apache.derby.iapi.error.StandardException.newException(Unknown Source)
        at
org.apache.derby.iapi.error.StandardException.newException(Unknown Source)
        at
org.apache.derby.impl.services.monitor.StorageFactoryService$10.run(Unknown
Source)
        at java.security.AccessController.doPrivileged(Native Method)
        at
org.apache.derby.impl.services.monitor.StorageFactoryService.createServiceRoot(Unknown
Source)
        at
org.apache.derby.impl.services.monitor.BaseMonitor.bootService(Unknown
Source)
        at
org.apache.derby.impl.services.monitor.BaseMonitor.createPersistentService(Unknown
Source)
        at
org.apache.derby.impl.services.monitor.FileMonitor.createPersistentService(Unknown
Source)
        at
org.apache.derby.iapi.services.monitor.Monitor.createPersistentService(Unknown
Source)
        at
org.apache.derby.impl.jdbc.EmbedConnection.createDatabase(Unknown Source)
        at org.apache.derby.impl.jdbc.EmbedConnection.<init>(Unknown Source)
        at
org.apache.derby.jdbc.InternalDriver.getNewEmbedConnection(Unknown Source)
        at org.apache.derby.jdbc.InternalDriver.connect(Unknown Source)
        at org.apache.derby.jdbc.InternalDriver.connect(Unknown Source)
        at org.apache.derby.jdbc.AutoloadedDriver.connect(Unknown Source)
        at java.sql.DriverManager.getConnection(DriverManager.java:571)
        at java.sql.DriverManager.getConnection(DriverManager.java:187)
        at
com.jolbox.bonecp.BoneCP.obtainRawInternalConnection(BoneCP.java:361)
        at com.jolbox.bonecp.BoneCP.<init>(BoneCP.java:416)
        at
com.jolbox.bonecp.BoneCPDataSource.getConnection(BoneCPDataSource.java:120)
        at
org.datanucleus.store.rdbms.ConnectionFactoryImpl$ManagedConnectionImpl.getConnection(ConnectionFactoryImpl.java:501)
        at
org.datanucleus.store.rdbms.RDBMSStoreManager.<init>(RDBMSStoreManager.java:298)
        at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native
Method)
        at
sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57)
        at
sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
        at java.lang.reflect.Constructor.newInstance(Constructor.java:526)
        at
org.datanucleus.plugin.NonManagedPluginRegistry.createExecutableExtension(NonManagedPluginRegistry.java:631)
        at
org.datanucleus.plugin.PluginManager.createExecutableExtension(PluginManager.java:301)
        at
org.datanucleus.NucleusContext.createStoreManagerForProperties(NucleusContext.java:1187)
        at
org.datanucleus.NucleusContext.initialise(NucleusContext.java:356)
        at
org.datanucleus.api.jdo.JDOPersistenceManagerFactory.freezeConfiguration(JDOPersistenceManagerFactory.java:775)
        at
org.datanucleus.api.jdo.JDOPersistenceManagerFactory.createPersistenceManagerFactory(JDOPersistenceManagerFactory.java:333)
        at
org.datanucleus.api.jdo.JDOPersistenceManagerFactory.getPersistenceManagerFactory(JDOPersistenceManagerFactory.java:202)
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
        at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
        at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
        at java.lang.reflect.Method.invoke(Method.java:606)
        at javax.jdo.JDOHelper$16.run(JDOHelper.java:1965)
        at java.security.AccessController.doPrivileged(Native Method)
        at javax.jdo.JDOHelper.invoke(JDOHelper.java:1960)
        at
javax.jdo.JDOHelper.invokeGetPersistenceManagerFactoryOnImplementation(JDOHelper.java:1166)
        at
javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:808)
        at
javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:701)
        at
org.apache.hadoop.hive.metastore.ObjectStore.getPMF(ObjectStore.java:411)
        at
org.apache.hadoop.hive.metastore.ObjectStore.getPersistenceManager(ObjectStore.java:440)
        at
org.apache.hadoop.hive.metastore.ObjectStore.initialize(ObjectStore.java:335)
        at
org.apache.hadoop.hive.metastore.ObjectStore.setConf(ObjectStore.java:291)
        at
org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:73)
        at
org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:133)
        at
org.apache.hadoop.hive.metastore.RawStoreProxy.<init>(RawStoreProxy.java:57)

Re: Zeppelin Spark2 / Hive issue

Posted by Ruslan Dautkhanov <da...@gmail.com>.
I am now not getting that error when using %sql, but Zeppelin/Hive jar
still creates a local metastore
instead of using our prod HMS, each time I use Zeppelin, it creates a Derby
files in bin/ directory from where I started Zeppelin.

$ ll
total 12
-rw-r--r-- 1 rdautkha gopher  746 Nov 18 16:06 derby.log
drwxr-xr-x 5 rdautkha gopher 4096 Nov 18 16:05 metastore_db
-rwxr-xr-x 1 rdautkha gopher   98 Nov 14 18:05 zeppelin.sh

Here's one of the stacks:

...
 INFO [2016-11-18 16:05:49,638] ({pool-2-thread-2}
ObjectStore.java[initialize]:333) - ObjectStore, initialize called
 INFO [2016-11-18 16:05:49,652] ({pool-2-thread-2}
Log4JLogger.java[info]:77) - Reading in results for query
"org.datanucleus.store.rdbms.query.SQLQuery@0" since the connection used is
closing
 INFO [2016-11-18 16:05:49,653] ({pool-2-thread-2}
MetaStoreDirectSql.java[<init>]:138) - Using direct SQL, underlying DB is
DERBY
 INFO [2016-11-18 16:05:49,653] ({pool-2-thread-2}
ObjectStore.java[setConf]:316) - Initialized ObjectStore
ERROR [2016-11-18 16:05:49,685] ({pool-2-thread-2}
RetryingHMSHandler.java[invokeInternal]:189) -
AlreadyExistsException(message:Database default already exists)
        at
org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.create_database(HiveMetaStore.java:941)
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
        at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
        at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
        at java.lang.reflect.Method.invoke(Method.java:606)
        at
org.apache.hadoop.hive.metastore.RetryingHMSHandler.invokeInternal(RetryingHMSHandler.java:138)
        at
org.apache.hadoop.hive.metastore.RetryingHMSHandler.invoke(RetryingHMSHandler.java:99)
        at com.sun.proxy.$Proxy29.create_database(Unknown Source)
        at
org.apache.hadoop.hive.metastore.HiveMetaStoreClient.createDatabase(HiveMetaStoreClient.java:656)
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
        at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
        at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
        at java.lang.reflect.Method.invoke(Method.java:606)
        at
org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.invoke(RetryingMetaStoreClient.java:105)
        at com.sun.proxy.$Proxy30.createDatabase(Unknown Source)
        at
org.apache.hadoop.hive.ql.metadata.Hive.createDatabase(Hive.java:350)
        at
org.apache.spark.sql.hive.client.HiveClientImpl$$anonfun$createDatabase$1.apply$mcV$sp(HiveClientImpl.scala:309)
        at
org.apache.spark.sql.hive.client.HiveClientImpl$$anonfun$createDatabase$1.apply(HiveClientImpl.scala:309)
        at
org.apache.spark.sql.hive.client.HiveClientImpl$$anonfun$createDatabase$1.apply(HiveClientImpl.scala:309)
        at
org.apache.spark.sql.hive.client.HiveClientImpl$$anonfun$withHiveState$1.apply(HiveClientImpl.scala:280)
        at
org.apache.spark.sql.hive.client.HiveClientImpl.liftedTree1$1(HiveClientImpl.scala:227)
        at
org.apache.spark.sql.hive.client.HiveClientImpl.retryLocked(HiveClientImpl.scala:226)
        at
org.apache.spark.sql.hive.client.HiveClientImpl.withHiveState(HiveClientImpl.scala:269)
        at
org.apache.spark.sql.hive.client.HiveClientImpl.createDatabase(HiveClientImpl.scala:308)
        at
org.apache.spark.sql.hive.HiveExternalCatalog$$anonfun$createDatabase$1.apply$mcV$sp(HiveExternalCatalog.scala:99)
        at
org.apache.spark.sql.hive.HiveExternalCatalog$$anonfun$createDatabase$1.apply(HiveExternalCatalog.scala:99)
        at
org.apache.spark.sql.hive.HiveExternalCatalog$$anonfun$createDatabase$1.apply(HiveExternalCatalog.scala:99)
        at
org.apache.spark.sql.hive.HiveExternalCatalog.withClient(HiveExternalCatalog.scala:72)
        at
org.apache.spark.sql.hive.HiveExternalCatalog.createDatabase(HiveExternalCatalog.scala:98)
        at
org.apache.spark.sql.catalyst.catalog.SessionCatalog.createDatabase(SessionCatalog.scala:147)
        at
org.apache.spark.sql.catalyst.catalog.SessionCatalog.<init>(SessionCatalog.scala:89)
        at
org.apache.spark.sql.hive.HiveSessionCatalog.<init>(HiveSessionCatalog.scala:51)
        at
org.apache.spark.sql.hive.HiveSessionState.catalog$lzycompute(HiveSessionState.scala:49)
        at
org.apache.spark.sql.hive.HiveSessionState.catalog(HiveSessionState.scala:48)
        at
org.apache.spark.sql.hive.HiveSessionState$$anon$1.<init>(HiveSessionState.scala:63)
        at
org.apache.spark.sql.hive.HiveSessionState.analyzer$lzycompute(HiveSessionState.scala:63)
        at
org.apache.spark.sql.hive.HiveSessionState.analyzer(HiveSessionState.scala:62)




-- 
Ruslan Dautkhanov

On Mon, Nov 14, 2016 at 7:23 AM, herman.yu@teeupdata.com <
herman.yu@teeupdata.com> wrote:

> You may check if you have hive-site.xml under zeppelin/spark  config
> folder…
>
> Thanks
> Herman.
>
>
>
> On Nov 14, 2016, at 03:07, Ruslan Dautkhanov <da...@gmail.com> wrote:
>
> Dear Apache Zeppelin User group,
>
> Got Zeppelin running, but can't get %spark.sql interpreter running
> correctly,
> getting [1] in console output.
>
> Running latest Zeppelin (0.6.2), Spark 2.0, Hive 1.1, Hadoop 2.6, Java 7.
>
> My understanding is that Spark wants to initialize Hive Context, and Hive
> isn't getting its config correclty so tries to create a Derby local
> database
> in ZEPPELIN_HOME/bin/
> Nope, we don't use Derby for HMS and it's using a remote RDBMS, so clearly
> Hive/Spark isn't getting Hive settings correctly.
>
> Btw, adding
> export ZEPPELIN_SPARK_USEHIVECONTEXT=false
> to zeppelin-env.sh didn't change anything.
>
> See [2] for zeppelin-env.sh .
>
>
> PS. I was trying to workaround this through setting
> export SPARK_CLASSPATH=$HIVE_HOME/hive/lib/*:/etc/hive/conf
> but it gives other problems - complains
> "Caused by: org.apache.spark.SparkException: Found both *spark.driver.extraClassPath
> and* *SPARK_CLASSPATH*. *Use only the former*."
> Sidenote: that's interesting that spark-shell with the same SPARK_CLASSPATH
> gives opposite message -
> " *SPARK_CLASSPATH *was detected (set to '...').
> *This is deprecated in Spark 1.0*+. "
>
>
>
> Thank you,
> Ruslan Dautkhanov
>
>
> [0]
>
> export JAVA_HOME=/usr/java/java7
> export ZEPPELIN_MEM="-Xms1024m -Xmx2048m -XX:MaxPermSize=512m"
>
> export ZEPPELIN_LOG_DIR="/home/rdautkha/zeppelin/log"
> export ZEPPELIN_PID_DIR="/home/rdautkha/zeppelin/run"
> export ZEPPELIN_WAR_TEMPDIR="/home/rdautkha/zeppelin/tmp"
> export ZEPPELIN_NOTEBOOK_DIR="/home/rdautkha/zeppelin/notebooks"
>
> export SPARK_HOME=/opt/cloudera/parcels/SPARK2/lib/spark2
> export SPARK_SUBMIT_OPTIONS="--principal rdautkhanov@CORP.EPSILON.COM
> --keytab /home/rdautkha/.kt --conf spark.driver.memory=7g --conf
> spark.executor.cores=2 --conf spark.executor.memory=8g"
> export SPARK_APP_NAME="Zeppelin (test notebook 1)"
>
> export HADOOP_CONF_DIR=/etc/hadoop/conf
>
> export PYSPARK_PYTHON="/opt/cloudera/parcels/Anaconda/bin/python2"
> export PYTHONPATH="/opt/cloudera/parcels/SPARK2/lib/spark2/
> python:/opt/cloudera/parcels/SPARK2/lib/spark2/python/lib/
> py4j-0.10.3-src.zip"
> export MASTER="yarn-client"
>
> # these last three lines (and their combinations) were added in attempt to
> resolve the the problem:
> export SPARK_CLASSPATH=/opt/cloudera/parcels/CDH/lib/hive/lib/*:/
> etc/hive/conf
> export ZEPPELIN_SPARK_USEHIVECONTEXT=false
> export HIVE_CONF_DIR=/etc/hive/conf
>
>
>
> [1]
>
> Caused by: ERROR XBM0H: Directory /opt/zeppelin/zeppelin-0.6.2-bin-all/bin/metastore_db
> cannot be created.
>         at org.apache.derby.iapi.error.StandardException.newException(Unknown
> Source)
>         at org.apache.derby.iapi.error.StandardException.newException(Unknown
> Source)
>         at org.apache.derby.impl.services.monitor.
> StorageFactoryService$10.run(Unknown Source)
>         at java.security.AccessController.doPrivileged(Native Method)
>         at org.apache.derby.impl.services.monitor.StorageFactoryService.createServiceRoot(Unknown
> Source)
>         at org.apache.derby.impl.services.monitor.BaseMonitor.bootService(Unknown
> Source)
>         at org.apache.derby.impl.services.monitor.BaseMonitor.
> createPersistentService(Unknown Source)
>         at org.apache.derby.impl.services.monitor.FileMonitor.
> createPersistentService(Unknown Source)
>         at org.apache.derby.iapi.services.monitor.Monitor.
> createPersistentService(Unknown Source)
>         ... 104 more
> ============= end nested exception, level (1) ===========
> ============= begin nested exception, level (2) ===========
> ERROR XBM0H: Directory /opt/zeppelin/zeppelin-0.6.2-bin-all/bin/metastore_db
> cannot be created.
>         at org.apache.derby.iapi.error.StandardException.newException(Unknown
> Source)
>         at org.apache.derby.iapi.error.StandardException.newException(Unknown
> Source)
>         at org.apache.derby.impl.services.monitor.
> StorageFactoryService$10.run(Unknown Source)
>         at java.security.AccessController.doPrivileged(Native Method)
>         at org.apache.derby.impl.services.monitor.StorageFactoryService.createServiceRoot(Unknown
> Source)
>         at org.apache.derby.impl.services.monitor.BaseMonitor.bootService(Unknown
> Source)
>         at org.apache.derby.impl.services.monitor.BaseMonitor.
> createPersistentService(Unknown Source)
>         at org.apache.derby.impl.services.monitor.FileMonitor.
> createPersistentService(Unknown Source)
>         at org.apache.derby.iapi.services.monitor.Monitor.
> createPersistentService(Unknown Source)
>         at org.apache.derby.impl.jdbc.EmbedConnection.createDatabase(Unknown
> Source)
>         at org.apache.derby.impl.jdbc.EmbedConnection.<init>(Unknown
> Source)
>         at org.apache.derby.jdbc.InternalDriver.getNewEmbedConnection(Unknown
> Source)
>         at org.apache.derby.jdbc.InternalDriver.connect(Unknown Source)
>         at org.apache.derby.jdbc.InternalDriver.connect(Unknown Source)
>         at org.apache.derby.jdbc.AutoloadedDriver.connect(Unknown Source)
>         at java.sql.DriverManager.getConnection(DriverManager.java:571)
>         at java.sql.DriverManager.getConnection(DriverManager.java:187)
>         at com.jolbox.bonecp.BoneCP.obtainRawInternalConnection(
> BoneCP.java:361)
>         at com.jolbox.bonecp.BoneCP.<init>(BoneCP.java:416)
>         at com.jolbox.bonecp.BoneCPDataSource.getConnection(
> BoneCPDataSource.java:120)
>         at org.datanucleus.store.rdbms.ConnectionFactoryImpl$
> ManagedConnectionImpl.getConnection(ConnectionFactoryImpl.java:501)
>         at org.datanucleus.store.rdbms.RDBMSStoreManager.<init>(
> RDBMSStoreManager.java:298)
>         at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native
> Method)
>         at sun.reflect.NativeConstructorAccessorImpl.newInstance(
> NativeConstructorAccessorImpl.java:57)
>         at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(
> DelegatingConstructorAccessorImpl.java:45)
>         at java.lang.reflect.Constructor.newInstance(Constructor.java:526)
>         at org.datanucleus.plugin.NonManagedPluginRegistry.
> createExecutableExtension(NonManagedPluginRegistry.java:631)
>         at org.datanucleus.plugin.PluginManager.createExecutableExtension(
> PluginManager.java:301)
>         at org.datanucleus.NucleusContext.createStoreManagerForPropertie
> s(NucleusContext.java:1187)
>         at org.datanucleus.NucleusContext.initialise(
> NucleusContext.java:356)
>         at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.
> freezeConfiguration(JDOPersistenceManagerFactory.java:775)
>         at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.
> createPersistenceManagerFactory(JDOPersistenceManagerFactory.java:333)
>         at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.
> getPersistenceManagerFactory(JDOPersistenceManagerFactory.java:202)
>         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>         at sun.reflect.NativeMethodAccessorImpl.invoke(
> NativeMethodAccessorImpl.java:57)
>         at sun.reflect.DelegatingMethodAccessorImpl.invoke(
> DelegatingMethodAccessorImpl.java:43)
>         at java.lang.reflect.Method.invoke(Method.java:606)
>         at javax.jdo.JDOHelper$16.run(JDOHelper.java:1965)
>         at java.security.AccessController.doPrivileged(Native Method)
>         at javax.jdo.JDOHelper.invoke(JDOHelper.java:1960)
>         at javax.jdo.JDOHelper.invokeGetPersistenceManagerFac
> toryOnImplementation(JDOHelper.java:1166)
>         at javax.jdo.JDOHelper.getPersistenceManagerFactory(
> JDOHelper.java:808)
>         at javax.jdo.JDOHelper.getPersistenceManagerFactory(
> JDOHelper.java:701)
>         at org.apache.hadoop.hive.metastore.ObjectStore.getPMF(
> ObjectStore.java:411)
>         at org.apache.hadoop.hive.metastore.ObjectStore.
> getPersistenceManager(ObjectStore.java:440)
>         at org.apache.hadoop.hive.metastore.ObjectStore.
> initialize(ObjectStore.java:335)
>         at org.apache.hadoop.hive.metastore.ObjectStore.setConf(
> ObjectStore.java:291)
>         at org.apache.hadoop.util.ReflectionUtils.setConf(
> ReflectionUtils.java:73)
>         at org.apache.hadoop.util.ReflectionUtils.newInstance(
> ReflectionUtils.java:133)
>         at org.apache.hadoop.hive.metastore.RawStoreProxy.<init>
> (RawStoreProxy.java:57)
>
>
>
>
>

Re: Zeppelin Spark2 / Hive issue

Posted by Ruslan Dautkhanov <da...@gmail.com>.
Thank you Herman!

That was it.




-- 
Ruslan Dautkhanov

On Mon, Nov 14, 2016 at 7:23 AM, herman.yu@teeupdata.com <
herman.yu@teeupdata.com> wrote:

> You may check if you have hive-site.xml under zeppelin/spark  config
> folder…
>
> Thanks
> Herman.
>
>
>
> On Nov 14, 2016, at 03:07, Ruslan Dautkhanov <da...@gmail.com> wrote:
>
> Dear Apache Zeppelin User group,
>
> Got Zeppelin running, but can't get %spark.sql interpreter running
> correctly,
> getting [1] in console output.
>
> Running latest Zeppelin (0.6.2), Spark 2.0, Hive 1.1, Hadoop 2.6, Java 7.
>
> My understanding is that Spark wants to initialize Hive Context, and Hive
> isn't getting its config correclty so tries to create a Derby local
> database
> in ZEPPELIN_HOME/bin/
> Nope, we don't use Derby for HMS and it's using a remote RDBMS, so clearly
> Hive/Spark isn't getting Hive settings correctly.
>
> Btw, adding
> export ZEPPELIN_SPARK_USEHIVECONTEXT=false
> to zeppelin-env.sh didn't change anything.
>
> See [2] for zeppelin-env.sh .
>
>
> PS. I was trying to workaround this through setting
> export SPARK_CLASSPATH=$HIVE_HOME/hive/lib/*:/etc/hive/conf
> but it gives other problems - complains
> "Caused by: org.apache.spark.SparkException: Found both *spark.driver.extraClassPath
> and* *SPARK_CLASSPATH*. *Use only the former*."
> Sidenote: that's interesting that spark-shell with the same SPARK_CLASSPATH
> gives opposite message -
> " *SPARK_CLASSPATH *was detected (set to '...').
> *This is deprecated in Spark 1.0*+. "
>
>
>
> Thank you,
> Ruslan Dautkhanov
>
>
> [0]
>
> export JAVA_HOME=/usr/java/java7
> export ZEPPELIN_MEM="-Xms1024m -Xmx2048m -XX:MaxPermSize=512m"
>
> export ZEPPELIN_LOG_DIR="/home/rdautkha/zeppelin/log"
> export ZEPPELIN_PID_DIR="/home/rdautkha/zeppelin/run"
> export ZEPPELIN_WAR_TEMPDIR="/home/rdautkha/zeppelin/tmp"
> export ZEPPELIN_NOTEBOOK_DIR="/home/rdautkha/zeppelin/notebooks"
>
> export SPARK_HOME=/opt/cloudera/parcels/SPARK2/lib/spark2
> export SPARK_SUBMIT_OPTIONS="--principal rdautkhanov@CORP.EPSILON.COM
> --keytab /home/rdautkha/.kt --conf spark.driver.memory=7g --conf
> spark.executor.cores=2 --conf spark.executor.memory=8g"
> export SPARK_APP_NAME="Zeppelin (test notebook 1)"
>
> export HADOOP_CONF_DIR=/etc/hadoop/conf
>
> export PYSPARK_PYTHON="/opt/cloudera/parcels/Anaconda/bin/python2"
> export PYTHONPATH="/opt/cloudera/parcels/SPARK2/lib/spark2/
> python:/opt/cloudera/parcels/SPARK2/lib/spark2/python/lib/
> py4j-0.10.3-src.zip"
> export MASTER="yarn-client"
>
> # these last three lines (and their combinations) were added in attempt to
> resolve the the problem:
> export SPARK_CLASSPATH=/opt/cloudera/parcels/CDH/lib/hive/lib/*:/
> etc/hive/conf
> export ZEPPELIN_SPARK_USEHIVECONTEXT=false
> export HIVE_CONF_DIR=/etc/hive/conf
>
>
>
> [1]
>
> Caused by: ERROR XBM0H: Directory /opt/zeppelin/zeppelin-0.6.2-bin-all/bin/metastore_db
> cannot be created.
>         at org.apache.derby.iapi.error.StandardException.newException(Unknown
> Source)
>         at org.apache.derby.iapi.error.StandardException.newException(Unknown
> Source)
>         at org.apache.derby.impl.services.monitor.
> StorageFactoryService$10.run(Unknown Source)
>         at java.security.AccessController.doPrivileged(Native Method)
>         at org.apache.derby.impl.services.monitor.StorageFactoryService.createServiceRoot(Unknown
> Source)
>         at org.apache.derby.impl.services.monitor.BaseMonitor.bootService(Unknown
> Source)
>         at org.apache.derby.impl.services.monitor.BaseMonitor.
> createPersistentService(Unknown Source)
>         at org.apache.derby.impl.services.monitor.FileMonitor.
> createPersistentService(Unknown Source)
>         at org.apache.derby.iapi.services.monitor.Monitor.
> createPersistentService(Unknown Source)
>         ... 104 more
> ============= end nested exception, level (1) ===========
> ============= begin nested exception, level (2) ===========
> ERROR XBM0H: Directory /opt/zeppelin/zeppelin-0.6.2-bin-all/bin/metastore_db
> cannot be created.
>         at org.apache.derby.iapi.error.StandardException.newException(Unknown
> Source)
>         at org.apache.derby.iapi.error.StandardException.newException(Unknown
> Source)
>         at org.apache.derby.impl.services.monitor.
> StorageFactoryService$10.run(Unknown Source)
>         at java.security.AccessController.doPrivileged(Native Method)
>         at org.apache.derby.impl.services.monitor.StorageFactoryService.createServiceRoot(Unknown
> Source)
>         at org.apache.derby.impl.services.monitor.BaseMonitor.bootService(Unknown
> Source)
>         at org.apache.derby.impl.services.monitor.BaseMonitor.
> createPersistentService(Unknown Source)
>         at org.apache.derby.impl.services.monitor.FileMonitor.
> createPersistentService(Unknown Source)
>         at org.apache.derby.iapi.services.monitor.Monitor.
> createPersistentService(Unknown Source)
>         at org.apache.derby.impl.jdbc.EmbedConnection.createDatabase(Unknown
> Source)
>         at org.apache.derby.impl.jdbc.EmbedConnection.<init>(Unknown
> Source)
>         at org.apache.derby.jdbc.InternalDriver.getNewEmbedConnection(Unknown
> Source)
>         at org.apache.derby.jdbc.InternalDriver.connect(Unknown Source)
>         at org.apache.derby.jdbc.InternalDriver.connect(Unknown Source)
>         at org.apache.derby.jdbc.AutoloadedDriver.connect(Unknown Source)
>         at java.sql.DriverManager.getConnection(DriverManager.java:571)
>         at java.sql.DriverManager.getConnection(DriverManager.java:187)
>         at com.jolbox.bonecp.BoneCP.obtainRawInternalConnection(
> BoneCP.java:361)
>         at com.jolbox.bonecp.BoneCP.<init>(BoneCP.java:416)
>         at com.jolbox.bonecp.BoneCPDataSource.getConnection(
> BoneCPDataSource.java:120)
>         at org.datanucleus.store.rdbms.ConnectionFactoryImpl$
> ManagedConnectionImpl.getConnection(ConnectionFactoryImpl.java:501)
>         at org.datanucleus.store.rdbms.RDBMSStoreManager.<init>(
> RDBMSStoreManager.java:298)
>         at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native
> Method)
>         at sun.reflect.NativeConstructorAccessorImpl.newInstance(
> NativeConstructorAccessorImpl.java:57)
>         at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(
> DelegatingConstructorAccessorImpl.java:45)
>         at java.lang.reflect.Constructor.newInstance(Constructor.java:526)
>         at org.datanucleus.plugin.NonManagedPluginRegistry.
> createExecutableExtension(NonManagedPluginRegistry.java:631)
>         at org.datanucleus.plugin.PluginManager.createExecutableExtension(
> PluginManager.java:301)
>         at org.datanucleus.NucleusContext.createStoreManagerForPropertie
> s(NucleusContext.java:1187)
>         at org.datanucleus.NucleusContext.initialise(
> NucleusContext.java:356)
>         at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.
> freezeConfiguration(JDOPersistenceManagerFactory.java:775)
>         at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.
> createPersistenceManagerFactory(JDOPersistenceManagerFactory.java:333)
>         at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.
> getPersistenceManagerFactory(JDOPersistenceManagerFactory.java:202)
>         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>         at sun.reflect.NativeMethodAccessorImpl.invoke(
> NativeMethodAccessorImpl.java:57)
>         at sun.reflect.DelegatingMethodAccessorImpl.invoke(
> DelegatingMethodAccessorImpl.java:43)
>         at java.lang.reflect.Method.invoke(Method.java:606)
>         at javax.jdo.JDOHelper$16.run(JDOHelper.java:1965)
>         at java.security.AccessController.doPrivileged(Native Method)
>         at javax.jdo.JDOHelper.invoke(JDOHelper.java:1960)
>         at javax.jdo.JDOHelper.invokeGetPersistenceManagerFac
> toryOnImplementation(JDOHelper.java:1166)
>         at javax.jdo.JDOHelper.getPersistenceManagerFactory(
> JDOHelper.java:808)
>         at javax.jdo.JDOHelper.getPersistenceManagerFactory(
> JDOHelper.java:701)
>         at org.apache.hadoop.hive.metastore.ObjectStore.getPMF(
> ObjectStore.java:411)
>         at org.apache.hadoop.hive.metastore.ObjectStore.
> getPersistenceManager(ObjectStore.java:440)
>         at org.apache.hadoop.hive.metastore.ObjectStore.
> initialize(ObjectStore.java:335)
>         at org.apache.hadoop.hive.metastore.ObjectStore.setConf(
> ObjectStore.java:291)
>         at org.apache.hadoop.util.ReflectionUtils.setConf(
> ReflectionUtils.java:73)
>         at org.apache.hadoop.util.ReflectionUtils.newInstance(
> ReflectionUtils.java:133)
>         at org.apache.hadoop.hive.metastore.RawStoreProxy.<init>
> (RawStoreProxy.java:57)
>
>
>
>
>

Re: Zeppelin Spark2 / Hive issue

Posted by "herman.yu@teeupdata.com" <he...@teeupdata.com>.
You may check if you have hive-site.xml under zeppelin/spark  config folder…

Thanks
Herman.



> On Nov 14, 2016, at 03:07, Ruslan Dautkhanov <da...@gmail.com> wrote:
> 
> Dear Apache Zeppelin User group,
> 
> Got Zeppelin running, but can't get %spark.sql interpreter running correctly, 
> getting [1] in console output.
> 
> Running latest Zeppelin (0.6.2), Spark 2.0, Hive 1.1, Hadoop 2.6, Java 7.
> 
> My understanding is that Spark wants to initialize Hive Context, and Hive
> isn't getting its config correclty so tries to create a Derby local database
> in ZEPPELIN_HOME/bin/ 
> Nope, we don't use Derby for HMS and it's using a remote RDBMS, so clearly
> Hive/Spark isn't getting Hive settings correctly.
> 
> Btw, adding
> export ZEPPELIN_SPARK_USEHIVECONTEXT=false
> to zeppelin-env.sh didn't change anything.
> 
> See [2] for zeppelin-env.sh .
> 
> 
> PS. I was trying to workaround this through setting
> export SPARK_CLASSPATH=$HIVE_HOME/hive/lib/*:/etc/hive/conf
> but it gives other problems - complains 
> "Caused by: org.apache.spark.SparkException: Found both spark.driver.extraClassPath and SPARK_CLASSPATH. Use only the former."
> Sidenote: that's interesting that spark-shell with the same SPARK_CLASSPATH
> gives opposite message - 
> " SPARK_CLASSPATH was detected (set to '...').
> This is deprecated in Spark 1.0+. "
> 
> 
> 
> Thank you,
> Ruslan Dautkhanov
> 
> 
> [0]
> 
> export JAVA_HOME=/usr/java/java7
> export ZEPPELIN_MEM="-Xms1024m -Xmx2048m -XX:MaxPermSize=512m"
> 
> export ZEPPELIN_LOG_DIR="/home/rdautkha/zeppelin/log"
> export ZEPPELIN_PID_DIR="/home/rdautkha/zeppelin/run"
> export ZEPPELIN_WAR_TEMPDIR="/home/rdautkha/zeppelin/tmp"
> export ZEPPELIN_NOTEBOOK_DIR="/home/rdautkha/zeppelin/notebooks"
> 
> export SPARK_HOME=/opt/cloudera/parcels/SPARK2/lib/spark2
> export SPARK_SUBMIT_OPTIONS="--principal rdautkhanov@CORP.EPSILON.COM <ma...@CORP.EPSILON.COM> --keytab /home/rdautkha/.kt --conf spark.driver.memory=7g --conf spark.executor.cores=2 --conf spark.executor.memory=8g"
> export SPARK_APP_NAME="Zeppelin (test notebook 1)"
> 
> export HADOOP_CONF_DIR=/etc/hadoop/conf
> 
> export PYSPARK_PYTHON="/opt/cloudera/parcels/Anaconda/bin/python2"
> export PYTHONPATH="/opt/cloudera/parcels/SPARK2/lib/spark2/python:/opt/cloudera/parcels/SPARK2/lib/spark2/python/lib/py4j-0.10.3-src.zip"
> export MASTER="yarn-client"
> 
> # these last three lines (and their combinations) were added in attempt to resolve the the problem:
> export SPARK_CLASSPATH=/opt/cloudera/parcels/CDH/lib/hive/lib/*:/etc/hive/conf
> export ZEPPELIN_SPARK_USEHIVECONTEXT=false
> export HIVE_CONF_DIR=/etc/hive/conf
> 
> 
> 
> [1]
> 
> Caused by: ERROR XBM0H: Directory /opt/zeppelin/zeppelin-0.6.2-bin-all/bin/metastore_db cannot be created.
>         at org.apache.derby.iapi.error.StandardException.newException(Unknown Source)
>         at org.apache.derby.iapi.error.StandardException.newException(Unknown Source)
>         at org.apache.derby.impl.services.monitor.StorageFactoryService$10.run(Unknown Source)
>         at java.security.AccessController.doPrivileged(Native Method)
>         at org.apache.derby.impl.services.monitor.StorageFactoryService.createServiceRoot(Unknown Source)
>         at org.apache.derby.impl.services.monitor.BaseMonitor.bootService(Unknown Source)
>         at org.apache.derby.impl.services.monitor.BaseMonitor.createPersistentService(Unknown Source)
>         at org.apache.derby.impl.services.monitor.FileMonitor.createPersistentService(Unknown Source)
>         at org.apache.derby.iapi.services.monitor.Monitor.createPersistentService(Unknown Source)
>         ... 104 more
> ============= end nested exception, level (1) ===========
> ============= begin nested exception, level (2) ===========
> ERROR XBM0H: Directory /opt/zeppelin/zeppelin-0.6.2-bin-all/bin/metastore_db cannot be created.
>         at org.apache.derby.iapi.error.StandardException.newException(Unknown Source)
>         at org.apache.derby.iapi.error.StandardException.newException(Unknown Source)
>         at org.apache.derby.impl.services.monitor.StorageFactoryService$10.run(Unknown Source)
>         at java.security.AccessController.doPrivileged(Native Method)
>         at org.apache.derby.impl.services.monitor.StorageFactoryService.createServiceRoot(Unknown Source)
>         at org.apache.derby.impl.services.monitor.BaseMonitor.bootService(Unknown Source)
>         at org.apache.derby.impl.services.monitor.BaseMonitor.createPersistentService(Unknown Source)
>         at org.apache.derby.impl.services.monitor.FileMonitor.createPersistentService(Unknown Source)
>         at org.apache.derby.iapi.services.monitor.Monitor.createPersistentService(Unknown Source)
>         at org.apache.derby.impl.jdbc.EmbedConnection.createDatabase(Unknown Source)
>         at org.apache.derby.impl.jdbc.EmbedConnection.<init>(Unknown Source)
>         at org.apache.derby.jdbc.InternalDriver.getNewEmbedConnection(Unknown Source)
>         at org.apache.derby.jdbc.InternalDriver.connect(Unknown Source)
>         at org.apache.derby.jdbc.InternalDriver.connect(Unknown Source)
>         at org.apache.derby.jdbc.AutoloadedDriver.connect(Unknown Source)
>         at java.sql.DriverManager.getConnection(DriverManager.java:571)
>         at java.sql.DriverManager.getConnection(DriverManager.java:187)
>         at com.jolbox.bonecp.BoneCP.obtainRawInternalConnection(BoneCP.java:361)
>         at com.jolbox.bonecp.BoneCP.<init>(BoneCP.java:416)
>         at com.jolbox.bonecp.BoneCPDataSource.getConnection(BoneCPDataSource.java:120)
>         at org.datanucleus.store.rdbms.ConnectionFactoryImpl$ManagedConnectionImpl.getConnection(ConnectionFactoryImpl.java:501)
>         at org.datanucleus.store.rdbms.RDBMSStoreManager.<init>(RDBMSStoreManager.java:298)
>         at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
>         at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57)
>         at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
>         at java.lang.reflect.Constructor.newInstance(Constructor.java:526)
>         at org.datanucleus.plugin.NonManagedPluginRegistry.createExecutableExtension(NonManagedPluginRegistry.java:631)
>         at org.datanucleus.plugin.PluginManager.createExecutableExtension(PluginManager.java:301)
>         at org.datanucleus.NucleusContext.createStoreManagerForProperties(NucleusContext.java:1187)
>         at org.datanucleus.NucleusContext.initialise(NucleusContext.java:356)
>         at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.freezeConfiguration(JDOPersistenceManagerFactory.java:775)
>         at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.createPersistenceManagerFactory(JDOPersistenceManagerFactory.java:333)
>         at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.getPersistenceManagerFactory(JDOPersistenceManagerFactory.java:202)
>         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>         at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>         at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>         at java.lang.reflect.Method.invoke(Method.java:606)
>         at javax.jdo.JDOHelper$16.run(JDOHelper.java:1965)
>         at java.security.AccessController.doPrivileged(Native Method)
>         at javax.jdo.JDOHelper.invoke(JDOHelper.java:1960)
>         at javax.jdo.JDOHelper.invokeGetPersistenceManagerFactoryOnImplementation(JDOHelper.java:1166)
>         at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:808)
>         at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:701)
>         at org.apache.hadoop.hive.metastore.ObjectStore.getPMF(ObjectStore.java:411)
>         at org.apache.hadoop.hive.metastore.ObjectStore.getPersistenceManager(ObjectStore.java:440)
>         at org.apache.hadoop.hive.metastore.ObjectStore.initialize(ObjectStore.java:335)
>         at org.apache.hadoop.hive.metastore.ObjectStore.setConf(ObjectStore.java:291)
>         at org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:73)
>         at org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:133)
>         at org.apache.hadoop.hive.metastore.RawStoreProxy.<init>(RawStoreProxy.java:57)
> 
> 
>