You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hudi.apache.org by GitBox <gi...@apache.org> on 2021/11/13 18:36:42 UTC

[GitHub] [hudi] nsivabalan opened a new issue #3990: [SUPPORT] Spark-sql fails to start if there is a concurrent spark-shell open

nsivabalan opened a new issue #3990:
URL: https://github.com/apache/hudi/issues/3990


   **_Tips before filing an issue_**
   
   - Have you gone through our [FAQs](https://cwiki.apache.org/confluence/display/HUDI/FAQ)?
   
   - Join the mailing list to engage in conversations and get faster support at dev-subscribe@hudi.apache.org.
   
   - If you have triaged this as a bug, then file an [issue](https://issues.apache.org/jira/projects/HUDI/issues) directly.
   
   **Describe the problem you faced**
   
   I had spark-shell open in one window and tried to launch spark-sql in another. but the spark-sql failed to launch with an exception claiming that another instance of Derby is already booted. 
   
   **To Reproduce**
   
   Steps to reproduce the behavior:
   
   1. launch spark-shell and make 1 or 2 commits. 
   2. launch spark-sql 
   3.
   4.
   
   **Expected behavior**
   
   A clear and concise description of what you expected to happen.
   
   **Environment Description**
   
   * Hudi version :
   
   * Spark version :
   
   * Hive version :
   
   * Hadoop version :
   
   * Storage (HDFS/S3/GCS..) :
   
   * Running on Docker? (yes/no) :
   
   
   **Additional context**
   
   Add any other context about the problem here.
   
   **Stacktrace**
   
   ```
   ./bin/spark-sql --jars ~/Documents/personal/projects/oct21_3/hudi/packaging/hudi-spark-bundle/target/hudi-spark-bundle_2.11-0.10.0-SNAPSHOT.jar --conf 'spark.serializer=org.apache.spark.serializer.KryoSerializer' --conf 'spark.sql.extensions=org.apache.spark.sql.hudi.HoodieSparkSessionExtension'  --conf 'spark.kryoserializer.buffer.max=1024m' --conf "spark.memory.storageFraction=0.8"  --conf "spark.driver.extraClassPath=-XX:NewSize=1g -XX:SurvivorRatio=2 -XX:+UseCompressedOops -XX:+UseConcMarkSweepGC -XX:+UseParNewGC -XX:CMSInitiatingOccupancyFraction=70"     --conf "spark.executor.extraClassPath=-XX:NewSize=1g -XX:SurvivorRatio=2 -XX:+UseCompressedOops -XX:+UseConcMarkSweepGC -XX:+UseParNewGC -XX:CMSInitiatingOccupancyFraction=70" --conf spark.rdd.compress=true --driver-memory 8g --conf "spark.memory.fraction=0.8" --driver-java-options "-Dlog4j.configuration=file:///Users/nsb/Documents/personal/tools/log4j/spark_sql.log4j.props"
   21/11/13 13:30:30 WARN Utils: Your hostname, Sivabalans-MacBook-Pro.local resolves to a loopback address: 127.0.0.1; using 10.0.0.202 instead (on interface en0)
   21/11/13 13:30:30 WARN Utils: Set SPARK_LOCAL_IP if you need to bind to another address
   21/11/13 13:30:31 WARN NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
   21/11/13 13:30:32 ERROR Schema: Failed initialising database.
   Unable to open a test connection to the given database. JDBC url = jdbc:derby:;databaseName=metastore_db;create=true, username = APP. Terminating connection pool (set lazyInit to true if you expect to start your database after your app). Original Exception: ------
   java.sql.SQLException: Failed to start database 'metastore_db' with class loader sun.misc.Launcher$AppClassLoader@5fcfe4b2, see the next exception for details.
   	at org.apache.derby.impl.jdbc.SQLExceptionFactory.getSQLException(Unknown Source)
   	at org.apache.derby.impl.jdbc.SQLExceptionFactory.getSQLException(Unknown Source)
   	at org.apache.derby.impl.jdbc.Util.seeNextException(Unknown Source)
   	at org.apache.derby.impl.jdbc.EmbedConnection.bootDatabase(Unknown Source)
   	at org.apache.derby.impl.jdbc.EmbedConnection.<init>(Unknown Source)
   	at org.apache.derby.jdbc.InternalDriver$1.run(Unknown Source)
   	at org.apache.derby.jdbc.InternalDriver$1.run(Unknown Source)
   	at java.security.AccessController.doPrivileged(Native Method)
   	at org.apache.derby.jdbc.InternalDriver.getNewEmbedConnection(Unknown Source)
   	at org.apache.derby.jdbc.InternalDriver.connect(Unknown Source)
   	at org.apache.derby.jdbc.InternalDriver.connect(Unknown Source)
   	at org.apache.derby.jdbc.AutoloadedDriver.connect(Unknown Source)
   	at java.sql.DriverManager.getConnection(DriverManager.java:664)
   	at java.sql.DriverManager.getConnection(DriverManager.java:208)
   	at com.jolbox.bonecp.BoneCP.obtainRawInternalConnection(BoneCP.java:361)
   	at com.jolbox.bonecp.BoneCP.<init>(BoneCP.java:416)
   	at com.jolbox.bonecp.BoneCPDataSource.getConnection(BoneCPDataSource.java:120)
   	at org.datanucleus.store.rdbms.ConnectionFactoryImpl$ManagedConnectionImpl.getConnection(ConnectionFactoryImpl.java:501)
   	at org.datanucleus.store.rdbms.RDBMSStoreManager.<init>(RDBMSStoreManager.java:298)
   	at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
   	at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
   	at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
   	at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
   	at org.datanucleus.plugin.NonManagedPluginRegistry.createExecutableExtension(NonManagedPluginRegistry.java:631)
   	at org.datanucleus.plugin.PluginManager.createExecutableExtension(PluginManager.java:301)
   	at org.datanucleus.NucleusContext.createStoreManagerForProperties(NucleusContext.java:1187)
   	at org.datanucleus.NucleusContext.initialise(NucleusContext.java:356)
   	at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.freezeConfiguration(JDOPersistenceManagerFactory.java:775)
   	at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.createPersistenceManagerFactory(JDOPersistenceManagerFactory.java:333)
   	at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.getPersistenceManagerFactory(JDOPersistenceManagerFactory.java:202)
   	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
   	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
   	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
   	at java.lang.reflect.Method.invoke(Method.java:498)
   	at javax.jdo.JDOHelper$16.run(JDOHelper.java:1965)
   	at java.security.AccessController.doPrivileged(Native Method)
   	at javax.jdo.JDOHelper.invoke(JDOHelper.java:1960)
   	at javax.jdo.JDOHelper.invokeGetPersistenceManagerFactoryOnImplementation(JDOHelper.java:1166)
   	at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:808)
   	at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:701)
   	at org.apache.hadoop.hive.metastore.ObjectStore.getPMF(ObjectStore.java:365)
   	at org.apache.hadoop.hive.metastore.ObjectStore.getPersistenceManager(ObjectStore.java:394)
   	at org.apache.hadoop.hive.metastore.ObjectStore.initialize(ObjectStore.java:291)
   	at org.apache.hadoop.hive.metastore.ObjectStore.setConf(ObjectStore.java:258)
   	at org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:76)
   	at org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:136)
   	at org.apache.hadoop.hive.metastore.RawStoreProxy.<init>(RawStoreProxy.java:57)
   	at org.apache.hadoop.hive.metastore.RawStoreProxy.getProxy(RawStoreProxy.java:66)
   	at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.newRawStore(HiveMetaStore.java:593)
   	at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMS(HiveMetaStore.java:571)
   	at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.createDefaultDB(HiveMetaStore.java:620)
   	at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.init(HiveMetaStore.java:461)
   	at org.apache.hadoop.hive.metastore.RetryingHMSHandler.<init>(RetryingHMSHandler.java:66)
   	at org.apache.hadoop.hive.metastore.RetryingHMSHandler.getProxy(RetryingHMSHandler.java:72)
   	at org.apache.hadoop.hive.metastore.HiveMetaStore.newRetryingHMSHandler(HiveMetaStore.java:5762)
   	at org.apache.hadoop.hive.metastore.HiveMetaStoreClient.<init>(HiveMetaStoreClient.java:199)
   	at org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient.<init>(SessionHiveMetaStoreClient.java:74)
   	at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
   	at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
   	at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
   	at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
   	at org.apache.hadoop.hive.metastore.MetaStoreUtils.newInstance(MetaStoreUtils.java:1521)
   	at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.<init>(RetryingMetaStoreClient.java:86)
   	at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:132)
   	at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:104)
   	at org.apache.hadoop.hive.ql.metadata.Hive.createMetaStoreClient(Hive.java:3005)
   	at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3024)
   	at org.apache.hadoop.hive.ql.metadata.Hive.getAllDatabases(Hive.java:1234)
   	at org.apache.hadoop.hive.ql.metadata.Hive.reloadFunctions(Hive.java:174)
   	at org.apache.hadoop.hive.ql.metadata.Hive.<clinit>(Hive.java:166)
   	at org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.java:503)
   	at org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver$.main(SparkSQLCLIDriver.scala:133)
   	at org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver.main(SparkSQLCLIDriver.scala)
   	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
   	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
   	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
   	at java.lang.reflect.Method.invoke(Method.java:498)
   	at org.apache.spark.deploy.JavaMainApplication.start(SparkApplication.scala:52)
   	at org.apache.spark.deploy.SparkSubmit.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:845)
   	at org.apache.spark.deploy.SparkSubmit.doRunMain$1(SparkSubmit.scala:161)
   	at org.apache.spark.deploy.SparkSubmit.submit(SparkSubmit.scala:184)
   	at org.apache.spark.deploy.SparkSubmit.doSubmit(SparkSubmit.scala:86)
   	at org.apache.spark.deploy.SparkSubmit$$anon$2.doSubmit(SparkSubmit.scala:920)
   	at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:929)
   	at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
   Caused by: ERROR XJ040: Failed to start database 'metastore_db' with class loader sun.misc.Launcher$AppClassLoader@5fcfe4b2, see the next exception for details.
   	at org.apache.derby.iapi.error.StandardException.newException(Unknown Source)
   	at org.apache.derby.impl.jdbc.SQLExceptionFactory.wrapArgsForTransportAcrossDRDA(Unknown Source)
   	... 85 more
   Caused by: ERROR XSDB6: Another instance of Derby may have already booted the database /Users/nsb/Documents/personal/tools/spark-2.4.7-bin-hadoop2.7/metastore_db.
   	at org.apache.derby.iapi.error.StandardException.newException(Unknown Source)
   	at org.apache.derby.iapi.error.StandardException.newException(Unknown Source)
   	at org.apache.derby.impl.store.raw.data.BaseDataFileFactory.privGetJBMSLockOnDB(Unknown Source)
   	at org.apache.derby.impl.store.raw.data.BaseDataFileFactory.run(Unknown Source)
   	at java.security.AccessController.doPrivileged(Native Method)
   	at org.apache.derby.impl.store.raw.data.BaseDataFileFactory.getJBMSLockOnDB(Unknown Source)
   	at org.apache.derby.impl.store.raw.data.BaseDataFileFactory.boot(Unknown Source)
   	at org.apache.derby.impl.services.monitor.BaseMonitor.boot(Unknown Source)
   	at org.apache.derby.impl.services.monitor.TopService.bootModule(Unknown Source)
   	at org.apache.derby.impl.services.monitor.BaseMonitor.startModule(Unknown Source)
   	at org.apache.derby.impl.services.monitor.FileMonitor.startModule(Unknown Source)
   	at org.apache.derby.iapi.services.monitor.Monitor.bootServiceModule(Unknown Source)
   	at org.apache.derby.impl.store.raw.RawStore$6.run(Unknown Source)
   	at java.security.AccessController.doPrivileged(Native Method)
   	at org.apache.derby.impl.store.raw.RawStore.bootServiceModule(Unknown Source)
   	at org.apache.derby.impl.store.raw.RawStore.boot(Unknown Source)
   	at org.apache.derby.impl.services.monitor.BaseMonitor.boot(Unknown Source)
   	at org.apache.derby.impl.services.monitor.TopService.bootModule(Unknown Source)
   	at org.apache.derby.impl.services.monitor.BaseMonitor.startModule(Unknown Source)
   	at org.apache.derby.impl.services.monitor.FileMonitor.startModule(Unknown Source)
   	at org.apache.derby.iapi.services.monitor.Monitor.bootServiceModule(Unknown Source)
   	at org.apache.derby.impl.store.access.RAMAccessManager$5.run(Unknown Source)
   	at java.security.AccessController.doPrivileged(Native Method)
   	at org.apache.derby.impl.store.access.RAMAccessManager.bootServiceModule(Unknown Source)
   	at org.apache.derby.impl.store.access.RAMAccessManager.boot(Unknown Source)
   	at org.apache.derby.impl.services.monitor.BaseMonitor.boot(Unknown Source)
   	at org.apache.derby.impl.services.monitor.TopService.bootModule(Unknown Source)
   	at org.apache.derby.impl.services.monitor.BaseMonitor.startModule(Unknown Source)
   	at org.apache.derby.impl.services.monitor.FileMonitor.startModule(Unknown Source)
   	at org.apache.derby.iapi.services.monitor.Monitor.bootServiceModule(Unknown Source)
   	at org.apache.derby.impl.db.BasicDatabase$5.run(Unknown Source)
   	at java.security.AccessController.doPrivileged(Native Method)
   	at org.apache.derby.impl.db.BasicDatabase.bootServiceModule(Unknown Source)
   	at org.apache.derby.impl.db.BasicDatabase.bootStore(Unknown Source)
   	at org.apache.derby.impl.db.BasicDatabase.boot(Unknown Source)
   	at org.apache.derby.impl.services.monitor.BaseMonitor.boot(Unknown Source)
   	at org.apache.derby.impl.services.monitor.TopService.bootModule(Unknown Source)
   	at org.apache.derby.impl.services.monitor.BaseMonitor.bootService(Unknown Source)
   	at org.apache.derby.impl.services.monitor.BaseMonitor.startProviderService(Unknown Source)
   	at org.apache.derby.impl.services.monitor.BaseMonitor.findProviderAndStartService(Unknown Source)
   	at org.apache.derby.impl.services.monitor.BaseMonitor.startPersistentService(Unknown Source)
   	at org.apache.derby.iapi.services.monitor.Monitor.startPersistentService(Unknown Source)
   	at org.apache.derby.impl.jdbc.EmbedConnection$4.run(Unknown Source)
   	at org.apache.derby.impl.jdbc.EmbedConnection$4.run(Unknown Source)
   	at java.security.AccessController.doPrivileged(Native Method)
   	at org.apache.derby.impl.jdbc.EmbedConnection.startPersistentService(Unknown Source)
   	... 82 more
   ------
   
   org.datanucleus.exceptions.NucleusDataStoreException: Unable to open a test connection to the given database. JDBC url = jdbc:derby:;databaseName=metastore_db;create=true, username = APP. Terminating connection pool (set lazyInit to true if you expect to start your database after your app). Original Exception: ------
   java.sql.SQLException: Failed to start database 'metastore_db' with class loader sun.misc.Launcher$AppClassLoader@5fcfe4b2, see the next exception for details.
   	at org.apache.derby.impl.jdbc.SQLExceptionFactory.getSQLException(Unknown Source)
   	at org.apache.derby.impl.jdbc.SQLExceptionFactory.getSQLException(Unknown Source)
   	at org.apache.derby.impl.jdbc.Util.seeNextException(Unknown Source)
   	at org.apache.derby.impl.jdbc.EmbedConnection.bootDatabase(Unknown Source)
   	at org.apache.derby.impl.jdbc.EmbedConnection.<init>(Unknown Source)
   	at org.apache.derby.jdbc.InternalDriver$1.run(Unknown Source)
   	at org.apache.derby.jdbc.InternalDriver$1.run(Unknown Source)
   	at java.security.AccessController.doPrivileged(Native Method)
   	at org.apache.derby.jdbc.InternalDriver.getNewEmbedConnection(Unknown Source)
   	at org.apache.derby.jdbc.InternalDriver.connect(Unknown Source)
   	at org.apache.derby.jdbc.InternalDriver.connect(Unknown Source)
   	at org.apache.derby.jdbc.AutoloadedDriver.connect(Unknown Source)
   	at java.sql.DriverManager.getConnection(DriverManager.java:664)
   	at java.sql.DriverManager.getConnection(DriverManager.java:208)
   	at com.jolbox.bonecp.BoneCP.obtainRawInternalConnection(BoneCP.java:361)
   	at com.jolbox.bonecp.BoneCP.<init>(BoneCP.java:416)
   	at com.jolbox.bonecp.BoneCPDataSource.getConnection(BoneCPDataSource.java:120)
   	at org.datanucleus.store.rdbms.ConnectionFactoryImpl$ManagedConnectionImpl.getConnection(ConnectionFactoryImpl.java:501)
   	at org.datanucleus.store.rdbms.RDBMSStoreManager.<init>(RDBMSStoreManager.java:298)
   	at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
   	at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
   	at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
   	at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
   	at org.datanucleus.plugin.NonManagedPluginRegistry.createExecutableExtension(NonManagedPluginRegistry.java:631)
   	at org.datanucleus.plugin.PluginManager.createExecutableExtension(PluginManager.java:301)
   	at org.datanucleus.NucleusContext.createStoreManagerForProperties(NucleusContext.java:1187)
   	at org.datanucleus.NucleusContext.initialise(NucleusContext.java:356)
   	at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.freezeConfiguration(JDOPersistenceManagerFactory.java:775)
   	at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.createPersistenceManagerFactory(JDOPersistenceManagerFactory.java:333)
   	at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.getPersistenceManagerFactory(JDOPersistenceManagerFactory.java:202)
   	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
   	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
   	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
   	at java.lang.reflect.Method.invoke(Method.java:498)
   	at javax.jdo.JDOHelper$16.run(JDOHelper.java:1965)
   	at java.security.AccessController.doPrivileged(Native Method)
   	at javax.jdo.JDOHelper.invoke(JDOHelper.java:1960)
   	at javax.jdo.JDOHelper.invokeGetPersistenceManagerFactoryOnImplementation(JDOHelper.java:1166)
   	at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:808)
   	at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:701)
   	at org.apache.hadoop.hive.metastore.ObjectStore.getPMF(ObjectStore.java:365)
   	at org.apache.hadoop.hive.metastore.ObjectStore.getPersistenceManager(ObjectStore.java:394)
   	at org.apache.hadoop.hive.metastore.ObjectStore.initialize(ObjectStore.java:291)
   	at org.apache.hadoop.hive.metastore.ObjectStore.setConf(ObjectStore.java:258)
   	at org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:76)
   	at org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:136)
   	at org.apache.hadoop.hive.metastore.RawStoreProxy.<init>(RawStoreProxy.java:57)
   	at org.apache.hadoop.hive.metastore.RawStoreProxy.getProxy(RawStoreProxy.java:66)
   	at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.newRawStore(HiveMetaStore.java:593)
   	at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMS(HiveMetaStore.java:571)
   	at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.createDefaultDB(HiveMetaStore.java:620)
   	at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.init(HiveMetaStore.java:461)
   	at org.apache.hadoop.hive.metastore.RetryingHMSHandler.<init>(RetryingHMSHandler.java:66)
   	at org.apache.hadoop.hive.metastore.RetryingHMSHandler.getProxy(RetryingHMSHandler.java:72)
   	at org.apache.hadoop.hive.metastore.HiveMetaStore.newRetryingHMSHandler(HiveMetaStore.java:5762)
   	at org.apache.hadoop.hive.metastore.HiveMetaStoreClient.<init>(HiveMetaStoreClient.java:199)
   	at org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient.<init>(SessionHiveMetaStoreClient.java:74)
   	at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
   	at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
   	at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
   	at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
   	at org.apache.hadoop.hive.metastore.MetaStoreUtils.newInstance(MetaStoreUtils.java:1521)
   	at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.<init>(RetryingMetaStoreClient.java:86)
   	at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:132)
   	at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:104)
   	at org.apache.hadoop.hive.ql.metadata.Hive.createMetaStoreClient(Hive.java:3005)
   	at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3024)
   	at org.apache.hadoop.hive.ql.metadata.Hive.getAllDatabases(Hive.java:1234)
   	at org.apache.hadoop.hive.ql.metadata.Hive.reloadFunctions(Hive.java:174)
   	at org.apache.hadoop.hive.ql.metadata.Hive.<clinit>(Hive.java:166)
   	at org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.java:503)
   	at org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver$.main(SparkSQLCLIDriver.scala:133)
   	at org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver.main(SparkSQLCLIDriver.scala)
   	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
   	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
   	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
   	at java.lang.reflect.Method.invoke(Method.java:498)
   	at org.apache.spark.deploy.JavaMainApplication.start(SparkApplication.scala:52)
   	at org.apache.spark.deploy.SparkSubmit.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:845)
   	at org.apache.spark.deploy.SparkSubmit.doRunMain$1(SparkSubmit.scala:161)
   	at org.apache.spark.deploy.SparkSubmit.submit(SparkSubmit.scala:184)
   	at org.apache.spark.deploy.SparkSubmit.doSubmit(SparkSubmit.scala:86)
   	at org.apache.spark.deploy.SparkSubmit$$anon$2.doSubmit(SparkSubmit.scala:920)
   	at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:929)
   	at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
   Caused by: ERROR XJ040: Failed to start database 'metastore_db' with class loader sun.misc.Launcher$AppClassLoader@5fcfe4b2, see the next exception for details.
   	at org.apache.derby.iapi.error.StandardException.newException(Unknown Source)
   	at org.apache.derby.impl.jdbc.SQLExceptionFactory.wrapArgsForTransportAcrossDRDA(Unknown Source)
   	... 85 more
   Caused by: ERROR XSDB6: Another instance of Derby may have already booted the database /Users/nsb/Documents/personal/tools/spark-2.4.7-bin-hadoop2.7/metastore_db.
   ```
   
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: commits-unsubscribe@hudi.apache.org

For queries about this service, please contact Infrastructure at:
users@infra.apache.org



[GitHub] [hudi] nsivabalan commented on issue #3990: [SUPPORT] Spark-sql fails to start if there is a concurrent spark-shell open

Posted by GitBox <gi...@apache.org>.
nsivabalan commented on issue #3990:
URL: https://github.com/apache/hudi/issues/3990#issuecomment-968861089


   I tried passing in `--conf "spark.sql.warehouse.dir=/tmp/spark-warehouse/"` but did not help. Only solution I found online is to delete the `metastore_db/dbex.lck` and then try launching. It worked for me. 
   
   In general, it is proposed to have only one spark-shell or spark-sql per instance(client). 
   
   Sources:
   https://stackoverflow.com/questions/34465516/caused-by-error-xsdb6-another-instance-of-derby-may-have-already-booted-the-da
   https://stackoverflow.com/questions/37442910/spark-shell-startup-errors
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: commits-unsubscribe@hudi.apache.org

For queries about this service, please contact Infrastructure at:
users@infra.apache.org



[GitHub] [hudi] nsivabalan closed issue #3990: [SUPPORT] Spark-sql fails to start if there is a concurrent spark-shell open

Posted by GitBox <gi...@apache.org>.
nsivabalan closed issue #3990:
URL: https://github.com/apache/hudi/issues/3990


   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: commits-unsubscribe@hudi.apache.org

For queries about this service, please contact Infrastructure at:
users@infra.apache.org