You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@dolphinscheduler.apache.org by GitBox <gi...@apache.org> on 2020/09/03 03:06:54 UTC

[GitHub] [incubator-dolphinscheduler] AfterYesterday opened a new issue #3654: [Question] 通过shell提交spark-submit报错:ERROR XBM0H: Directory /tmp/dolphinscheduler/exec/process/1/3/3/4/metastore_db cannot be created

AfterYesterday opened a new issue #3654:
URL: https://github.com/apache/incubator-dolphinscheduler/issues/3654


   [INFO] 2020-09-03 10:57:31.411  - [taskAppId=TASK-3-3-7]:[121] -  -> Warning: Master yarn-client is deprecated since 2.0. Please use master "yarn" with specified deploy mode instead.
   	20/09/03 10:57:31 INFO spark.SparkContext: Running Spark version 2.4.0.cloudera2
   [INFO] 2020-09-03 10:57:33.234  - [taskAppId=TASK-3-3-7]:[121] -  -> 20/09/03 10:57:31 INFO spark.SparkContext: Submitted application: sparkSqlTest
   	20/09/03 10:57:31 INFO spark.SecurityManager: Changing view acls to: hdfs
   	20/09/03 10:57:31 INFO spark.SecurityManager: Changing modify acls to: hdfs
   	20/09/03 10:57:31 INFO spark.SecurityManager: Changing view acls groups to: 
   	20/09/03 10:57:31 INFO spark.SecurityManager: Changing modify acls groups to: 
   	20/09/03 10:57:31 INFO spark.SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users  with view permissions: Set(hdfs); groups with view permissions: Set(); users  with modify permissions: Set(hdfs); groups with modify permissions: Set()
   	20/09/03 10:57:31 INFO util.Utils: max retries is 16
   	20/09/03 10:57:31 INFO util.Utils: Successfully started service 'sparkDriver' on port 35977.
   	20/09/03 10:57:31 INFO spark.SparkEnv: Registering MapOutputTracker
   	20/09/03 10:57:31 INFO spark.SparkEnv: Registering BlockManagerMaster
   	20/09/03 10:57:31 INFO storage.BlockManagerMasterEndpoint: Using org.apache.spark.storage.DefaultTopologyMapper for getting topology information
   	20/09/03 10:57:31 INFO storage.BlockManagerMasterEndpoint: BlockManagerMasterEndpoint up
   	20/09/03 10:57:31 INFO storage.DiskBlockManager: Created local directory at /tmp/blockmgr-c87ef2c5-a62b-4e48-a849-d70660e4d66e
   	20/09/03 10:57:31 INFO memory.MemoryStore: MemoryStore started with capacity 366.3 MB
   	20/09/03 10:57:31 INFO spark.SparkEnv: Registering OutputCommitCoordinator
   	20/09/03 10:57:31 INFO util.log: Logging initialized @2045ms
   	20/09/03 10:57:32 INFO server.Server: jetty-9.3.z-SNAPSHOT, build timestamp: unknown, git hash: unknown
   	20/09/03 10:57:32 INFO server.Server: Started @2119ms
   	20/09/03 10:57:32 INFO util.Utils: max retries is 16
   	20/09/03 10:57:32 INFO server.AbstractConnector: Started ServerConnector@7f4d9395{HTTP/1.1,[http/1.1]}{0.0.0.0:4040}
   	20/09/03 10:57:32 INFO util.Utils: Successfully started service 'SparkUI' on port 4040.
   	20/09/03 10:57:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@3003697{/jobs,null,AVAILABLE,@Spark}
   	20/09/03 10:57:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@2aa27288{/jobs/json,null,AVAILABLE,@Spark}
   	20/09/03 10:57:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@7f34a967{/jobs/job,null,AVAILABLE,@Spark}
   	20/09/03 10:57:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@240139e1{/jobs/job/json,null,AVAILABLE,@Spark}
   	20/09/03 10:57:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@7ea4d397{/stages,null,AVAILABLE,@Spark}
   	20/09/03 10:57:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@49298ce7{/stages/json,null,AVAILABLE,@Spark}
   	20/09/03 10:57:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@253c1256{/stages/stage,null,AVAILABLE,@Spark}
   	20/09/03 10:57:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@55f45b92{/stages/stage/json,null,AVAILABLE,@Spark}
   	20/09/03 10:57:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@109f5dd8{/stages/pool,null,AVAILABLE,@Spark}
   	20/09/03 10:57:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@67fe380b{/stages/pool/json,null,AVAILABLE,@Spark}
   	20/09/03 10:57:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@4a325eb9{/storage,null,AVAILABLE,@Spark}
   	20/09/03 10:57:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@3dedb4a6{/storage/json,null,AVAILABLE,@Spark}
   	20/09/03 10:57:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@57f64f5e{/storage/rdd,null,AVAILABLE,@Spark}
   	20/09/03 10:57:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@415e0bcb{/storage/rdd/json,null,AVAILABLE,@Spark}
   	20/09/03 10:57:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@194152cf{/environment,null,AVAILABLE,@Spark}
   	20/09/03 10:57:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@49d98dc5{/environment/json,null,AVAILABLE,@Spark}
   	20/09/03 10:57:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@2c30b71f{/executors,null,AVAILABLE,@Spark}
   	20/09/03 10:57:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@1d81e101{/executors/json,null,AVAILABLE,@Spark}
   	20/09/03 10:57:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@ec50f54{/executors/threadDump,null,AVAILABLE,@Spark}
   	20/09/03 10:57:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@bf71cec{/executors/threadDump/json,null,AVAILABLE,@Spark}
   	20/09/03 10:57:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@22d6cac2{/static,null,AVAILABLE,@Spark}
   	20/09/03 10:57:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@6d868997{/,null,AVAILABLE,@Spark}
   	20/09/03 10:57:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@2c383e33{/api,null,AVAILABLE,@Spark}
   	20/09/03 10:57:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@4287d447{/jobs/job/kill,null,AVAILABLE,@Spark}
   	20/09/03 10:57:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@3af37506{/stages/stage/kill,null,AVAILABLE,@Spark}
   	20/09/03 10:57:32 INFO ui.SparkUI: Bound SparkUI to 0.0.0.0, and started at http://dsjzt-hadoop-2:4040
   	20/09/03 10:57:32 INFO spark.SparkContext: Added JAR file:/sparkjob/scalatest-1.0-SNAPSHOT.jar at spark://dsjzt-hadoop-2:35977/jars/scalatest-1.0-SNAPSHOT.jar with timestamp 1599101852188
   	20/09/03 10:57:32 INFO util.Utils: Using initial executors = 2, max of spark.dynamicAllocation.initialExecutors, spark.dynamicAllocation.minExecutors and spark.executor.instances
   	20/09/03 10:57:33 INFO client.RMProxy: Connecting to ResourceManager at dsjzt-hadoop-1/10.7.11.9:8032
   [INFO] 2020-09-03 10:57:35.226  - [taskAppId=TASK-3-3-7]:[121] -  -> 20/09/03 10:57:33 INFO yarn.Client: Requesting a new application from cluster with 3 NodeManagers
   	20/09/03 10:57:33 INFO yarn.Client: Verifying our application has not requested more than the maximum memory capability of the cluster (48119 MB per container)
   	20/09/03 10:57:33 INFO yarn.Client: Will allocate AM container, with 896 MB memory including 384 MB overhead
   	20/09/03 10:57:33 INFO yarn.Client: Setting up container launch context for our AM
   	20/09/03 10:57:33 INFO yarn.Client: Setting up the launch environment for our AM container
   	20/09/03 10:57:33 INFO yarn.Client: Preparing resources for our AM container
   	20/09/03 10:57:33 INFO yarn.Client: Uploading resource file:/tmp/spark-60cfad97-9e20-460f-b947-cce3b9092bb2/__spark_conf__1905511865034791586.zip -> hdfs://dsjzt-db:8020/user/hdfs/.sparkStaging/application_1598319282181_0237/__spark_conf__.zip
   	20/09/03 10:57:34 INFO spark.SecurityManager: Changing view acls to: hdfs
   	20/09/03 10:57:34 INFO spark.SecurityManager: Changing modify acls to: hdfs
   	20/09/03 10:57:34 INFO spark.SecurityManager: Changing view acls groups to: 
   	20/09/03 10:57:34 INFO spark.SecurityManager: Changing modify acls groups to: 
   	20/09/03 10:57:34 INFO spark.SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users  with view permissions: Set(hdfs); groups with view permissions: Set(); users  with modify permissions: Set(hdfs); groups with modify permissions: Set()
   	20/09/03 10:57:35 INFO yarn.Client: Submitting application application_1598319282181_0237 to ResourceManager
   [INFO] 2020-09-03 10:57:36.272  - [taskAppId=TASK-3-3-7]:[121] -  -> 20/09/03 10:57:35 INFO impl.YarnClientImpl: Submitted application application_1598319282181_0237
   	20/09/03 10:57:35 INFO cluster.SchedulerExtensionServices: Starting Yarn extension services with app application_1598319282181_0237 and attemptId None
   	20/09/03 10:57:35 INFO yarn.SparkRackResolver: Got an error when resolving hostNames. Falling back to /default-rack for all
   	20/09/03 10:57:36 INFO yarn.SparkRackResolver: Got an error when resolving hostNames. Falling back to /default-rack for all
   [INFO] 2020-09-03 10:57:37.278  - [taskAppId=TASK-3-3-7]:[121] -  -> 20/09/03 10:57:36 INFO yarn.Client: Application report for application_1598319282181_0237 (state: ACCEPTED)
   	20/09/03 10:57:36 INFO yarn.Client: 
   		 client token: N/A
   		 diagnostics: N/A
   		 ApplicationMaster host: N/A
   		 ApplicationMaster RPC port: -1
   		 queue: root.users.hdfs
   		 start time: 1599101850311
   		 final status: UNDEFINED
   		 tracking URL: http://dsjzt-hadoop-1:8088/proxy/application_1598319282181_0237/
   		 user: hdfs
   	20/09/03 10:57:37 INFO yarn.SparkRackResolver: Got an error when resolving hostNames. Falling back to /default-rack for all
   	20/09/03 10:57:37 INFO yarn.Client: Application report for application_1598319282181_0237 (state: ACCEPTED)
   [INFO] 2020-09-03 10:57:38.280  - [taskAppId=TASK-3-3-7]:[121] -  -> 20/09/03 10:57:38 INFO yarn.SparkRackResolver: Got an error when resolving hostNames. Falling back to /default-rack for all
   	20/09/03 10:57:38 INFO yarn.Client: Application report for application_1598319282181_0237 (state: ACCEPTED)
   [INFO] 2020-09-03 10:57:39.282  - [taskAppId=TASK-3-3-7]:[121] -  -> 20/09/03 10:57:38 INFO cluster.YarnClientSchedulerBackend: Add WebUI Filter. org.apache.hadoop.yarn.server.webproxy.amfilter.AmIpFilter, Map(PROXY_HOSTS -> dsjzt-hadoop-1, PROXY_URI_BASES -> http://dsjzt-hadoop-1:8088/proxy/application_1598319282181_0237), /proxy/application_1598319282181_0237
   	20/09/03 10:57:38 INFO ui.JettyUtils: Adding filter org.apache.hadoop.yarn.server.webproxy.amfilter.AmIpFilter to /jobs, /jobs/json, /jobs/job, /jobs/job/json, /stages, /stages/json, /stages/stage, /stages/stage/json, /stages/pool, /stages/pool/json, /storage, /storage/json, /storage/rdd, /storage/rdd/json, /environment, /environment/json, /executors, /executors/json, /executors/threadDump, /executors/threadDump/json, /static, /, /api, /jobs/job/kill, /stages/stage/kill.
   	20/09/03 10:57:39 INFO cluster.YarnSchedulerBackend$YarnSchedulerEndpoint: ApplicationMaster registered as NettyRpcEndpointRef(spark-client://YarnAM)
   	20/09/03 10:57:39 INFO yarn.SparkRackResolver: Got an error when resolving hostNames. Falling back to /default-rack for all
   	20/09/03 10:57:39 INFO yarn.Client: Application report for application_1598319282181_0237 (state: RUNNING)
   [INFO] 2020-09-03 10:57:41.272  - [taskAppId=TASK-3-3-7]:[121] -  -> 20/09/03 10:57:39 INFO yarn.Client: 
   		 client token: N/A
   		 diagnostics: N/A
   		 ApplicationMaster host: 10.7.11.9
   		 ApplicationMaster RPC port: -1
   		 queue: root.users.hdfs
   		 start time: 1599101850311
   		 final status: UNDEFINED
   		 tracking URL: http://dsjzt-hadoop-1:8088/proxy/application_1598319282181_0237/
   		 user: hdfs
   	20/09/03 10:57:39 INFO cluster.YarnClientSchedulerBackend: Application application_1598319282181_0237 has started running.
   	20/09/03 10:57:39 INFO util.Utils: max retries is 16
   	20/09/03 10:57:39 INFO util.Utils: Successfully started service 'org.apache.spark.network.netty.NettyBlockTransferService' on port 40754.
   	20/09/03 10:57:39 INFO netty.NettyBlockTransferService: Server created on dsjzt-hadoop-2:40754
   	20/09/03 10:57:39 INFO storage.BlockManager: Using org.apache.spark.storage.RandomBlockReplicationPolicy for block replication policy
   	20/09/03 10:57:39 INFO storage.BlockManagerMaster: Registering BlockManager BlockManagerId(driver, dsjzt-hadoop-2, 40754, None)
   	20/09/03 10:57:39 INFO storage.BlockManagerMasterEndpoint: Registering block manager dsjzt-hadoop-2:40754 with 366.3 MB RAM, BlockManagerId(driver, dsjzt-hadoop-2, 40754, None)
   	20/09/03 10:57:39 INFO storage.BlockManagerMaster: Registered BlockManager BlockManagerId(driver, dsjzt-hadoop-2, 40754, None)
   	20/09/03 10:57:39 INFO storage.BlockManager: external shuffle service port = 7337
   	20/09/03 10:57:39 INFO storage.BlockManager: Initialized BlockManager: BlockManagerId(driver, dsjzt-hadoop-2, 40754, None)
   	20/09/03 10:57:39 INFO ui.JettyUtils: Adding filter org.apache.hadoop.yarn.server.webproxy.amfilter.AmIpFilter to /metrics/json.
   	20/09/03 10:57:39 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@191ec193{/metrics/json,null,AVAILABLE,@Spark}
   	20/09/03 10:57:39 INFO scheduler.EventLoggingListener: Logging events to hdfs://dsjzt-db:8020/user/spark/spark2ApplicationHistory/application_1598319282181_0237
   	20/09/03 10:57:39 INFO util.Utils: Using initial executors = 2, max of spark.dynamicAllocation.initialExecutors, spark.dynamicAllocation.minExecutors and spark.executor.instances
   	20/09/03 10:57:39 WARN lineage.LineageWriter: Lineage directory /var/log/spark2/lineage doesn't exist or is not writable. Lineage for this application will be disabled.
   	20/09/03 10:57:39 INFO util.Utils: Extension com.cloudera.spark.lineage.NavigatorAppListener not being initialized.
   	20/09/03 10:57:40 INFO yarn.SparkRackResolver: Got an error when resolving hostNames. Falling back to /default-rack for all
   	20/09/03 10:57:41 INFO yarn.SparkRackResolver: Got an error when resolving hostNames. Falling back to /default-rack for all
   [INFO] 2020-09-03 10:57:42.491  - [taskAppId=TASK-3-3-7]:[121] -  -> 20/09/03 10:57:42 INFO yarn.SparkRackResolver: Got an error when resolving hostNames. Falling back to /default-rack for all
   	20/09/03 10:57:42 INFO cluster.YarnSchedulerBackend$YarnDriverEndpoint: Registered executor NettyRpcEndpointRef(spark-client://Executor) (10.7.11.9:34350) with ID 1
   [INFO] 2020-09-03 10:57:45.199  - [taskAppId=TASK-3-3-7]:[121] -  -> 20/09/03 10:57:42 INFO spark.ExecutorAllocationManager: New executor 1 has registered (new total is 1)
   	20/09/03 10:57:42 INFO cluster.YarnSchedulerBackend$YarnDriverEndpoint: Registered executor NettyRpcEndpointRef(spark-client://Executor) (10.7.11.9:34352) with ID 2
   	20/09/03 10:57:42 INFO spark.ExecutorAllocationManager: New executor 2 has registered (new total is 2)
   	20/09/03 10:57:42 INFO storage.BlockManagerMasterEndpoint: Registering block manager dsjzt-hadoop-1:37298 with 5.2 GB RAM, BlockManagerId(1, dsjzt-hadoop-1, 37298, None)
   	20/09/03 10:57:42 INFO cluster.YarnClientSchedulerBackend: SchedulerBackend is ready for scheduling beginning after reached minRegisteredResourcesRatio: 0.8
   	20/09/03 10:57:43 WARN lineage.LineageWriter: Lineage directory /var/log/spark2/lineage doesn't exist or is not writable. Lineage for this application will be disabled.
   	Thu Sep 03 10:57:45 CST 2020 Thread[main,5,main] java.io.FileNotFoundException: derby.log (Permission denied)
   [INFO] 2020-09-03 10:57:45.330  - [taskAppId=TASK-3-3-7]:[121] -  -> Thu Sep 03 10:57:45 CST 2020 Thread[main,5,main] Cleanup action starting
   	ERROR XBM0H: Directory /tmp/dolphinscheduler/exec/process/1/3/3/7/metastore_db cannot be created.
   		at org.apache.derby.iapi.error.StandardException.newException(Unknown Source)
   		at org.apache.derby.iapi.error.StandardException.newException(Unknown Source)
   		at org.apache.derby.impl.services.monitor.StorageFactoryService$10.run(Unknown Source)
   		at java.security.AccessController.doPrivileged(Native Method)
   		at org.apache.derby.impl.services.monitor.StorageFactoryService.createServiceRoot(Unknown Source)
   		at org.apache.derby.impl.services.monitor.BaseMonitor.bootService(Unknown Source)
   		at org.apache.derby.impl.services.monitor.BaseMonitor.createPersistentService(Unknown Source)
   		at org.apache.derby.impl.services.monitor.FileMonitor.createPersistentService(Unknown Source)
   		at org.apache.derby.iapi.services.monitor.Monitor.createPersistentService(Unknown Source)
   		at org.apache.derby.impl.jdbc.EmbedConnection.createDatabase(Unknown Source)
   		at org.apache.derby.impl.jdbc.EmbedConnection.<init>(Unknown Source)
   		at org.apache.derby.jdbc.InternalDriver.getNewEmbedConnection(Unknown Source)
   		at org.apache.derby.jdbc.InternalDriver.connect(Unknown Source)
   		at org.apache.derby.jdbc.InternalDriver.connect(Unknown Source)
   		at org.apache.derby.jdbc.AutoloadedDriver.connect(Unknown Source)
   		at java.sql.DriverManager.getConnection(DriverManager.java:664)
   		at java.sql.DriverManager.getConnection(DriverManager.java:208)
   		at com.jolbox.bonecp.BoneCP.obtainRawInternalConnection(BoneCP.java:349)
   		at com.jolbox.bonecp.BoneCP.<init>(BoneCP.java:416)
   		at com.jolbox.bonecp.BoneCPDataSource.getConnection(BoneCPDataSource.java:120)
   		at org.datanucleus.store.rdbms.ConnectionFactoryImpl$ManagedConnectionImpl.getConnection(ConnectionFactoryImpl.java:501)
   		at org.datanucleus.store.rdbms.RDBMSStoreManager.<init>(RDBMSStoreManager.java:298)
   		at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
   		at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
   		at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
   		at java.lang.reflect.Constructor.newInstance(Constructor.java:408)
   		at org.datanucleus.plugin.NonManagedPluginRegistry.createExecutableExtension(NonManagedPluginRegistry.java:631)
   		at org.datanucleus.plugin.PluginManager.createExecutableExtension(PluginManager.java:301)
   		at org.datanucleus.NucleusContext.createStoreManagerForProperties(NucleusContext.java:1187)
   		at org.datanucleus.NucleusContext.initialise(NucleusContext.java:356)
   		at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.freezeConfiguration(JDOPersistenceManagerFactory.java:775)
   		at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.createPersistenceManagerFactory(JDOPersistenceManagerFactory.java:333)
   		at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.getPersistenceManagerFactory(JDOPersistenceManagerFactory.java:202)
   		at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
   		at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
   		at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
   		at java.lang.reflect.Method.invoke(Method.java:483)
   		at javax.jdo.JDOHelper$16.run(JDOHelper.java:1965)
   		at java.security.AccessController.doPrivileged(Native Method)
   		at javax.jdo.JDOHelper.invoke(JDOHelper.java:1960)
   		at javax.jdo.JDOHelper.invokeGetPersistenceManagerFactoryOnImplementation(JDOHelper.java:1166)
   		at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:808)
   		at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:701)
   		at org.apache.hadoop.hive.metastore.ObjectStore.getPMF(ObjectStore.java:419)
   		at org.apache.hadoop.hive.metastore.ObjectStore.getPersistenceManager(ObjectStore.java:448)
   		at org.apache.hadoop.hive.metastore.ObjectStore.initialize(ObjectStore.java:343)
   		at org.apache.hadoop.hive.metastore.ObjectStore.setConf(ObjectStore.java:299)
   		at org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:73)
   		at org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:133)
   		at org.apache.hadoop.hive.metastore.RawStoreProxy.<init>(RawStoreProxy.java:60)
   		at org.apache.hadoop.hive.metastore.RawStoreProxy.getProxy(RawStoreProxy.java:69)
   		at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.newRawStore(HiveMetaStore.java:685)
   		at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMS(HiveMetaStore.java:663)
   		at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.createDefaultDB(HiveMetaStore.java:712)
   		at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.init(HiveMetaStore.java:511)
   		at org.apache.hadoop.hive.metastore.RetryingHMSHandler.<init>(RetryingHMSHandler.java:78)
   		at org.apache.hadoop.hive.metastore.RetryingHMSHandler.getProxy(RetryingHMSHandler.java:84)
   		at org.apache.hadoop.hive.metastore.HiveMetaStore.newRetryingHMSHandler(HiveMetaStore.java:6517)
   		at org.apache.hadoop.hive.metastore.HiveMetaStoreClient.<init>(HiveMetaStoreClient.java:207)
   		at org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient.<init>(SessionHiveMetaStoreClient.java:74)
   		at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
   		at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
   [INFO] 2020-09-03 10:57:45.331  - [taskAppId=TASK-3-3-7]:[121] -  -> 	at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
   		at java.lang.reflect.Constructor.newInstance(Constructor.java:408)
   		at org.apache.hadoop.hive.metastore.MetaStoreUtils.newInstance(MetaStoreUtils.java:1660)
   		at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.<init>(RetryingMetaStoreClient.java:67)
   		at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:82)
   		at org.apache.hadoop.hive.ql.metadata.Hive.createMetaStoreClient(Hive.java:3411)
   		at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3430)
   		at org.apache.hadoop.hive.ql.metadata.Hive.getAllFunctions(Hive.java:3655)
   		at org.apache.hadoop.hive.ql.metadata.Hive.reloadFunctions(Hive.java:231)
   		at org.apache.hadoop.hive.ql.metadata.Hive.registerAllFunctionsOnce(Hive.java:215)
   		at org.apache.hadoop.hive.ql.metadata.Hive.<init>(Hive.java:338)
   		at org.apache.hadoop.hive.ql.metadata.Hive.get(Hive.java:299)
   		at org.apache.hadoop.hive.ql.metadata.Hive.get(Hive.java:274)
   		at org.apache.spark.sql.hive.client.HiveClientImpl.org$apache$spark$sql$hive$client$HiveClientImpl$$client(HiveClientImpl.scala:246)
   		at org.apache.spark.sql.hive.client.HiveClientImpl$$anonfun$withHiveState$1.apply(HiveClientImpl.scala:268)
   		at org.apache.spark.sql.hive.client.HiveClientImpl.liftedTree1$1(HiveClientImpl.scala:213)
   		at org.apache.spark.sql.hive.client.HiveClientImpl.retryLocked(HiveClientImpl.scala:212)
   		at org.apache.spark.sql.hive.client.HiveClientImpl.withHiveState(HiveClientImpl.scala:258)
   		at org.apache.spark.sql.hive.client.HiveClientImpl.databaseExists(HiveClientImpl.scala:348)
   		at org.apache.spark.sql.hive.HiveExternalCatalog$$anonfun$databaseExists$1.apply$mcZ$sp(HiveExternalCatalog.scala:217)
   		at org.apache.spark.sql.hive.HiveExternalCatalog$$anonfun$databaseExists$1.apply(HiveExternalCatalog.scala:217)
   		at org.apache.spark.sql.hive.HiveExternalCatalog$$anonfun$databaseExists$1.apply(HiveExternalCatalog.scala:217)
   		at org.apache.spark.sql.hive.HiveExternalCatalog.withClient(HiveExternalCatalog.scala:99)
   		at org.apache.spark.sql.hive.HiveExternalCatalog.databaseExists(HiveExternalCatalog.scala:216)
   		at org.apache.spark.sql.internal.SharedState.externalCatalog$lzycompute(SharedState.scala:114)
   		at org.apache.spark.sql.internal.SharedState.externalCatalog(SharedState.scala:102)
   		at org.apache.spark.sql.internal.SharedState.globalTempViewManager$lzycompute(SharedState.scala:141)
   		at org.apache.spark.sql.internal.SharedState.globalTempViewManager(SharedState.scala:136)
   		at org.apache.spark.sql.hive.HiveSessionStateBuilder$$anonfun$2.apply(HiveSessionStateBuilder.scala:55)
   		at org.apache.spark.sql.hive.HiveSessionStateBuilder$$anonfun$2.apply(HiveSessionStateBuilder.scala:55)
   		at org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager$lzycompute(SessionCatalog.scala:91)
   		at org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager(SessionCatalog.scala:91)
   		at org.apache.spark.sql.catalyst.catalog.SessionCatalog.isTemporaryTable(SessionCatalog.scala:736)
   		at org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$.isRunningDirectlyOnFiles(Analyzer.scala:749)
   		at org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$.resolveRelation(Analyzer.scala:683)
   		at org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$$anonfun$apply$8.applyOrElse(Analyzer.scala:715)
   		at org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$$anonfun$apply$8.applyOrElse(Analyzer.scala:708)
   		at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1$$anonfun$apply$1.apply(AnalysisHelper.scala:90)
   		at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1$$anonfun$apply$1.apply(AnalysisHelper.scala:90)
   		at org.apache.spark.sql.catalyst.trees.CurrentOrigin$.withOrigin(TreeNode.scala:70)
   		at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1.apply(AnalysisHelper.scala:89)
   		at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1.apply(AnalysisHelper.scala:86)
   		at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$.allowInvokingTransformsInAnalyzer(AnalysisHelper.scala:194)
   		at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$class.resolveOperatorsUp(AnalysisHelper.scala:86)
   		at org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.resolveOperatorsUp(LogicalPlan.scala:29)
   		at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1$$anonfun$1.apply(AnalysisHelper.scala:87)
   		at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1$$anonfun$1.apply(AnalysisHelper.scala:87)
   		at org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$4.apply(TreeNode.scala:326)
   		at org.apache.spark.sql.catalyst.trees.TreeNode.mapProductIterator(TreeNode.scala:187)
   		at org.apache.spark.sql.catalyst.trees.TreeNode.mapChildren(TreeNode.scala:324)
   		at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1.apply(AnalysisHelper.scala:87)
   		at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1.apply(AnalysisHelper.scala:86)
   		at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$.allowInvokingTransformsInAnalyzer(AnalysisHelper.scala:194)
   		at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$class.resolveOperatorsUp(AnalysisHelper.scala:86)
   		at org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.resolveOperatorsUp(LogicalPlan.scala:29)
   		at org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$.apply(Analyzer.scala:708)
   		at org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$.apply(Analyzer.scala:654)
   		at org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1$$anonfun$apply$1.apply(RuleExecutor.scala:87)
   		at org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1$$anonfun$apply$1.apply(RuleExecutor.scala:84)
   		at scala.collection.LinearSeqOptimized$class.foldLeft(LinearSeqOptimized.scala:124)
   		at scala.collection.immutable.List.foldLeft(List.scala:84)
   		at org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1.apply(RuleExecutor.scala:84)
   		at org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1.apply(RuleExecutor.scala:76)
   		at scala.collection.immutable.List.foreach(List.scala:392)
   [INFO] 2020-09-03 10:57:45.335  - [taskAppId=TASK-3-3-7]:[121] -  -> 	at org.apache.spark.sql.catalyst.rules.RuleExecutor.execute(RuleExecutor.scala:76)
   		at org.apache.spark.sql.catalyst.analysis.Analyzer.org$apache$spark$sql$catalyst$analysis$Analyzer$$executeSameContext(Analyzer.scala:127)
   		at org.apache.spark.sql.catalyst.analysis.Analyzer.execute(Analyzer.scala:121)
   		at org.apache.spark.sql.catalyst.analysis.Analyzer$$anonfun$executeAndCheck$1.apply(Analyzer.scala:106)
   		at org.apache.spark.sql.catalyst.analysis.Analyzer$$anonfun$executeAndCheck$1.apply(Analyzer.scala:105)
   		at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$.markInAnalyzer(AnalysisHelper.scala:201)
   		at org.apache.spark.sql.catalyst.analysis.Analyzer.executeAndCheck(Analyzer.scala:105)
   		at org.apache.spark.sql.execution.QueryExecution.analyzed$lzycompute(QueryExecution.scala:57)
   		at org.apache.spark.sql.execution.QueryExecution.analyzed(QueryExecution.scala:55)
   		at org.apache.spark.sql.execution.QueryExecution.assertAnalyzed(QueryExecution.scala:47)
   		at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:78)
   		at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:642)
   		at dsjzt.sparkSqlTest$.main(sparkSqlTest.scala:14)
   		at dsjzt.sparkSqlTest.main(sparkSqlTest.scala)
   		at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
   		at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
   		at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
   		at java.lang.reflect.Method.invoke(Method.java:483)
   		at org.apache.spark.deploy.JavaMainApplication.start(SparkApplication.scala:52)
   		at org.apache.spark.deploy.SparkSubmit.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:849)
   		at org.apache.spark.deploy.SparkSubmit.doRunMain$1(SparkSubmit.scala:167)
   		at org.apache.spark.deploy.SparkSubmit.submit(SparkSubmit.scala:195)
   		at org.apache.spark.deploy.SparkSubmit.doSubmit(SparkSubmit.scala:86)
   		at org.apache.spark.deploy.SparkSubmit$$anon$2.doSubmit(SparkSubmit.scala:924)
   		at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:933)
   		at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
   	Cleanup action completed
   	Thu Sep 03 10:57:45 CST 2020 Thread[main,5,main] Cleanup action starting
   	java.sql.SQLException: Failed to create database 'metastore_db', see the next exception for details.
   		at org.apache.derby.impl.jdbc.SQLExceptionFactory.getSQLException(Unknown Source)
   		at org.apache.derby.impl.jdbc.SQLExceptionFactory.getSQLException(Unknown Source)
   		at org.apache.derby.impl.jdbc.Util.seeNextException(Unknown Source)
   		at org.apache.derby.impl.jdbc.EmbedConnection.createDatabase(Unknown Source)
   		at org.apache.derby.impl.jdbc.EmbedConnection.<init>(Unknown Source)
   		at org.apache.derby.jdbc.InternalDriver.getNewEmbedConnection(Unknown Source)
   		at org.apache.derby.jdbc.InternalDriver.connect(Unknown Source)
   		at org.apache.derby.jdbc.InternalDriver.connect(Unknown Source)
   		at org.apache.derby.jdbc.AutoloadedDriver.connect(Unknown Source)
   		at java.sql.DriverManager.getConnection(DriverManager.java:664)
   		at java.sql.DriverManager.getConnection(DriverManager.java:208)
   		at com.jolbox.bonecp.BoneCP.obtainRawInternalConnection(BoneCP.java:349)
   		at com.jolbox.bonecp.BoneCP.<init>(BoneCP.java:416)
   		at com.jolbox.bonecp.BoneCPDataSource.getConnection(BoneCPDataSource.java:120)
   		at org.datanucleus.store.rdbms.ConnectionFactoryImpl$ManagedConnectionImpl.getConnection(ConnectionFactoryImpl.java:501)
   		at org.datanucleus.store.rdbms.RDBMSStoreManager.<init>(RDBMSStoreManager.java:298)
   		at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
   		at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
   		at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
   		at java.lang.reflect.Constructor.newInstance(Constructor.java:408)
   		at org.datanucleus.plugin.NonManagedPluginRegistry.createExecutableExtension(NonManagedPluginRegistry.java:631)
   		at org.datanucleus.plugin.PluginManager.createExecutableExtension(PluginManager.java:301)
   		at org.datanucleus.NucleusContext.createStoreManagerForProperties(NucleusContext.java:1187)
   		at org.datanucleus.NucleusContext.initialise(NucleusContext.java:356)
   		at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.freezeConfiguration(JDOPersistenceManagerFactory.java:775)
   		at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.createPersistenceManagerFactory(JDOPersistenceManagerFactory.java:333)
   		at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.getPersistenceManagerFactory(JDOPersistenceManagerFactory.java:202)
   		at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
   		at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
   		at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
   		at java.lang.reflect.Method.invoke(Method.java:483)
   		at javax.jdo.JDOHelper$16.run(JDOHelper.java:1965)
   		at java.security.AccessController.doPrivileged(Native Method)
   		at javax.jdo.JDOHelper.invoke(JDOHelper.java:1960)
   		at javax.jdo.JDOHelper.invokeGetPersistenceManagerFactoryOnImplementation(JDOHelper.java:1166)
   [INFO] 2020-09-03 10:57:45.336  - [taskAppId=TASK-3-3-7]:[121] -  -> 	at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:808)
   		at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:701)
   		at org.apache.hadoop.hive.metastore.ObjectStore.getPMF(ObjectStore.java:419)
   		at org.apache.hadoop.hive.metastore.ObjectStore.getPersistenceManager(ObjectStore.java:448)
   		at org.apache.hadoop.hive.metastore.ObjectStore.initialize(ObjectStore.java:343)
   		at org.apache.hadoop.hive.metastore.ObjectStore.setConf(ObjectStore.java:299)
   		at org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:73)
   		at org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:133)
   		at org.apache.hadoop.hive.metastore.RawStoreProxy.<init>(RawStoreProxy.java:60)
   		at org.apache.hadoop.hive.metastore.RawStoreProxy.getProxy(RawStoreProxy.java:69)
   		at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.newRawStore(HiveMetaStore.java:685)
   		at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMS(HiveMetaStore.java:663)
   		at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.createDefaultDB(HiveMetaStore.java:712)
   		at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.init(HiveMetaStore.java:511)
   		at org.apache.hadoop.hive.metastore.RetryingHMSHandler.<init>(RetryingHMSHandler.java:78)
   		at org.apache.hadoop.hive.metastore.RetryingHMSHandler.getProxy(RetryingHMSHandler.java:84)
   		at org.apache.hadoop.hive.metastore.HiveMetaStore.newRetryingHMSHandler(HiveMetaStore.java:6517)
   		at org.apache.hadoop.hive.metastore.HiveMetaStoreClient.<init>(HiveMetaStoreClient.java:207)
   		at org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient.<init>(SessionHiveMetaStoreClient.java:74)
   		at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
   		at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
   		at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
   		at java.lang.reflect.Constructor.newInstance(Constructor.java:408)
   		at org.apache.hadoop.hive.metastore.MetaStoreUtils.newInstance(MetaStoreUtils.java:1660)
   		at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.<init>(RetryingMetaStoreClient.java:67)
   		at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:82)
   		at org.apache.hadoop.hive.ql.metadata.Hive.createMetaStoreClient(Hive.java:3411)
   		at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3430)
   		at org.apache.hadoop.hive.ql.metadata.Hive.getAllFunctions(Hive.java:3655)
   		at org.apache.hadoop.hive.ql.metadata.Hive.reloadFunctions(Hive.java:231)
   		at org.apache.hadoop.hive.ql.metadata.Hive.registerAllFunctionsOnce(Hive.java:215)
   		at org.apache.hadoop.hive.ql.metadata.Hive.<init>(Hive.java:338)
   		at org.apache.hadoop.hive.ql.metadata.Hive.get(Hive.java:299)
   		at org.apache.hadoop.hive.ql.metadata.Hive.get(Hive.java:274)
   		at org.apache.spark.sql.hive.client.HiveClientImpl.org$apache$spark$sql$hive$client$HiveClientImpl$$client(HiveClientImpl.scala:246)
   		at org.apache.spark.sql.hive.client.HiveClientImpl$$anonfun$withHiveState$1.apply(HiveClientImpl.scala:268)
   		at org.apache.spark.sql.hive.client.HiveClientImpl.liftedTree1$1(HiveClientImpl.scala:213)
   		at org.apache.spark.sql.hive.client.HiveClientImpl.retryLocked(HiveClientImpl.scala:212)
   		at org.apache.spark.sql.hive.client.HiveClientImpl.withHiveState(HiveClientImpl.scala:258)
   		at org.apache.spark.sql.hive.client.HiveClientImpl.databaseExists(HiveClientImpl.scala:348)
   		at org.apache.spark.sql.hive.HiveExternalCatalog$$anonfun$databaseExists$1.apply$mcZ$sp(HiveExternalCatalog.scala:217)
   		at org.apache.spark.sql.hive.HiveExternalCatalog$$anonfun$databaseExists$1.apply(HiveExternalCatalog.scala:217)
   		at org.apache.spark.sql.hive.HiveExternalCatalog$$anonfun$databaseExists$1.apply(HiveExternalCatalog.scala:217)
   		at org.apache.spark.sql.hive.HiveExternalCatalog.withClient(HiveExternalCatalog.scala:99)
   		at org.apache.spark.sql.hive.HiveExternalCatalog.databaseExists(HiveExternalCatalog.scala:216)
   		at org.apache.spark.sql.internal.SharedState.externalCatalog$lzycompute(SharedState.scala:114)
   		at org.apache.spark.sql.internal.SharedState.externalCatalog(SharedState.scala:102)
   		at org.apache.spark.sql.internal.SharedState.globalTempViewManager$lzycompute(SharedState.scala:141)
   		at org.apache.spark.sql.internal.SharedState.globalTempViewManager(SharedState.scala:136)
   		at org.apache.spark.sql.hive.HiveSessionStateBuilder$$anonfun$2.apply(HiveSessionStateBuilder.scala:55)
   		at org.apache.spark.sql.hive.HiveSessionStateBuilder$$anonfun$2.apply(HiveSessionStateBuilder.scala:55)
   		at org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager$lzycompute(SessionCatalog.scala:91)
   		at org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager(SessionCatalog.scala:91)
   		at org.apache.spark.sql.catalyst.catalog.SessionCatalog.isTemporaryTable(SessionCatalog.scala:736)
   		at org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$.isRunningDirectlyOnFiles(Analyzer.scala:749)
   		at org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$.resolveRelation(Analyzer.scala:683)
   		at org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$$anonfun$apply$8.applyOrElse(Analyzer.scala:715)
   		at org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$$anonfun$apply$8.applyOrElse(Analyzer.scala:708)
   		at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1$$anonfun$apply$1.apply(AnalysisHelper.scala:90)
   		at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1$$anonfun$apply$1.apply(AnalysisHelper.scala:90)
   		at org.apache.spark.sql.catalyst.trees.CurrentOrigin$.withOrigin(TreeNode.scala:70)
   		at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1.apply(AnalysisHelper.scala:89)
   		at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1.apply(AnalysisHelper.scala:86)
   		at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$.allowInvokingTransformsInAnalyzer(AnalysisHelper.scala:194)
   [INFO] 2020-09-03 10:57:45.336  - [taskAppId=TASK-3-3-7]:[121] -  -> 	at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$class.resolveOperatorsUp(AnalysisHelper.scala:86)
   		at org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.resolveOperatorsUp(LogicalPlan.scala:29)
   		at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1$$anonfun$1.apply(AnalysisHelper.scala:87)
   		at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1$$anonfun$1.apply(AnalysisHelper.scala:87)
   		at org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$4.apply(TreeNode.scala:326)
   		at org.apache.spark.sql.catalyst.trees.TreeNode.mapProductIterator(TreeNode.scala:187)
   		at org.apache.spark.sql.catalyst.trees.TreeNode.mapChildren(TreeNode.scala:324)
   		at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1.apply(AnalysisHelper.scala:87)
   		at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1.apply(AnalysisHelper.scala:86)
   		at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$.allowInvokingTransformsInAnalyzer(AnalysisHelper.scala:194)
   		at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$class.resolveOperatorsUp(AnalysisHelper.scala:86)
   		at org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.resolveOperatorsUp(LogicalPlan.scala:29)
   		at org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$.apply(Analyzer.scala:708)
   		at org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$.apply(Analyzer.scala:654)
   		at org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1$$anonfun$apply$1.apply(RuleExecutor.scala:87)
   		at org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1$$anonfun$apply$1.apply(RuleExecutor.scala:84)
   		at scala.collection.LinearSeqOptimized$class.foldLeft(LinearSeqOptimized.scala:124)
   		at scala.collection.immutable.List.foldLeft(List.scala:84)
   		at org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1.apply(RuleExecutor.scala:84)
   		at org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1.apply(RuleExecutor.scala:76)
   		at scala.collection.immutable.List.foreach(List.scala:392)
   		at org.apache.spark.sql.catalyst.rules.RuleExecutor.execute(RuleExecutor.scala:76)
   		at org.apache.spark.sql.catalyst.analysis.Analyzer.org$apache$spark$sql$catalyst$analysis$Analyzer$$executeSameContext(Analyzer.scala:127)
   		at org.apache.spark.sql.catalyst.analysis.Analyzer.execute(Analyzer.scala:121)
   		at org.apache.spark.sql.catalyst.analysis.Analyzer$$anonfun$executeAndCheck$1.apply(Analyzer.scala:106)
   		at org.apache.spark.sql.catalyst.analysis.Analyzer$$anonfun$executeAndCheck$1.apply(Analyzer.scala:105)
   		at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$.markInAnalyzer(AnalysisHelper.scala:201)
   		at org.apache.spark.sql.catalyst.analysis.Analyzer.executeAndCheck(Analyzer.scala:105)
   		at org.apache.spark.sql.execution.QueryExecution.analyzed$lzycompute(QueryExecution.scala:57)
   		at org.apache.spark.sql.execution.QueryExecution.analyzed(QueryExecution.scala:55)
   		at org.apache.spark.sql.execution.QueryExecution.assertAnalyzed(QueryExecution.scala:47)
   		at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:78)
   		at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:642)
   		at dsjzt.sparkSqlTest$.main(sparkSqlTest.scala:14)
   		at dsjzt.sparkSqlTest.main(sparkSqlTest.scala)
   		at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
   		at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
   		at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
   		at java.lang.reflect.Method.invoke(Method.java:483)
   		at org.apache.spark.deploy.JavaMainApplication.start(SparkApplication.scala:52)
   		at org.apache.spark.deploy.SparkSubmit.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:849)
   		at org.apache.spark.deploy.SparkSubmit.doRunMain$1(SparkSubmit.scala:167)
   		at org.apache.spark.deploy.SparkSubmit.submit(SparkSubmit.scala:195)
   		at org.apache.spark.deploy.SparkSubmit.doSubmit(SparkSubmit.scala:86)
   		at org.apache.spark.deploy.SparkSubmit$$anon$2.doSubmit(SparkSubmit.scala:924)
   		at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:933)
   		at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
   	Caused by: ERROR XJ041: Failed to create database 'metastore_db', see the next exception for details.
   		at org.apache.derby.iapi.error.StandardException.newException(Unknown Source)
   		at org.apache.derby.impl.jdbc.SQLExceptionFactory.wrapArgsForTransportAcrossDRDA(Unknown Source)
   		... 146 more
   	Caused by: ERROR XBM0H: Directory /tmp/dolphinscheduler/exec/process/1/3/3/7/metastore_db cannot be created.
   		at org.apache.derby.iapi.error.StandardException.newException(Unknown Source)
   		at org.apache.derby.iapi.error.StandardException.newException(Unknown Source)
   		at org.apache.derby.impl.services.monitor.StorageFactoryService$10.run(Unknown Source)
   		at java.security.AccessController.doPrivileged(Native Method)
   		at org.apache.derby.impl.services.monitor.StorageFactoryService.createServiceRoot(Unknown Source)
   		at org.apache.derby.impl.services.monitor.BaseMonitor.bootService(Unknown Source)
   		at org.apache.derby.impl.services.monitor.BaseMonitor.createPersistentService(Unknown Source)
   		at org.apache.derby.impl.services.monitor.FileMonitor.createPersistentService(Unknown Source)
   		at org.apache.derby.iapi.services.monitor.Monitor.createPersistentService(Unknown Source)
   		... 143 more
   	============= begin nested exception, level (1) ===========
   	java.sql.SQLException: Directory /tmp/dolphinscheduler/exec/process/1/3/3/7/metastore_db cannot be created.
   [INFO] 2020-09-03 10:57:45.337  - [taskAppId=TASK-3-3-7]:[121] -  -> 	at org.apache.derby.impl.jdbc.SQLExceptionFactory.getSQLException(Unknown Source)
   		at org.apache.derby.impl.jdbc.Util.generateCsSQLException(Unknown Source)
   		at org.apache.derby.impl.jdbc.TransactionResourceImpl.wrapInSQLException(Unknown Source)
   		at org.apache.derby.impl.jdbc.TransactionResourceImpl.handleException(Unknown Source)
   		at org.apache.derby.impl.jdbc.EmbedConnection.handleException(Unknown Source)
   		at org.apache.derby.impl.jdbc.EmbedConnection.createDatabase(Unknown Source)
   		at org.apache.derby.impl.jdbc.EmbedConnection.<init>(Unknown Source)
   		at org.apache.derby.jdbc.InternalDriver.getNewEmbedConnection(Unknown Source)
   		at org.apache.derby.jdbc.InternalDriver.connect(Unknown Source)
   		at org.apache.derby.jdbc.InternalDriver.connect(Unknown Source)
   		at org.apache.derby.jdbc.AutoloadedDriver.connect(Unknown Source)
   		at java.sql.DriverManager.getConnection(DriverManager.java:664)
   		at java.sql.DriverManager.getConnection(DriverManager.java:208)
   		at com.jolbox.bonecp.BoneCP.obtainRawInternalConnection(BoneCP.java:349)
   		at com.jolbox.bonecp.BoneCP.<init>(BoneCP.java:416)
   		at com.jolbox.bonecp.BoneCPDataSource.getConnection(BoneCPDataSource.java:120)
   		at org.datanucleus.store.rdbms.ConnectionFactoryImpl$ManagedConnectionImpl.getConnection(ConnectionFactoryImpl.java:501)
   		at org.datanucleus.store.rdbms.RDBMSStoreManager.<init>(RDBMSStoreManager.java:298)
   		at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
   		at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
   		at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
   		at java.lang.reflect.Constructor.newInstance(Constructor.java:408)
   		at org.datanucleus.plugin.NonManagedPluginRegistry.createExecutableExtension(NonManagedPluginRegistry.java:631)
   		at org.datanucleus.plugin.PluginManager.createExecutableExtension(PluginManager.java:301)
   		at org.datanucleus.NucleusContext.createStoreManagerForProperties(NucleusContext.java:1187)
   		at org.datanucleus.NucleusContext.initialise(NucleusContext.java:356)
   		at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.freezeConfiguration(JDOPersistenceManagerFactory.java:775)
   		at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.createPersistenceManagerFactory(JDOPersistenceManagerFactory.java:333)
   		at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.getPersistenceManagerFactory(JDOPersistenceManagerFactory.java:202)
   		at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
   		at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
   		at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
   		at java.lang.reflect.Method.invoke(Method.java:483)
   		at javax.jdo.JDOHelper$16.run(JDOHelper.java:1965)
   		at java.security.AccessController.doPrivileged(Native Method)
   		at javax.jdo.JDOHelper.invoke(JDOHelper.java:1960)
   		at javax.jdo.JDOHelper.invokeGetPersistenceManagerFactoryOnImplementation(JDOHelper.java:1166)
   		at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:808)
   		at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:701)
   		at org.apache.hadoop.hive.metastore.ObjectStore.getPMF(ObjectStore.java:419)
   		at org.apache.hadoop.hive.metastore.ObjectStore.getPersistenceManager(ObjectStore.java:448)
   		at org.apache.hadoop.hive.metastore.ObjectStore.initialize(ObjectStore.java:343)
   		at org.apache.hadoop.hive.metastore.ObjectStore.setConf(ObjectStore.java:299)
   		at org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:73)
   		at org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:133)
   		at org.apache.hadoop.hive.metastore.RawStoreProxy.<init>(RawStoreProxy.java:60)
   		at org.apache.hadoop.hive.metastore.RawStoreProxy.getProxy(RawStoreProxy.java:69)
   		at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.newRawStore(HiveMetaStore.java:685)
   		at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMS(HiveMetaStore.java:663)
   		at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.createDefaultDB(HiveMetaStore.java:712)
   		at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.init(HiveMetaStore.java:511)
   		at org.apache.hadoop.hive.metastore.RetryingHMSHandler.<init>(RetryingHMSHandler.java:78)
   		at org.apache.hadoop.hive.metastore.RetryingHMSHandler.getProxy(RetryingHMSHandler.java:84)
   		at org.apache.hadoop.hive.metastore.HiveMetaStore.newRetryingHMSHandler(HiveMetaStore.java:6517)
   		at org.apache.hadoop.hive.metastore.HiveMetaStoreClient.<init>(HiveMetaStoreClient.java:207)
   		at org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient.<init>(SessionHiveMetaStoreClient.java:74)
   		at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
   		at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
   		at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
   		at java.lang.reflect.Constructor.newInstance(Constructor.java:408)
   		at org.apache.hadoop.hive.metastore.MetaStoreUtils.newInstance(MetaStoreUtils.java:1660)
   		at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.<init>(RetryingMetaStoreClient.java:67)
   		at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:82)
   		at org.apache.hadoop.hive.ql.metadata.Hive.createMetaStoreClient(Hive.java:3411)
   [INFO] 2020-09-03 10:57:45.337  - [taskAppId=TASK-3-3-7]:[121] -  -> 	at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3430)
   		at org.apache.hadoop.hive.ql.metadata.Hive.getAllFunctions(Hive.java:3655)
   		at org.apache.hadoop.hive.ql.metadata.Hive.reloadFunctions(Hive.java:231)
   		at org.apache.hadoop.hive.ql.metadata.Hive.registerAllFunctionsOnce(Hive.java:215)
   		at org.apache.hadoop.hive.ql.metadata.Hive.<init>(Hive.java:338)
   		at org.apache.hadoop.hive.ql.metadata.Hive.get(Hive.java:299)
   		at org.apache.hadoop.hive.ql.metadata.Hive.get(Hive.java:274)
   		at org.apache.spark.sql.hive.client.HiveClientImpl.org$apache$spark$sql$hive$client$HiveClientImpl$$client(HiveClientImpl.scala:246)
   		at org.apache.spark.sql.hive.client.HiveClientImpl$$anonfun$withHiveState$1.apply(HiveClientImpl.scala:268)
   		at org.apache.spark.sql.hive.client.HiveClientImpl.liftedTree1$1(HiveClientImpl.scala:213)
   		at org.apache.spark.sql.hive.client.HiveClientImpl.retryLocked(HiveClientImpl.scala:212)
   		at org.apache.spark.sql.hive.client.HiveClientImpl.withHiveState(HiveClientImpl.scala:258)
   		at org.apache.spark.sql.hive.client.HiveClientImpl.databaseExists(HiveClientImpl.scala:348)
   		at org.apache.spark.sql.hive.HiveExternalCatalog$$anonfun$databaseExists$1.apply$mcZ$sp(HiveExternalCatalog.scala:217)
   		at org.apache.spark.sql.hive.HiveExternalCatalog$$anonfun$databaseExists$1.apply(HiveExternalCatalog.scala:217)
   		at org.apache.spark.sql.hive.HiveExternalCatalog$$anonfun$databaseExists$1.apply(HiveExternalCatalog.scala:217)
   		at org.apache.spark.sql.hive.HiveExternalCatalog.withClient(HiveExternalCatalog.scala:99)
   		at org.apache.spark.sql.hive.HiveExternalCatalog.databaseExists(HiveExternalCatalog.scala:216)
   		at org.apache.spark.sql.internal.SharedState.externalCatalog$lzycompute(SharedState.scala:114)
   		at org.apache.spark.sql.internal.SharedState.externalCatalog(SharedState.scala:102)
   		at org.apache.spark.sql.internal.SharedState.globalTempViewManager$lzycompute(SharedState.scala:141)
   		at org.apache.spark.sql.internal.SharedState.globalTempViewManager(SharedState.scala:136)
   		at org.apache.spark.sql.hive.HiveSessionStateBuilder$$anonfun$2.apply(HiveSessionStateBuilder.scala:55)
   		at org.apache.spark.sql.hive.HiveSessionStateBuilder$$anonfun$2.apply(HiveSessionStateBuilder.scala:55)
   		at org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager$lzycompute(SessionCatalog.scala:91)
   		at org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager(SessionCatalog.scala:91)
   		at org.apache.spark.sql.catalyst.catalog.SessionCatalog.isTemporaryTable(SessionCatalog.scala:736)
   		at org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$.isRunningDirectlyOnFiles(Analyzer.scala:749)
   		at org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$.resolveRelation(Analyzer.scala:683)
   		at org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$$anonfun$apply$8.applyOrElse(Analyzer.scala:715)
   		at org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$$anonfun$apply$8.applyOrElse(Analyzer.scala:708)
   		at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1$$anonfun$apply$1.apply(AnalysisHelper.scala:90)
   		at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1$$anonfun$apply$1.apply(AnalysisHelper.scala:90)
   		at org.apache.spark.sql.catalyst.trees.CurrentOrigin$.withOrigin(TreeNode.scala:70)
   		at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1.apply(AnalysisHelper.scala:89)
   		at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1.apply(AnalysisHelper.scala:86)
   		at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$.allowInvokingTransformsInAnalyzer(AnalysisHelper.scala:194)
   		at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$class.resolveOperatorsUp(AnalysisHelper.scala:86)
   		at org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.resolveOperatorsUp(LogicalPlan.scala:29)
   		at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1$$anonfun$1.apply(AnalysisHelper.scala:87)
   		at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1$$anonfun$1.apply(AnalysisHelper.scala:87)
   		at org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$4.apply(TreeNode.scala:326)
   		at org.apache.spark.sql.catalyst.trees.TreeNode.mapProductIterator(TreeNode.scala:187)
   		at org.apache.spark.sql.catalyst.trees.TreeNode.mapChildren(TreeNode.scala:324)
   		at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1.apply(AnalysisHelper.scala:87)
   		at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1.apply(AnalysisHelper.scala:86)
   		at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$.allowInvokingTransformsInAnalyzer(AnalysisHelper.scala:194)
   		at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$class.resolveOperatorsUp(AnalysisHelper.scala:86)
   		at org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.resolveOperatorsUp(LogicalPlan.scala:29)
   		at org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$.apply(Analyzer.scala:708)
   		at org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$.apply(Analyzer.scala:654)
   		at org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1$$anonfun$apply$1.apply(RuleExecutor.scala:87)
   		at org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1$$anonfun$apply$1.apply(RuleExecutor.scala:84)
   		at scala.collection.LinearSeqOptimized$class.foldLeft(LinearSeqOptimized.scala:124)
   		at scala.collection.immutable.List.foldLeft(List.scala:84)
   		at org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1.apply(RuleExecutor.scala:84)
   		at org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1.apply(RuleExecutor.scala:76)
   		at scala.collection.immutable.List.foreach(List.scala:392)
   		at org.apache.spark.sql.catalyst.rules.RuleExecutor.execute(RuleExecutor.scala:76)
   		at org.apache.spark.sql.catalyst.analysis.Analyzer.org$apache$spark$sql$catalyst$analysis$Analyzer$$executeSameContext(Analyzer.scala:127)
   		at org.apache.spark.sql.catalyst.analysis.Analyzer.execute(Analyzer.scala:121)
   		at org.apache.spark.sql.catalyst.analysis.Analyzer$$anonfun$executeAndCheck$1.apply(Analyzer.scala:106)
   		at org.apache.spark.sql.catalyst.analysis.Analyzer$$anonfun$executeAndCheck$1.apply(Analyzer.scala:105)
   		at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$.markInAnalyzer(AnalysisHelper.scala:201)
   [INFO] 2020-09-03 10:57:45.338  - [taskAppId=TASK-3-3-7]:[121] -  -> 	at org.apache.spark.sql.catalyst.analysis.Analyzer.executeAndCheck(Analyzer.scala:105)
   		at org.apache.spark.sql.execution.QueryExecution.analyzed$lzycompute(QueryExecution.scala:57)
   		at org.apache.spark.sql.execution.QueryExecution.analyzed(QueryExecution.scala:55)
   		at org.apache.spark.sql.execution.QueryExecution.assertAnalyzed(QueryExecution.scala:47)
   		at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:78)
   		at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:642)
   		at dsjzt.sparkSqlTest$.main(sparkSqlTest.scala:14)
   		at dsjzt.sparkSqlTest.main(sparkSqlTest.scala)
   		at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
   		at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
   		at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
   		at java.lang.reflect.Method.invoke(Method.java:483)
   		at org.apache.spark.deploy.JavaMainApplication.start(SparkApplication.scala:52)
   		at org.apache.spark.deploy.SparkSubmit.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:849)
   		at org.apache.spark.deploy.SparkSubmit.doRunMain$1(SparkSubmit.scala:167)
   		at org.apache.spark.deploy.SparkSubmit.submit(SparkSubmit.scala:195)
   		at org.apache.spark.deploy.SparkSubmit.doSubmit(SparkSubmit.scala:86)
   		at org.apache.spark.deploy.SparkSubmit$$anon$2.doSubmit(SparkSubmit.scala:924)
   		at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:933)
   		at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
   	Caused by: ERROR XBM0H: Directory /tmp/dolphinscheduler/exec/process/1/3/3/7/metastore_db cannot be created.
   		at org.apache.derby.iapi.error.StandardException.newException(Unknown Source)
   		at org.apache.derby.iapi.error.StandardException.newException(Unknown Source)
   		at org.apache.derby.impl.services.monitor.StorageFactoryService$10.run(Unknown Source)
   		at java.security.AccessController.doPrivileged(Native Method)
   		at org.apache.derby.impl.services.monitor.StorageFactoryService.createServiceRoot(Unknown Source)
   		at org.apache.derby.impl.services.monitor.BaseMonitor.bootService(Unknown Source)
   		at org.apache.derby.impl.services.monitor.BaseMonitor.createPersistentService(Unknown Source)
   		at org.apache.derby.impl.services.monitor.FileMonitor.createPersistentService(Unknown Source)
   		at org.apache.derby.iapi.services.monitor.Monitor.createPersistentService(Unknown Source)
   		... 143 more
   	============= end nested exception, level (1) ===========
   	============= begin nested exception, level (2) ===========
   	ERROR XBM0H: Directory /tmp/dolphinscheduler/exec/process/1/3/3/7/metastore_db cannot be created.
   		at org.apache.derby.iapi.error.StandardException.newException(Unknown Source)
   		at org.apache.derby.iapi.error.StandardException.newException(Unknown Source)
   		at org.apache.derby.impl.services.monitor.StorageFactoryService$10.run(Unknown Source)
   		at java.security.AccessController.doPrivileged(Native Method)
   		at org.apache.derby.impl.services.monitor.StorageFactoryService.createServiceRoot(Unknown Source)
   		at org.apache.derby.impl.services.monitor.BaseMonitor.bootService(Unknown Source)
   		at org.apache.derby.impl.services.monitor.BaseMonitor.createPersistentService(Unknown Source)
   		at org.apache.derby.impl.services.monitor.FileMonitor.createPersistentService(Unknown Source)
   		at org.apache.derby.iapi.services.monitor.Monitor.createPersistentService(Unknown Source)
   		at org.apache.derby.impl.jdbc.EmbedConnection.createDatabase(Unknown Source)
   		at org.apache.derby.impl.jdbc.EmbedConnection.<init>(Unknown Source)
   		at org.apache.derby.jdbc.InternalDriver.getNewEmbedConnection(Unknown Source)
   		at org.apache.derby.jdbc.InternalDriver.connect(Unknown Source)
   		at org.apache.derby.jdbc.InternalDriver.connect(Unknown Source)
   		at org.apache.derby.jdbc.AutoloadedDriver.connect(Unknown Source)
   		at java.sql.DriverManager.getConnection(DriverManager.java:664)
   		at java.sql.DriverManager.getConnection(DriverManager.java:208)
   		at com.jolbox.bonecp.BoneCP.obtainRawInternalConnection(BoneCP.java:349)
   		at com.jolbox.bonecp.BoneCP.<init>(BoneCP.java:416)
   		at com.jolbox.bonecp.BoneCPDataSource.getConnection(BoneCPDataSource.java:120)
   		at org.datanucleus.store.rdbms.ConnectionFactoryImpl$ManagedConnectionImpl.getConnection(ConnectionFactoryImpl.java:501)
   		at org.datanucleus.store.rdbms.RDBMSStoreManager.<init>(RDBMSStoreManager.java:298)
   		at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
   		at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
   		at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
   		at java.lang.reflect.Constructor.newInstance(Constructor.java:408)
   		at org.datanucleus.plugin.NonManagedPluginRegistry.createExecutableExtension(NonManagedPluginRegistry.java:631)
   		at org.datanucleus.plugin.PluginManager.createExecutableExtension(PluginManager.java:301)
   		at org.datanucleus.NucleusContext.createStoreManagerForProperties(NucleusContext.java:1187)
   		at org.datanucleus.NucleusContext.initialise(NucleusContext.java:356)
   [INFO] 2020-09-03 10:57:45.338  - [taskAppId=TASK-3-3-7]:[121] -  -> 	at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.freezeConfiguration(JDOPersistenceManagerFactory.java:775)
   		at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.createPersistenceManagerFactory(JDOPersistenceManagerFactory.java:333)
   		at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.getPersistenceManagerFactory(JDOPersistenceManagerFactory.java:202)
   		at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
   		at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
   		at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
   		at java.lang.reflect.Method.invoke(Method.java:483)
   		at javax.jdo.JDOHelper$16.run(JDOHelper.java:1965)
   		at java.security.AccessController.doPrivileged(Native Method)
   		at javax.jdo.JDOHelper.invoke(JDOHelper.java:1960)
   		at javax.jdo.JDOHelper.invokeGetPersistenceManagerFactoryOnImplementation(JDOHelper.java:1166)
   		at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:808)
   		at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:701)
   		at org.apache.hadoop.hive.metastore.ObjectStore.getPMF(ObjectStore.java:419)
   		at org.apache.hadoop.hive.metastore.ObjectStore.getPersistenceManager(ObjectStore.java:448)
   		at org.apache.hadoop.hive.metastore.ObjectStore.initialize(ObjectStore.java:343)
   		at org.apache.hadoop.hive.metastore.ObjectStore.setConf(ObjectStore.java:299)
   		at org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:73)
   		at org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:133)
   		at org.apache.hadoop.hive.metastore.RawStoreProxy.<init>(RawStoreProxy.java:60)
   		at org.apache.hadoop.hive.metastore.RawStoreProxy.getProxy(RawStoreProxy.java:69)
   		at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.newRawStore(HiveMetaStore.java:685)
   		at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMS(HiveMetaStore.java:663)
   		at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.createDefaultDB(HiveMetaStore.java:712)
   		at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.init(HiveMetaStore.java:511)
   		at org.apache.hadoop.hive.metastore.RetryingHMSHandler.<init>(RetryingHMSHandler.java:78)
   		at org.apache.hadoop.hive.metastore.RetryingHMSHandler.getProxy(RetryingHMSHandler.java:84)
   		at org.apache.hadoop.hive.metastore.HiveMetaStore.newRetryingHMSHandler(HiveMetaStore.java:6517)
   		at org.apache.hadoop.hive.metastore.HiveMetaStoreClient.<init>(HiveMetaStoreClient.java:207)
   		at org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient.<init>(SessionHiveMetaStoreClient.java:74)
   		at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
   		at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
   		at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
   		at java.lang.reflect.Constructor.newInstance(Constructor.java:408)
   		at org.apache.hadoop.hive.metastore.MetaStoreUtils.newInstance(MetaStoreUtils.java:1660)
   		at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.<init>(RetryingMetaStoreClient.java:67)
   		at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:82)
   		at org.apache.hadoop.hive.ql.metadata.Hive.createMetaStoreClient(Hive.java:3411)
   		at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3430)
   		at org.apache.hadoop.hive.ql.metadata.Hive.getAllFunctions(Hive.java:3655)
   		at org.apache.hadoop.hive.ql.metadata.Hive.reloadFunctions(Hive.java:231)
   		at org.apache.hadoop.hive.ql.metadata.Hive.registerAllFunctionsOnce(Hive.java:215)
   		at org.apache.hadoop.hive.ql.metadata.Hive.<init>(Hive.java:338)
   		at org.apache.hadoop.hive.ql.metadata.Hive.get(Hive.java:299)
   		at org.apache.hadoop.hive.ql.metadata.Hive.get(Hive.java:274)
   		at org.apache.spark.sql.hive.client.HiveClientImpl.org$apache$spark$sql$hive$client$HiveClientImpl$$client(HiveClientImpl.scala:246)
   		at org.apache.spark.sql.hive.client.HiveClientImpl$$anonfun$withHiveState$1.apply(HiveClientImpl.scala:268)
   		at org.apache.spark.sql.hive.client.HiveClientImpl.liftedTree1$1(HiveClientImpl.scala:213)
   		at org.apache.spark.sql.hive.client.HiveClientImpl.retryLocked(HiveClientImpl.scala:212)
   		at org.apache.spark.sql.hive.client.HiveClientImpl.withHiveState(HiveClientImpl.scala:258)
   		at org.apache.spark.sql.hive.client.HiveClientImpl.databaseExists(HiveClientImpl.scala:348)
   		at org.apache.spark.sql.hive.HiveExternalCatalog$$anonfun$databaseExists$1.apply$mcZ$sp(HiveExternalCatalog.scala:217)
   		at org.apache.spark.sql.hive.HiveExternalCatalog$$anonfun$databaseExists$1.apply(HiveExternalCatalog.scala:217)
   		at org.apache.spark.sql.hive.HiveExternalCatalog$$anonfun$databaseExists$1.apply(HiveExternalCatalog.scala:217)
   		at org.apache.spark.sql.hive.HiveExternalCatalog.withClient(HiveExternalCatalog.scala:99)
   		at org.apache.spark.sql.hive.HiveExternalCatalog.databaseExists(HiveExternalCatalog.scala:216)
   		at org.apache.spark.sql.internal.SharedState.externalCatalog$lzycompute(SharedState.scala:114)
   		at org.apache.spark.sql.internal.SharedState.externalCatalog(SharedState.scala:102)
   		at org.apache.spark.sql.internal.SharedState.globalTempViewManager$lzycompute(SharedState.scala:141)
   		at org.apache.spark.sql.internal.SharedState.globalTempViewManager(SharedState.scala:136)
   		at org.apache.spark.sql.hive.HiveSessionStateBuilder$$anonfun$2.apply(HiveSessionStateBuilder.scala:55)
   		at org.apache.spark.sql.hive.HiveSessionStateBuilder$$anonfun$2.apply(HiveSessionStateBuilder.scala:55)
   		at org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager$lzycompute(SessionCatalog.scala:91)
   		at org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager(SessionCatalog.scala:91)
   [INFO] 2020-09-03 10:57:45.338  - [taskAppId=TASK-3-3-7]:[121] -  -> 	at org.apache.spark.sql.catalyst.catalog.SessionCatalog.isTemporaryTable(SessionCatalog.scala:736)
   		at org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$.isRunningDirectlyOnFiles(Analyzer.scala:749)
   		at org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$.resolveRelation(Analyzer.scala:683)
   		at org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$$anonfun$apply$8.applyOrElse(Analyzer.scala:715)
   		at org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$$anonfun$apply$8.applyOrElse(Analyzer.scala:708)
   		at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1$$anonfun$apply$1.apply(AnalysisHelper.scala:90)
   		at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1$$anonfun$apply$1.apply(AnalysisHelper.scala:90)
   		at org.apache.spark.sql.catalyst.trees.CurrentOrigin$.withOrigin(TreeNode.scala:70)
   		at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1.apply(AnalysisHelper.scala:89)
   		at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1.apply(AnalysisHelper.scala:86)
   		at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$.allowInvokingTransformsInAnalyzer(AnalysisHelper.scala:194)
   		at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$class.resolveOperatorsUp(AnalysisHelper.scala:86)
   		at org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.resolveOperatorsUp(LogicalPlan.scala:29)
   		at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1$$anonfun$1.apply(AnalysisHelper.scala:87)
   		at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1$$anonfun$1.apply(AnalysisHelper.scala:87)
   		at org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$4.apply(TreeNode.scala:326)
   		at org.apache.spark.sql.catalyst.trees.TreeNode.mapProductIterator(TreeNode.scala:187)
   		at org.apache.spark.sql.catalyst.trees.TreeNode.mapChildren(TreeNode.scala:324)
   		at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1.apply(AnalysisHelper.scala:87)
   		at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1.apply(AnalysisHelper.scala:86)
   		at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$.allowInvokingTransformsInAnalyzer(AnalysisHelper.scala:194)
   		at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$class.resolveOperatorsUp(AnalysisHelper.scala:86)
   		at org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.resolveOperatorsUp(LogicalPlan.scala:29)
   		at org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$.apply(Analyzer.scala:708)
   		at org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$.apply(Analyzer.scala:654)
   		at org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1$$anonfun$apply$1.apply(RuleExecutor.scala:87)
   		at org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1$$anonfun$apply$1.apply(RuleExecutor.scala:84)
   		at scala.collection.LinearSeqOptimized$class.foldLeft(LinearSeqOptimized.scala:124)
   		at scala.collection.immutable.List.foldLeft(List.scala:84)
   		at org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1.apply(RuleExecutor.scala:84)
   		at org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1.apply(RuleExecutor.scala:76)
   		at scala.collection.immutable.List.foreach(List.scala:392)
   		at org.apache.spark.sql.catalyst.rules.RuleExecutor.execute(RuleExecutor.scala:76)
   		at org.apache.spark.sql.catalyst.analysis.Analyzer.org$apache$spark$sql$catalyst$analysis$Analyzer$$executeSameContext(Analyzer.scala:127)
   		at org.apache.spark.sql.catalyst.analysis.Analyzer.execute(Analyzer.scala:121)
   		at org.apache.spark.sql.catalyst.analysis.Analyzer$$anonfun$executeAndCheck$1.apply(Analyzer.scala:106)
   		at org.apache.spark.sql.catalyst.analysis.Analyzer$$anonfun$executeAndCheck$1.apply(Analyzer.scala:105)
   		at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$.markInAnalyzer(AnalysisHelper.scala:201)
   		at org.apache.spark.sql.catalyst.analysis.Analyzer.executeAndCheck(Analyzer.scala:105)
   		at org.apache.spark.sql.execution.QueryExecution.analyzed$lzycompute(QueryExecution.scala:57)
   		at org.apache.spark.sql.execution.QueryExecution.analyzed(QueryExecution.scala:55)
   		at org.apache.spark.sql.execution.QueryExecution.assertAnalyzed(QueryExecution.scala:47)
   		at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:78)
   		at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:642)
   		at dsjzt.sparkSqlTest$.main(sparkSqlTest.scala:14)
   		at dsjzt.sparkSqlTest.main(sparkSqlTest.scala)
   		at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
   		at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
   		at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
   		at java.lang.reflect.Method.invoke(Method.java:483)
   		at org.apache.spark.deploy.JavaMainApplication.start(SparkApplication.scala:52)
   		at org.apache.spark.deploy.SparkSubmit.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:849)
   		at org.apache.spark.deploy.SparkSubmit.doRunMain$1(SparkSubmit.scala:167)
   		at org.apache.spark.deploy.SparkSubmit.submit(SparkSubmit.scala:195)
   		at org.apache.spark.deploy.SparkSubmit.doSubmit(SparkSubmit.scala:86)
   		at org.apache.spark.deploy.SparkSubmit$$anon$2.doSubmit(SparkSubmit.scala:924)
   		at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:933)
   		at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
   	============= end nested exception, level (2) ===========
   	Cleanup action completed
   	Thu Sep 03 10:57:45 CST 2020 Thread[main,5,main] Cleanup action starting
   	ERROR XBM0H: Directory /tmp/dolphinscheduler/exec/process/1/3/3/7/metastore_db cannot be created.
   		at org.apache.derby.iapi.error.StandardException.newException(Unknown Source)
   		at org.apache.derby.iapi.error.StandardException.newException(Unknown Source)
   [INFO] 2020-09-03 10:57:45.339  - [taskAppId=TASK-3-3-7]:[121] -  -> 	at org.apache.derby.impl.services.monitor.StorageFactoryService$10.run(Unknown Source)
   		at java.security.AccessController.doPrivileged(Native Method)
   		at org.apache.derby.impl.services.monitor.StorageFactoryService.createServiceRoot(Unknown Source)
   		at org.apache.derby.impl.services.monitor.BaseMonitor.bootService(Unknown Source)
   		at org.apache.derby.impl.services.monitor.BaseMonitor.createPersistentService(Unknown Source)
   		at org.apache.derby.impl.services.monitor.FileMonitor.createPersistentService(Unknown Source)
   		at org.apache.derby.iapi.services.monitor.Monitor.createPersistentService(Unknown Source)
   		at org.apache.derby.impl.jdbc.EmbedConnection.createDatabase(Unknown Source)
   		at org.apache.derby.impl.jdbc.EmbedConnection.<init>(Unknown Source)
   		at org.apache.derby.jdbc.InternalDriver.getNewEmbedConnection(Unknown Source)
   		at org.apache.derby.jdbc.InternalDriver.connect(Unknown Source)
   		at org.apache.derby.jdbc.InternalDriver.connect(Unknown Source)
   		at org.apache.derby.jdbc.AutoloadedDriver.connect(Unknown Source)
   		at java.sql.DriverManager.getConnection(DriverManager.java:664)
   		at java.sql.DriverManager.getConnection(DriverManager.java:208)
   		at com.jolbox.bonecp.BoneCP.obtainRawInternalConnection(BoneCP.java:361)
   		at com.jolbox.bonecp.BoneCP.<init>(BoneCP.java:416)
   		at com.jolbox.bonecp.BoneCPDataSource.getConnection(BoneCPDataSource.java:120)
   		at org.datanucleus.store.rdbms.ConnectionFactoryImpl$ManagedConnectionImpl.getConnection(ConnectionFactoryImpl.java:501)
   		at org.datanucleus.store.rdbms.RDBMSStoreManager.<init>(RDBMSStoreManager.java:298)
   		at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
   		at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
   		at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
   		at java.lang.reflect.Constructor.newInstance(Constructor.java:408)
   		at org.datanucleus.plugin.NonManagedPluginRegistry.createExecutableExtension(NonManagedPluginRegistry.java:631)
   		at org.datanucleus.plugin.PluginManager.createExecutableExtension(PluginManager.java:301)
   		at org.datanucleus.NucleusContext.createStoreManagerForProperties(NucleusContext.java:1187)
   		at org.datanucleus.NucleusContext.initialise(NucleusContext.java:356)
   		at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.freezeConfiguration(JDOPersistenceManagerFactory.java:775)
   		at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.createPersistenceManagerFactory(JDOPersistenceManagerFactory.java:333)
   		at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.getPersistenceManagerFactory(JDOPersistenceManagerFactory.java:202)
   		at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
   		at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
   		at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
   		at java.lang.reflect.Method.invoke(Method.java:483)
   		at javax.jdo.JDOHelper$16.run(JDOHelper.java:1965)
   		at java.security.AccessController.doPrivileged(Native Method)
   		at javax.jdo.JDOHelper.invoke(JDOHelper.java:1960)
   		at javax.jdo.JDOHelper.invokeGetPersistenceManagerFactoryOnImplementation(JDOHelper.java:1166)
   		at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:808)
   		at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:701)
   		at org.apache.hadoop.hive.metastore.ObjectStore.getPMF(ObjectStore.java:419)
   		at org.apache.hadoop.hive.metastore.ObjectStore.getPersistenceManager(ObjectStore.java:448)
   		at org.apache.hadoop.hive.metastore.ObjectStore.initialize(ObjectStore.java:343)
   		at org.apache.hadoop.hive.metastore.ObjectStore.setConf(ObjectStore.java:299)
   		at org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:73)
   		at org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:133)
   		at org.apache.hadoop.hive.metastore.RawStoreProxy.<init>(RawStoreProxy.java:60)
   		at org.apache.hadoop.hive.metastore.RawStoreProxy.getProxy(RawStoreProxy.java:69)
   		at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.newRawStore(HiveMetaStore.java:685)
   		at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMS(HiveMetaStore.java:663)
   		at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.createDefaultDB(HiveMetaStore.java:712)
   		at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.init(HiveMetaStore.java:511)
   		at org.apache.hadoop.hive.metastore.RetryingHMSHandler.<init>(RetryingHMSHandler.java:78)
   		at org.apache.hadoop.hive.metastore.RetryingHMSHandler.getProxy(RetryingHMSHandler.java:84)
   		at org.apache.hadoop.hive.metastore.HiveMetaStore.newRetryingHMSHandler(HiveMetaStore.java:6517)
   		at org.apache.hadoop.hive.metastore.HiveMetaStoreClient.<init>(HiveMetaStoreClient.java:207)
   		at org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient.<init>(SessionHiveMetaStoreClient.java:74)
   		at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
   		at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
   		at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
   		at java.lang.reflect.Constructor.newInstance(Constructor.java:408)
   		at org.apache.hadoop.hive.metastore.MetaStoreUtils.newInstance(MetaStoreUtils.java:1660)
   		at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.<init>(RetryingMetaStoreClient.java:67)
   [INFO] 2020-09-03 10:57:45.339  - [taskAppId=TASK-3-3-7]:[121] -  -> 	at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:82)
   		at org.apache.hadoop.hive.ql.metadata.Hive.createMetaStoreClient(Hive.java:3411)
   		at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3430)
   		at org.apache.hadoop.hive.ql.metadata.Hive.getAllFunctions(Hive.java:3655)
   		at org.apache.hadoop.hive.ql.metadata.Hive.reloadFunctions(Hive.java:231)
   		at org.apache.hadoop.hive.ql.metadata.Hive.registerAllFunctionsOnce(Hive.java:215)
   		at org.apache.hadoop.hive.ql.metadata.Hive.<init>(Hive.java:338)
   		at org.apache.hadoop.hive.ql.metadata.Hive.get(Hive.java:299)
   		at org.apache.hadoop.hive.ql.metadata.Hive.get(Hive.java:274)
   		at org.apache.spark.sql.hive.client.HiveClientImpl.org$apache$spark$sql$hive$client$HiveClientImpl$$client(HiveClientImpl.scala:246)
   		at org.apache.spark.sql.hive.client.HiveClientImpl$$anonfun$withHiveState$1.apply(HiveClientImpl.scala:268)
   		at org.apache.spark.sql.hive.client.HiveClientImpl.liftedTree1$1(HiveClientImpl.scala:213)
   		at org.apache.spark.sql.hive.client.HiveClientImpl.retryLocked(HiveClientImpl.scala:212)
   		at org.apache.spark.sql.hive.client.HiveClientImpl.withHiveState(HiveClientImpl.scala:258)
   		at org.apache.spark.sql.hive.client.HiveClientImpl.databaseExists(HiveClientImpl.scala:348)
   		at org.apache.spark.sql.hive.HiveExternalCatalog$$anonfun$databaseExists$1.apply$mcZ$sp(HiveExternalCatalog.scala:217)
   		at org.apache.spark.sql.hive.HiveExternalCatalog$$anonfun$databaseExists$1.apply(HiveExternalCatalog.scala:217)
   		at org.apache.spark.sql.hive.HiveExternalCatalog$$anonfun$databaseExists$1.apply(HiveExternalCatalog.scala:217)
   		at org.apache.spark.sql.hive.HiveExternalCatalog.withClient(HiveExternalCatalog.scala:99)
   		at org.apache.spark.sql.hive.HiveExternalCatalog.databaseExists(HiveExternalCatalog.scala:216)
   		at org.apache.spark.sql.internal.SharedState.externalCatalog$lzycompute(SharedState.scala:114)
   		at org.apache.spark.sql.internal.SharedState.externalCatalog(SharedState.scala:102)
   		at org.apache.spark.sql.internal.SharedState.globalTempViewManager$lzycompute(SharedState.scala:141)
   		at org.apache.spark.sql.internal.SharedState.globalTempViewManager(SharedState.scala:136)
   		at org.apache.spark.sql.hive.HiveSessionStateBuilder$$anonfun$2.apply(HiveSessionStateBuilder.scala:55)
   		at org.apache.spark.sql.hive.HiveSessionStateBuilder$$anonfun$2.apply(HiveSessionStateBuilder.scala:55)
   		at org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager$lzycompute(SessionCatalog.scala:91)
   		at org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager(SessionCatalog.scala:91)
   		at org.apache.spark.sql.catalyst.catalog.SessionCatalog.isTemporaryTable(SessionCatalog.scala:736)
   		at org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$.isRunningDirectlyOnFiles(Analyzer.scala:749)
   		at org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$.resolveRelation(Analyzer.scala:683)
   		at org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$$anonfun$apply$8.applyOrElse(Analyzer.scala:715)
   		at org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$$anonfun$apply$8.applyOrElse(Analyzer.scala:708)
   		at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1$$anonfun$apply$1.apply(AnalysisHelper.scala:90)
   		at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1$$anonfun$apply$1.apply(AnalysisHelper.scala:90)
   		at org.apache.spark.sql.catalyst.trees.CurrentOrigin$.withOrigin(TreeNode.scala:70)
   		at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1.apply(AnalysisHelper.scala:89)
   		at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1.apply(AnalysisHelper.scala:86)
   		at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$.allowInvokingTransformsInAnalyzer(AnalysisHelper.scala:194)
   		at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$class.resolveOperatorsUp(AnalysisHelper.scala:86)
   		at org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.resolveOperatorsUp(LogicalPlan.scala:29)
   		at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1$$anonfun$1.apply(AnalysisHelper.scala:87)
   		at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1$$anonfun$1.apply(AnalysisHelper.scala:87)
   		at org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$4.apply(TreeNode.scala:326)
   		at org.apache.spark.sql.catalyst.trees.TreeNode.mapProductIterator(TreeNode.scala:187)
   		at org.apache.spark.sql.catalyst.trees.TreeNode.mapChildren(TreeNode.scala:324)
   		at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1.apply(AnalysisHelper.scala:87)
   		at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1.apply(AnalysisHelper.scala:86)
   		at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$.allowInvokingTransformsInAnalyzer(AnalysisHelper.scala:194)
   		at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$class.resolveOperatorsUp(AnalysisHelper.scala:86)
   		at org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.resolveOperatorsUp(LogicalPlan.scala:29)
   		at org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$.apply(Analyzer.scala:708)
   		at org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$.apply(Analyzer.scala:654)
   		at org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1$$anonfun$apply$1.apply(RuleExecutor.scala:87)
   		at org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1$$anonfun$apply$1.apply(RuleExecutor.scala:84)
   		at scala.collection.LinearSeqOptimized$class.foldLeft(LinearSeqOptimized.scala:124)
   		at scala.collection.immutable.List.foldLeft(List.scala:84)
   		at org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1.apply(RuleExecutor.scala:84)
   		at org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1.apply(RuleExecutor.scala:76)
   		at scala.collection.immutable.List.foreach(List.scala:392)
   		at org.apache.spark.sql.catalyst.rules.RuleExecutor.execute(RuleExecutor.scala:76)
   		at org.apache.spark.sql.catalyst.analysis.Analyzer.org$apache$spark$sql$catalyst$analysis$Analyzer$$executeSameContext(Analyzer.scala:127)
   		at org.apache.spark.sql.catalyst.analysis.Analyzer.execute(Analyzer.scala:121)
   		at org.apache.spark.sql.catalyst.analysis.Analyzer$$anonfun$executeAndCheck$1.apply(Analyzer.scala:106)
   [INFO] 2020-09-03 10:57:45.340  - [taskAppId=TASK-3-3-7]:[121] -  -> 	at org.apache.spark.sql.catalyst.analysis.Analyzer$$anonfun$executeAndCheck$1.apply(Analyzer.scala:105)
   		at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$.markInAnalyzer(AnalysisHelper.scala:201)
   		at org.apache.spark.sql.catalyst.analysis.Analyzer.executeAndCheck(Analyzer.scala:105)
   		at org.apache.spark.sql.execution.QueryExecution.analyzed$lzycompute(QueryExecution.scala:57)
   		at org.apache.spark.sql.execution.QueryExecution.analyzed(QueryExecution.scala:55)
   		at org.apache.spark.sql.execution.QueryExecution.assertAnalyzed(QueryExecution.scala:47)
   		at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:78)
   		at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:642)
   		at dsjzt.sparkSqlTest$.main(sparkSqlTest.scala:14)
   		at dsjzt.sparkSqlTest.main(sparkSqlTest.scala)
   		at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
   		at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
   		at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
   		at java.lang.reflect.Method.invoke(Method.java:483)
   		at org.apache.spark.deploy.JavaMainApplication.start(SparkApplication.scala:52)
   		at org.apache.spark.deploy.SparkSubmit.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:849)
   		at org.apache.spark.deploy.SparkSubmit.doRunMain$1(SparkSubmit.scala:167)
   		at org.apache.spark.deploy.SparkSubmit.submit(SparkSubmit.scala:195)
   		at org.apache.spark.deploy.SparkSubmit.doSubmit(SparkSubmit.scala:86)
   		at org.apache.spark.deploy.SparkSubmit$$anon$2.doSubmit(SparkSubmit.scala:924)
   		at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:933)
   		at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
   	Cleanup action completed
   	Thu Sep 03 10:57:45 CST 2020 Thread[main,5,main] Cleanup action starting
   	java.sql.SQLException: Failed to create database 'metastore_db', see the next exception for details.
   		at org.apache.derby.impl.jdbc.SQLExceptionFactory.getSQLException(Unknown Source)
   		at org.apache.derby.impl.jdbc.SQLExceptionFactory.getSQLException(Unknown Source)
   		at org.apache.derby.impl.jdbc.Util.seeNextException(Unknown Source)
   		at org.apache.derby.impl.jdbc.EmbedConnection.createDatabase(Unknown Source)
   		at org.apache.derby.impl.jdbc.EmbedConnection.<init>(Unknown Source)
   		at org.apache.derby.jdbc.InternalDriver.getNewEmbedConnection(Unknown Source)
   		at org.apache.derby.jdbc.InternalDriver.connect(Unknown Source)
   		at org.apache.derby.jdbc.InternalDriver.connect(Unknown Source)
   		at org.apache.derby.jdbc.AutoloadedDriver.connect(Unknown Source)
   		at java.sql.DriverManager.getConnection(DriverManager.java:664)
   		at java.sql.DriverManager.getConnection(DriverManager.java:208)
   		at com.jolbox.bonecp.BoneCP.obtainRawInternalConnection(BoneCP.java:361)
   		at com.jolbox.bonecp.BoneCP.<init>(BoneCP.java:416)
   		at com.jolbox.bonecp.BoneCPDataSource.getConnection(BoneCPDataSource.java:120)
   		at org.datanucleus.store.rdbms.ConnectionFactoryImpl$ManagedConnectionImpl.getConnection(ConnectionFactoryImpl.java:501)
   		at org.datanucleus.store.rdbms.RDBMSStoreManager.<init>(RDBMSStoreManager.java:298)
   		at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
   		at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
   		at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
   		at java.lang.reflect.Constructor.newInstance(Constructor.java:408)
   		at org.datanucleus.plugin.NonManagedPluginRegistry.createExecutableExtension(NonManagedPluginRegistry.java:631)
   		at org.datanucleus.plugin.PluginManager.createExecutableExtension(PluginManager.java:301)
   		at org.datanucleus.NucleusContext.createStoreManagerForProperties(NucleusContext.java:1187)
   		at org.datanucleus.NucleusContext.initialise(NucleusContext.java:356)
   		at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.freezeConfiguration(JDOPersistenceManagerFactory.java:775)
   		at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.createPersistenceManagerFactory(JDOPersistenceManagerFactory.java:333)
   		at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.getPersistenceManagerFactory(JDOPersistenceManagerFactory.java:202)
   		at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
   		at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
   		at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
   		at java.lang.reflect.Method.invoke(Method.java:483)
   		at javax.jdo.JDOHelper$16.run(JDOHelper.java:1965)
   		at java.security.AccessController.doPrivileged(Native Method)
   		at javax.jdo.JDOHelper.invoke(JDOHelper.java:1960)
   		at javax.jdo.JDOHelper.invokeGetPersistenceManagerFactoryOnImplementation(JDOHelper.java:1166)
   		at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:808)
   		at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:701)
   		at org.apache.hadoop.hive.metastore.ObjectStore.getPMF(ObjectStore.java:419)
   		at org.apache.hadoop.hive.metastore.ObjectStore.getPersistenceManager(ObjectStore.java:448)
   [INFO] 2020-09-03 10:57:45.340  - [taskAppId=TASK-3-3-7]:[121] -  -> 	at org.apache.hadoop.hive.metastore.ObjectStore.initialize(ObjectStore.java:343)
   		at org.apache.hadoop.hive.metastore.ObjectStore.setConf(ObjectStore.java:299)
   		at org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:73)
   		at org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:133)
   		at org.apache.hadoop.hive.metastore.RawStoreProxy.<init>(RawStoreProxy.java:60)
   		at org.apache.hadoop.hive.metastore.RawStoreProxy.getProxy(RawStoreProxy.java:69)
   		at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.newRawStore(HiveMetaStore.java:685)
   		at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMS(HiveMetaStore.java:663)
   		at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.createDefaultDB(HiveMetaStore.java:712)
   		at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.init(HiveMetaStore.java:511)
   		at org.apache.hadoop.hive.metastore.RetryingHMSHandler.<init>(RetryingHMSHandler.java:78)
   		at org.apache.hadoop.hive.metastore.RetryingHMSHandler.getProxy(RetryingHMSHandler.java:84)
   		at org.apache.hadoop.hive.metastore.HiveMetaStore.newRetryingHMSHandler(HiveMetaStore.java:6517)
   		at org.apache.hadoop.hive.metastore.HiveMetaStoreClient.<init>(HiveMetaStoreClient.java:207)
   		at org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient.<init>(SessionHiveMetaStoreClient.java:74)
   		at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
   		at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
   		at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
   		at java.lang.reflect.Constructor.newInstance(Constructor.java:408)
   		at org.apache.hadoop.hive.metastore.MetaStoreUtils.newInstance(MetaStoreUtils.java:1660)
   		at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.<init>(RetryingMetaStoreClient.java:67)
   		at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:82)
   		at org.apache.hadoop.hive.ql.metadata.Hive.createMetaStoreClient(Hive.java:3411)
   		at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3430)
   		at org.apache.hadoop.hive.ql.metadata.Hive.getAllFunctions(Hive.java:3655)
   		at org.apache.hadoop.hive.ql.metadata.Hive.reloadFunctions(Hive.java:231)
   		at org.apache.hadoop.hive.ql.metadata.Hive.registerAllFunctionsOnce(Hive.java:215)
   		at org.apache.hadoop.hive.ql.metadata.Hive.<init>(Hive.java:338)
   		at org.apache.hadoop.hive.ql.metadata.Hive.get(Hive.java:299)
   		at org.apache.hadoop.hive.ql.metadata.Hive.get(Hive.java:274)
   		at org.apache.spark.sql.hive.client.HiveClientImpl.org$apache$spark$sql$hive$client$HiveClientImpl$$client(HiveClientImpl.scala:246)
   		at org.apache.spark.sql.hive.client.HiveClientImpl$$anonfun$withHiveState$1.apply(HiveClientImpl.scala:268)
   		at org.apache.spark.sql.hive.client.HiveClientImpl.liftedTree1$1(HiveClientImpl.scala:213)
   		at org.apache.spark.sql.hive.client.HiveClientImpl.retryLocked(HiveClientImpl.scala:212)
   		at org.apache.spark.sql.hive.client.HiveClientImpl.withHiveState(HiveClientImpl.scala:258)
   		at org.apache.spark.sql.hive.client.HiveClientImpl.databaseExists(HiveClientImpl.scala:348)
   		at org.apache.spark.sql.hive.HiveExternalCatalog$$anonfun$databaseExists$1.apply$mcZ$sp(HiveExternalCatalog.scala:217)
   		at org.apache.spark.sql.hive.HiveExternalCatalog$$anonfun$databaseExists$1.apply(HiveExternalCatalog.scala:217)
   		at org.apache.spark.sql.hive.HiveExternalCatalog$$anonfun$databaseExists$1.apply(HiveExternalCatalog.scala:217)
   		at org.apache.spark.sql.hive.HiveExternalCatalog.withClient(HiveExternalCatalog.scala:99)
   		at org.apache.spark.sql.hive.HiveExternalCatalog.databaseExists(HiveExternalCatalog.scala:216)
   		at org.apache.spark.sql.internal.SharedState.externalCatalog$lzycompute(SharedState.scala:114)
   =================================================================================
   1:/tmp/dolphinscheduler已经给dolphinscheduler用户赋权过了
   2:mysql的jdbc包也放在对应的lib下了
   3:shell命令、hadoop命令、hive -e命令都是没问题的,可以跑
   请问还有什么其他的地方会影响到这个吗?


----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
users@infra.apache.org



[GitHub] [incubator-dolphinscheduler] xingchun-chen commented on issue #3654: [Question] 通过shell提交spark-submit报错:ERROR XBM0H: Directory /tmp/dolphinscheduler/exec/process/1/3/3/4/metastore_db cannot be created

Posted by GitBox <gi...@apache.org>.
xingchun-chen commented on issue #3654:
URL: https://github.com/apache/incubator-dolphinscheduler/issues/3654#issuecomment-720871234


   Thank you very much for your feedback, as the problem has been resolved, it will be closed


----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
users@infra.apache.org



[GitHub] [incubator-dolphinscheduler] AfterYesterday commented on issue #3654: [Question] 通过shell提交spark-submit报错:ERROR XBM0H: Directory /tmp/dolphinscheduler/exec/process/1/3/3/4/metastore_db cannot be created

Posted by GitBox <gi...@apache.org>.
AfterYesterday commented on issue #3654:
URL: https://github.com/apache/incubator-dolphinscheduler/issues/3654#issuecomment-687711075


   经过这两天的测试,dolphinscheduler不能通过shell调用带有连接hive的spark 脚本,是由于我将CDH5.16.1的spark升级到了2.4,
   (但是这个问题我在集群直接提交时,是没有的)。试了几种方案:
   1:分别尝试了dolphinscheduler1.3的1、2两个版本,无效;
   2:修改dolphinscheduler的pom文件,重新编译,使hive版本对应CDH5,无效;
   3:升级hive1.1到1.2版本(直接升级2.X版本失败,可能由于CDH版本有限制),无效,还有种通过parcels升级,还没有尝试(接下来会尝试一下);
   4:添加CDH5.16.1自带的spark1.6服务,通过该服务提交的脚本可以正常执行。
   以上方案4有效,方案3验证中
   考虑到大家对spark版本的需求,建议使用CDH6
   
   当前解决方案:
   1:使用spark1.6,不做改动;
   2:升级CDH5到6.X;


----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
users@infra.apache.org



[GitHub] [incubator-dolphinscheduler] xingchun-chen closed issue #3654: [Question] 通过shell提交spark-submit报错:ERROR XBM0H: Directory /tmp/dolphinscheduler/exec/process/1/3/3/4/metastore_db cannot be created

Posted by GitBox <gi...@apache.org>.
xingchun-chen closed issue #3654:
URL: https://github.com/apache/incubator-dolphinscheduler/issues/3654


   


----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
users@infra.apache.org



[GitHub] [incubator-dolphinscheduler] yy0812 commented on issue #3654: [Question] 通过shell提交spark-submit报错:ERROR XBM0H: Directory /tmp/dolphinscheduler/exec/process/1/3/3/4/metastore_db cannot be created

Posted by GitBox <gi...@apache.org>.
yy0812 commented on issue #3654:
URL: https://github.com/apache/incubator-dolphinscheduler/issues/3654#issuecomment-686226153


   What is the tenant number of the user to whom you created this task? When the task is executed, the tenant number is executed as the user


----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
users@infra.apache.org



[GitHub] [incubator-dolphinscheduler] AfterYesterday commented on issue #3654: [Question] 通过shell提交spark-submit报错:ERROR XBM0H: Directory /tmp/dolphinscheduler/exec/process/1/3/3/4/metastore_db cannot be created

Posted by GitBox <gi...@apache.org>.
AfterYesterday commented on issue #3654:
URL: https://github.com/apache/incubator-dolphinscheduler/issues/3654#issuecomment-686250557


   hello ,my  tenant number is 'hdfs' ,I think This is not a problem caused by permissions。
   I've tried a lot of approaches:like 'sudo chown -R dolphinscheduler:dolphinscheduler /tmp/dolphinscheduler' and 'chmod 777 /tmp/dolphinscheduler' ,but it`s not use
   


----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
users@infra.apache.org