You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@hive.apache.org by "meiyoula (JIRA)" <ji...@apache.org> on 2016/11/22 01:57:58 UTC
[jira] [Created] (HIVE-15256) One session close and delete
resourceDir, cause others open failed
meiyoula created HIVE-15256:
-------------------------------
Summary: One session close and delete resourceDir, cause others open failed
Key: HIVE-15256
URL: https://issues.apache.org/jira/browse/HIVE-15256
Project: Hive
Issue Type: Bug
Components: Clients
Reporter: meiyoula
*resourceDir* is shared to clients. When one connected client closes, it will delete *resourceDir*. At the same time, other clients are opening session, they will failed.
Exception is below:
{quote}
Error opening session: | org.apache.hive.service.cli.thrift.ThriftCLIService.OpenSession(ThriftCLIService.java:536)
java.lang.RuntimeException: ExitCodeException exitCode=1: chmod: cannot access `/opt/huawei/Bigdata/tmp/spark/dlresources': No such file or directory
at org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.java:528)
at org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.java:477)
at org.apache.spark.sql.hive.client.ClientWrapper.<init>(ClientWrapper.scala:229)
at org.apache.spark.sql.hive.client.IsolatedClientLoader.createClient(IsolatedClientLoader.scala:191)
at org.apache.spark.sql.hive.client.ClientWrapper.newSession(ClientWrapper.scala:1053)
at org.apache.spark.sql.hive.HiveContext.newSession(HiveContext.scala:93)
at org.apache.spark.sql.hive.thriftserver.SparkSQLSessionManager.openSession(SparkSQLSessionManager.scala:89)
at org.apache.hive.service.cli.CLIService.openSessionWithImpersonation(CLIService.java:189)
at org.apache.hive.service.cli.thrift.ThriftCLIService.getSessionHandle(ThriftCLIService.java:654)
at org.apache.hive.service.cli.thrift.ThriftCLIService.OpenSession(ThriftCLIService.java:522)
at org.apache.hive.service.cli.thrift.TCLIService$Processor$OpenSession.getResult(TCLIService.java:1257)
at org.apache.hive.service.cli.thrift.TCLIService$Processor$OpenSession.getResult(TCLIService.java:1242)
at org.apache.thrift.ProcessFunction.process(ProcessFunction.java:39)
at org.apache.thrift.TBaseProcessor.process(TBaseProcessor.java:39)
at org.apache.hadoop.hive.thrift.HadoopThriftAuthBridge$Server$TUGIAssumingProcessor.process(HadoopThriftAuthBridge.java:690)
at org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:285)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at java.lang.Thread.run(Thread.java:745)
Caused by: ExitCodeException exitCode=1: chmod: cannot access `/opt/huawei/Bigdata/tmp/spark/dlresources': No such file or directory
at org.apache.hadoop.util.Shell.runCommand(Shell.java:561)
at org.apache.hadoop.util.Shell.run(Shell.java:472)
at org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:738)
at org.apache.hadoop.util.Shell.execCommand(Shell.java:831)
at org.apache.hadoop.util.Shell.execCommand(Shell.java:814)
at org.apache.hadoop.fs.RawLocalFileSystem.setPermission(RawLocalFileSystem.java:744)
at org.apache.hadoop.fs.RawLocalFileSystem.mkOneDirWithMode(RawLocalFileSystem.java:502)
at org.apache.hadoop.fs.RawLocalFileSystem.mkdirsWithOptionalPermission(RawLocalFileSystem.java:542)
at org.apache.hadoop.fs.RawLocalFileSystem.mkdirs(RawLocalFileSystem.java:520)
at org.apache.hadoop.fs.FilterFileSystem.mkdirs(FilterFileSystem.java:340)
at org.apache.hadoop.hive.ql.session.SessionState.createPath(SessionState.java:656)
at org.apache.hadoop.hive.ql.session.SessionState.createSessionDirs(SessionState.java:584)
at org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.java:514)
... 18 more
{quote}
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)