You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "WEI ZHU (JIRA)" <ji...@apache.org> on 2015/12/03 06:42:11 UTC
[jira] [Commented] (SPARK-10528) spark-shell throws
java.lang.RuntimeException: The root scratch dir: /tmp/hive on HDFS should
be writable.
[ https://issues.apache.org/jira/browse/SPARK-10528?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15037296#comment-15037296 ]
WEI ZHU commented on SPARK-10528:
---------------------------------
I came to the same problem when using precompiled binaries - 1.5.2-hadoop 2.6
when I run ./spark-shell, it throws the following exception. and I found the ./pyspark working well for me. is this a bug?
hive.metastore.schema.verification is not enabled so recording the schema version 1.2.0
15/12/03 02:04:08 WARN ObjectStore: Failed to get database default, returning NoSuchObjectException
java.lang.RuntimeException: java.lang.RuntimeException: The root scratch dir: /tmp/hive on HDFS should be writable. Current permissions are: rwx------
at org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.java:522)
at org.apache.spark.sql.hive.client.ClientWrapper.<init>(ClientWrapper.scala:171)
at org.apache.spark.sql.hive.HiveContext.executionHive$lzycompute(HiveContext.scala:162)
at org.apache.spark.sql.hive.HiveContext.executionHive(HiveContext.scala:160)
at org.apache.spark.sql.hive.HiveContext.<init>(HiveContext.scala:167)
at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57)
at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
p.s. the permissions of hdfs folder look okay to me as below.
[root@sandbox bin]# hadoop fs -ls /tmp
Found 3 items
drwxrwxrwx - hive hdfs 0 2015-12-01 05:35 /tmp/hive
drwxrwxrwx - root hdfs 0 2015-08-12 00:37 /tmp/temp-1409396078
drwxrwxrwx - hdfs hdfs 0 2014-12-16 19:30 /tmp/udfs
> spark-shell throws java.lang.RuntimeException: The root scratch dir: /tmp/hive on HDFS should be writable.
> ----------------------------------------------------------------------------------------------------------
>
> Key: SPARK-10528
> URL: https://issues.apache.org/jira/browse/SPARK-10528
> Project: Spark
> Issue Type: Bug
> Components: Spark Shell
> Affects Versions: 1.5.0
> Environment: Windows 7 x64
> Reporter: Aliaksei Belablotski
> Priority: Minor
>
> Starting spark-shell throws
> java.lang.RuntimeException: java.lang.RuntimeException: The root scratch dir: /tmp/hive on HDFS should be writable. Current permissions are: rw-rw-rw-
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)
---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org