You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Dongjoon Hyun (JIRA)" <ji...@apache.org> on 2019/05/23 23:04:02 UTC

[jira] [Resolved] (SPARK-18673) Dataframes doesn't work on Hadoop 3.x; Hive rejects Hadoop version

     [ https://issues.apache.org/jira/browse/SPARK-18673?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

Dongjoon Hyun resolved SPARK-18673.
-----------------------------------
       Resolution: Fixed
         Assignee: Yuming Wang
    Fix Version/s: 3.0.0

This is superseded and resolved via SPARK-23710 . Please note that the following is the result of master branch with `hadoop-3.2` profile.

{code}
scala> spark.version
res0: String = 3.0.0-SNAPSHOT

scala> org.apache.hadoop.util.VersionInfo.getVersion
res1: String = 3.2.0

scala> org.apache.hadoop.hive.shims.ShimLoader.getMajorVersion
res2: String = 0.23

scala> spark.range(1).show()
+---+
| id|
+---+
|  0|
+---+
{code}

> Dataframes doesn't work on Hadoop 3.x; Hive rejects Hadoop version
> ------------------------------------------------------------------
>
>                 Key: SPARK-18673
>                 URL: https://issues.apache.org/jira/browse/SPARK-18673
>             Project: Spark
>          Issue Type: Bug
>          Components: SQL
>    Affects Versions: 2.1.0
>         Environment: Spark built with -Dhadoop.version=3.0.0-alpha2-SNAPSHOT 
>            Reporter: Steve Loughran
>            Assignee: Yuming Wang
>            Priority: Major
>             Fix For: 3.0.0
>
>
> Spark Dataframes fail to run on Hadoop 3.0.x, because hive.jar's shimloader considers 3.x to be an unknown Hadoop version.
> Hive itself will have to fix this; as Spark uses its own hive 1.2.x JAR, it will need to be updated to match.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org