You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Wenchen Fan (Jira)" <ji...@apache.org> on 2020/02/26 18:02:00 UTC

[jira] [Resolved] (SPARK-27619) MapType should be prohibited in hash expressions

     [ https://issues.apache.org/jira/browse/SPARK-27619?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

Wenchen Fan resolved SPARK-27619.
---------------------------------
    Fix Version/s: 3.0.0
         Assignee: Rakesh Raushan
       Resolution: Fixed

> MapType should be prohibited in hash expressions
> ------------------------------------------------
>
>                 Key: SPARK-27619
>                 URL: https://issues.apache.org/jira/browse/SPARK-27619
>             Project: Spark
>          Issue Type: Bug
>          Components: SQL
>    Affects Versions: 2.3.4, 2.4.0, 2.4.1, 2.4.2, 2.4.3, 2.4.4, 3.0.0
>            Reporter: Josh Rosen
>            Assignee: Rakesh Raushan
>            Priority: Blocker
>              Labels: correctness
>             Fix For: 3.0.0
>
>
> Spark currently allows MapType expressions to be used as input to hash expressions, but I think that this should be prohibited because Spark SQL does not support map equality.
> Currently, Spark SQL's map hashcodes are sensitive to the insertion order of map elements:
> {code:java}
> val a = spark.createDataset(Map(1->1, 2->2) :: Nil)
> val b = spark.createDataset(Map(2->2, 1->1) :: Nil)
> // Demonstration of how Scala Map equality is unaffected by insertion order:
> assert(Map(1->1, 2->2).hashCode() == Map(2->2, 1->1).hashCode())
> assert(Map(1->1, 2->2) == Map(2->2, 1->1))
> assert(a.first() == b.first())
> // In contrast, this will print two different hashcodes:
> println(Seq(a, b).map(_.selectExpr("hash(*)").first())){code}
> This behavior might be surprising to Scala developers.
> I think there's precedence for banning the use of MapType here because we already prohibit MapType in aggregation / joins / equality comparisons (SPARK-9415) and set operations (SPARK-19893).
> If we decide that we want this to be an error then it might also be a good idea to add a {{spark.sql.legacy}} flag as an escape-hatch to re-enable the old and buggy behavior (in case applications were relying on it in cases where it just so happens to be safe-by-accident (e.g. maps which only have one entry)).
> Alternatively, we could support hashing here if we implemented support for comparable map types (SPARK-18134).



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org