You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Apache Spark (JIRA)" <ji...@apache.org> on 2018/04/17 06:34:00 UTC

[jira] [Commented] (SPARK-23564) the optimized logical plan about Left anti join should be further optimization

    [ https://issues.apache.org/jira/browse/SPARK-23564?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16440476#comment-16440476 ] 

Apache Spark commented on SPARK-23564:
--------------------------------------

User 'cloud-fan' has created a pull request for this issue:
https://github.com/apache/spark/pull/21083

> the  optimized logical plan about Left anti join should be further optimization
> -------------------------------------------------------------------------------
>
>                 Key: SPARK-23564
>                 URL: https://issues.apache.org/jira/browse/SPARK-23564
>             Project: Spark
>          Issue Type: Improvement
>          Components: SQL
>    Affects Versions: 2.3.0
>            Reporter: KaiXinXIaoLei
>            Priority: Major
>
> The Optimized Logical Plan of the query '*select * from tt1 left anti join tt2 on tt2.i = tt1.i*' is 
>  
> {code:java}
> == Optimized Logical Plan ==
> Join LeftAnti, (i#2 = i#0)
> :- HiveTableRelation `default`.`tt1`, org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe, [i#0, s#1]
> +- Project [i#2]
> +- HiveTableRelation `default`.`tt2`, org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe, [i#2, s#3]
> {code}
>  
>  
> this plan can be further optimization by 'Filter isnotnull' of right table, as follow:
> {code:java}
> == Optimized Logical Plan ==
> Join LeftAnti, (i#2 = i#0)
> :- HiveTableRelation `default`.`tt1`, org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe, [i#0, s#1]
> +- Project [i#2]
>   +- Filter isnotnull(i#3)
>     +- HiveTableRelation `default`.`tt2`, org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe, [i#2, s#3]
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org