You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Dongjoon Hyun (JIRA)" <ji...@apache.org> on 2016/07/11 03:05:10 UTC

[jira] [Comment Edited] (SPARK-16475) Broadcast Hint for SQL Queries

    [ https://issues.apache.org/jira/browse/SPARK-16475?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15370041#comment-15370041 ] 

Dongjoon Hyun edited comment on SPARK-16475 at 7/11/16 3:04 AM:
----------------------------------------------------------------

Of course. It's not finished yet.
I'm working with priority queue. Never mind about that.


was (Author: dongjoon):
Of course. It's not finished yet.
I'm working with priority queue. Never mind about that.

Or, you had better break up this into some sub issues.
I'm sure [~hvanhovell] can do the SQL Parser part very fast.

> Broadcast Hint for SQL Queries
> ------------------------------
>
>                 Key: SPARK-16475
>                 URL: https://issues.apache.org/jira/browse/SPARK-16475
>             Project: Spark
>          Issue Type: Improvement
>            Reporter: Reynold Xin
>         Attachments: BroadcastHintinSparkSQL.pdf
>
>
> Broadcast hint is a way for users to manually annotate a query and suggest to the query optimizer the join method. It is very useful when the query optimizer cannot make optimal decision with respect to join methods due to conservativeness or the lack of proper statistics.
> The DataFrame API has broadcast hint since Spark 1.5. However, we do not have an equivalent functionality in SQL queries. We propose adding Hive-style broadcast hint to Spark SQL.
> For more information, please see the attached document. One note about the doc: in addition to supporting "MAPJOIN", we should also support "BROADCASTJOIN".



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org