You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Herman van Hovell (JIRA)" <ji...@apache.org> on 2016/11/16 20:04:58 UTC

[jira] [Commented] (SPARK-16795) Spark's HiveThriftServer should be able to use multiple sqlContexts

    [ https://issues.apache.org/jira/browse/SPARK-16795?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15671462#comment-15671462 ] 

Herman van Hovell commented on SPARK-16795:
-------------------------------------------

Spark uses one Hive client per spark context. This is known to cause performance issue in a highly concurrent environment because multiple sessions are contending for a single Hive client. This is a duplicate of SPARK-14003, and I am closing this one as such. 

> Spark's HiveThriftServer should be able to use multiple sqlContexts
> -------------------------------------------------------------------
>
>                 Key: SPARK-16795
>                 URL: https://issues.apache.org/jira/browse/SPARK-16795
>             Project: Spark
>          Issue Type: Wish
>    Affects Versions: 2.0.0
>            Reporter: Furcy Pin
>
> It seems that when sending multiple Hive queries to the thrift server, the server cannot parallelize the query plannings, because
> it uses only one global sqlContext.
> This make the server very inefficient at handling many small concurrent queries.
> It would be nice to have it use a pool of sqlContexts instead, with a configurable maximum number of contexts.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org