You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Herman van Hovell (JIRA)" <ji...@apache.org> on 2016/11/16 20:04:58 UTC
[jira] [Commented] (SPARK-16795) Spark's HiveThriftServer should be
able to use multiple sqlContexts
[ https://issues.apache.org/jira/browse/SPARK-16795?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15671462#comment-15671462 ]
Herman van Hovell commented on SPARK-16795:
-------------------------------------------
Spark uses one Hive client per spark context. This is known to cause performance issue in a highly concurrent environment because multiple sessions are contending for a single Hive client. This is a duplicate of SPARK-14003, and I am closing this one as such.
> Spark's HiveThriftServer should be able to use multiple sqlContexts
> -------------------------------------------------------------------
>
> Key: SPARK-16795
> URL: https://issues.apache.org/jira/browse/SPARK-16795
> Project: Spark
> Issue Type: Wish
> Affects Versions: 2.0.0
> Reporter: Furcy Pin
>
> It seems that when sending multiple Hive queries to the thrift server, the server cannot parallelize the query plannings, because
> it uses only one global sqlContext.
> This make the server very inefficient at handling many small concurrent queries.
> It would be nice to have it use a pool of sqlContexts instead, with a configurable maximum number of contexts.
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)
---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org