You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@hive.apache.org by Tao Li <tl...@hortonworks.com> on 2016/08/23 00:53:16 UTC

Question about Hive on Spark

Hi,

Looks like the Spark client (SparkClientImpl class) submits Sparks jobs to the YARN cluster by forking a process and kicking off spark-submit script. Are we provisioning new containers every time we submit a job? There could be a perf hit by doing that.

Thanks.

Re: Question about Hive on Spark

Posted by Xuefu Zhang <xu...@uber.com>.
That happens per session or if certain configs are changed in the session.

On Mon, Aug 22, 2016 at 5:53 PM, Tao Li <tl...@hortonworks.com> wrote:

> Hi,
>
> Looks like the Spark client (SparkClientImpl class) submits Sparks jobs to
> the YARN cluster by forking a process and kicking off spark-submit script.
> Are we provisioning new containers every time we submit a job? There could
> be a perf hit by doing that.
>
> Thanks.
>