You are viewing a plain text version of this content. The canonical link for it is here.
Posted to user@spark.apache.org by haopu <hw...@qilinsoft.com> on 2014/07/09 04:37:34 UTC

slower worker node in the cluster

In a standalone cluster, is there way to specify the stage to be running on a
faster worker?

That stage is reading HDFS file and then doing some filter operations.  The
tasks are assigned to the slower worker also, but the slower worker delays
to launch because it's running some tasks from other stages.

So I think it may be better to assign stage to a worker. Any suggestions?

And will the cluster on Yarn help?



--
View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/slower-worker-node-in-the-cluster-tp9125.html
Sent from the Apache Spark User List mailing list archive at Nabble.com.

Re: slower worker node in the cluster

Posted by Chris Fregly <ch...@fregly.com>.
perhaps creating Fair Scheduler Pools might help?  there's no way to pin
certain nodes to a pool, but you can specify minShares (cpu's).  not sure
if that would help, but worth looking in to.


On Tue, Jul 8, 2014 at 7:37 PM, haopu <hw...@qilinsoft.com> wrote:

> In a standalone cluster, is there way to specify the stage to be running
> on a
> faster worker?
>
> That stage is reading HDFS file and then doing some filter operations.  The
> tasks are assigned to the slower worker also, but the slower worker delays
> to launch because it's running some tasks from other stages.
>
> So I think it may be better to assign stage to a worker. Any suggestions?
>
> And will the cluster on Yarn help?
>
>
>
> --
> View this message in context:
> http://apache-spark-user-list.1001560.n3.nabble.com/slower-worker-node-in-the-cluster-tp9125.html
> Sent from the Apache Spark User List mailing list archive at Nabble.com.
>