You are viewing a plain text version of this content. The canonical link for it is here.
Posted to user@spark.apache.org by map reduced <k3...@gmail.com> on 2016/10/06 22:09:32 UTC

Zombie Driver process (Standalone Cluster)

Hi,

I am noticing zombie driver processes running on my standalone cluster. Not
sure about the reason - but node (with driver on it) restart may be a
potential cause.
What's interesting about these is: SparkUI doesn't recognize it as a
running driver, hence no 'kill' option there. Also, if I try running
command line driver kill command, it says no such driver found running.

bin/spark-class org.apache.spark.deploy.Client kill
spark://master-url:31498 driver-20161005204049-0031

Gives:
Driver driver-20161005204049-0031 has already finished or does not exist.

Here's another screenshot which shows that Spark Master UI doesn't find it
as a running driver:

[image: Inline image 1]
But it's really running on one of the workers:

[image: Inline image 2]
(Sorry can't share hostnames etc).

Any idea why this is happening and how to kill it? (I don't have a way to
ssh into that machine).

Thanks,
KP