You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Ari Gesher (JIRA)" <ji...@apache.org> on 2017/02/28 05:46:45 UTC

[jira] [Created] (SPARK-19764) Executors hang with supposedly running task that are really finished.

Ari Gesher created SPARK-19764:
----------------------------------

             Summary: Executors hang with supposedly running task that are really finished.
                 Key: SPARK-19764
                 URL: https://issues.apache.org/jira/browse/SPARK-19764
             Project: Spark
          Issue Type: Bug
          Components: PySpark, Spark Core
    Affects Versions: 2.0.2
         Environment: Ubuntu 16.04 LTS
OpenJDK Runtime Environment (build 1.8.0_121-8u121-b13-0ubuntu1.16.04.2-b13)
Spark 2.0.2 - Spark Cluster Manager
            Reporter: Ari Gesher
         Attachments: executor-2.log

We've come across a job that won't finish.  Running on a six-node cluster, each of the executors end up with 5-7 tasks that are never marked as completed.

Here's an excerpt from the web UI:

||Index  ▴||ID||Attempt||Status||Locality Level||Executor ID / Host||Launch Time||Duration||Scheduler Delay||Task Deserialization Time||GC Time||Result Serialization Time||Getting Result Time||Peak Execution Memory||Shuffle Read Size / Records||Errors||
|105	| 1131	| 0	| SUCCESS	|PROCESS_LOCAL	|4 / 172.31.24.171 |	2017/02/27 22:51:36 |	1.9 min |	9 ms |	4 ms |	0.7 s |	2 ms|	6 ms|	384.1 MB| 	90.3 MB / 572	| |
|106|	1168|	0|	RUNNING	|ANY|	2 / 172.31.16.112|	2017/02/27 22:53:25|	6.5 h	|0 ms|	0 ms|	1 s	|0 ms|	0 ms|	|384.1 MB	|98.7 MB / 624 | |	

However, the Executor reports the task as finished: 
{noformat}
17/02/27 22:53:25 INFO Executor: Running task 106.0 in stage 5.0 (TID 1168)
17/02/27 22:55:29 INFO Executor: Finished task 106.0 in stage 5.0 (TID 1168). 2633558 bytes result sent via BlockManager)
{noformat}


Full log from this executor attached.




--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org