You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Sital Kedia (JIRA)" <ji...@apache.org> on 2016/04/15 03:57:25 UTC

[jira] [Created] (SPARK-14649) DagScheduler runs duplicate tasks on fetch failure

Sital Kedia created SPARK-14649:
-----------------------------------

             Summary: DagScheduler runs duplicate tasks on fetch failure
                 Key: SPARK-14649
                 URL: https://issues.apache.org/jira/browse/SPARK-14649
             Project: Spark
          Issue Type: Bug
          Components: Scheduler
            Reporter: Sital Kedia


When running a job we found out that there are many duplicate tasks running after fetch failure in a stage. The issue is that when submitting tasks for a stage, the dag scheduler submits all the pending tasks (tasks whose output is not available). But out those pending tasks, some tasks might already be running on the cluster. The dag scheduler need to submit only non-running tasks for a stage. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org