You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Sital Kedia (JIRA)" <ji...@apache.org> on 2016/04/15 03:57:25 UTC
[jira] [Created] (SPARK-14649) DagScheduler runs duplicate tasks on
fetch failure
Sital Kedia created SPARK-14649:
-----------------------------------
Summary: DagScheduler runs duplicate tasks on fetch failure
Key: SPARK-14649
URL: https://issues.apache.org/jira/browse/SPARK-14649
Project: Spark
Issue Type: Bug
Components: Scheduler
Reporter: Sital Kedia
When running a job we found out that there are many duplicate tasks running after fetch failure in a stage. The issue is that when submitting tasks for a stage, the dag scheduler submits all the pending tasks (tasks whose output is not available). But out those pending tasks, some tasks might already be running on the cluster. The dag scheduler need to submit only non-running tasks for a stage.
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)
---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org