You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@spark.apache.org by va...@apache.org on 2019/01/24 22:17:54 UTC

[spark] branch branch-2.3 updated: [SPARK-26682][SQL] Use taskAttemptID instead of attemptNumber for Had…

This is an automated email from the ASF dual-hosted git repository.

vanzin pushed a commit to branch branch-2.3
in repository https://gitbox.apache.org/repos/asf/spark.git


The following commit(s) were added to refs/heads/branch-2.3 by this push:
     new ded902c  [SPARK-26682][SQL] Use taskAttemptID instead of attemptNumber for Had…
ded902c is described below

commit ded902c3a90a9340e551091d554245df5982590c
Author: Ryan Blue <bl...@apache.org>
AuthorDate: Thu Jan 24 14:17:38 2019 -0800

    [SPARK-26682][SQL] Use taskAttemptID instead of attemptNumber for Had…
    
    ## What changes were proposed in this pull request?
    
    Updates the attempt ID used by FileFormatWriter. Tasks in stage attempts use the same task attempt number and could conflict. Using Spark's task attempt ID guarantees that Hadoop TaskAttemptID instances are unique.
    
    This is a backport of d5a97c1 to the 2.3 branch.
    
    ## How was this patch tested?
    
    Existing tests. Also validated that we no longer detect this failure case in our logs after deployment.
    
    Closes #23640 from rdblue/SPARK-26682-backport-to-2.3.
    
    Authored-by: Ryan Blue <bl...@apache.org>
    Signed-off-by: Marcelo Vanzin <va...@cloudera.com>
---
 .../org/apache/spark/sql/execution/datasources/FileFormatWriter.scala   | 2 +-
 1 file changed, 1 insertion(+), 1 deletion(-)

diff --git a/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/FileFormatWriter.scala b/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/FileFormatWriter.scala
index 1d80a69..2f701ed 100644
--- a/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/FileFormatWriter.scala
+++ b/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/FileFormatWriter.scala
@@ -198,7 +198,7 @@ object FileFormatWriter extends Logging {
             description = description,
             sparkStageId = taskContext.stageId(),
             sparkPartitionId = taskContext.partitionId(),
-            sparkAttemptNumber = taskContext.attemptNumber(),
+            sparkAttemptNumber = taskContext.taskAttemptId().toInt & Integer.MAX_VALUE,
             committer,
             iterator = iter)
         },


---------------------------------------------------------------------
To unsubscribe, e-mail: commits-unsubscribe@spark.apache.org
For additional commands, e-mail: commits-help@spark.apache.org