You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hudi.apache.org by "sivabalan narayanan (Jira)" <ji...@apache.org> on 2022/04/11 03:20:00 UTC

[jira] [Assigned] (HUDI-3848) Restore fails when files pertaining to a commit has been cleaned up

     [ https://issues.apache.org/jira/browse/HUDI-3848?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

sivabalan narayanan reassigned HUDI-3848:
-----------------------------------------

    Assignee: sivabalan narayanan

> Restore fails when files pertaining to a commit has been cleaned up
> -------------------------------------------------------------------
>
>                 Key: HUDI-3848
>                 URL: https://issues.apache.org/jira/browse/HUDI-3848
>             Project: Apache Hudi
>          Issue Type: Bug
>            Reporter: sivabalan narayanan
>            Assignee: sivabalan narayanan
>            Priority: Major
>
> DC1, DC2, savepoint DC2, DC3, C4 (compaction), DC5, DC6, C7 (compaction). 
> DC8, Clean9 (this will clean up file slice 2 since file slice 1 is savepointed). Hence all files added as part of DC5, DC6 will be cleanedup. 
>  
> DC10. Restore to DC2. 
> Fails, bcoz, recently we moved to list based rollback. files to be deleted are fetched from commit metadata. in this case, DC5 and DC6 data files are already cleaned up and hence the failure. 
>  
> {code:java}
> 22/04/10 22:35:41 ERROR Executor: Exception in task 1.0 in stage 24.0 (TID 79)
> java.io.FileNotFoundException: File file:/tmp/hudi_trips_cow/asia/india/chennai/.4bd46734-6490-4efd-806f-eb7a2b8c36f6-0_20220410223029836.log.2_2-271-544 does not exist
>     at org.apache.hadoop.fs.RawLocalFileSystem.listStatus(RawLocalFileSystem.java:431)
>     at org.apache.hadoop.fs.FileSystem.listStatus(FileSystem.java:1517)
>     at org.apache.hadoop.fs.FileSystem.listStatus(FileSystem.java:1557)
>     at org.apache.hadoop.fs.ChecksumFileSystem.listStatus(ChecksumFileSystem.java:674)
>     at org.apache.hadoop.fs.FileSystem.listStatus(FileSystem.java:1517)
>     at org.apache.hadoop.fs.FileSystem.listStatus(FileSystem.java:1594)
>     at org.apache.hudi.common.fs.HoodieWrapperFileSystem.lambda$listStatus$21(HoodieWrapperFileSystem.java:595)
>     at org.apache.hudi.common.fs.HoodieWrapperFileSystem.executeFuncWithTimeMetrics(HoodieWrapperFileSystem.java:101)
>     at org.apache.hudi.common.fs.HoodieWrapperFileSystem.listStatus(HoodieWrapperFileSystem.java:594)
>     at org.apache.hudi.table.action.rollback.ListingBasedRollbackStrategy.fetchFilesFromCommitMetadata(ListingBasedRollbackStrategy.java:242)
>     at org.apache.hudi.table.action.rollback.ListingBasedRollbackStrategy.fetchFilesFromInstant(ListingBasedRollbackStrategy.java:228)
>     at org.apache.hudi.table.action.rollback.ListingBasedRollbackStrategy.lambda$getRollbackRequests$762d0ff4$1(ListingBasedRollbackStrategy.java:101)
>     at org.apache.hudi.client.common.HoodieSparkEngineContext.lambda$flatMap$7d470b86$1(HoodieSparkEngineContext.java:137)
>     at org.apache.spark.api.java.JavaRDDLike$$anonfun$fn$1$1.apply(JavaRDDLike.scala:125)
>     at scala.collection.Iterator$$anon$12.nextCur(Iterator.scala:435)
>     at scala.collection.Iterator$$anon$12.hasNext(Iterator.scala:441)
>     at scala.collection.Iterator$class.foreach(Iterator.scala:891)
>     at scala.collection.AbstractIterator.foreach(Iterator.scala:1334)
>     at scala.collection.generic.Growable$class.$plus$plus$eq(Growable.scala:59)
>     at scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:104)
>     at scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:48)
>     at scala.collection.TraversableOnce$class.to(TraversableOnce.scala:310)
>     at scala.collection.AbstractIterator.to(Iterator.scala:1334)
>     at scala.collection.TraversableOnce$class.toBuffer(TraversableOnce.scala:302)
>     at scala.collection.AbstractIterator.toBuffer(Iterator.scala:1334)
>     at scala.collection.TraversableOnce$class.toArray(TraversableOnce.scala:289)
>     at scala.collection.AbstractIterator.toArray(Iterator.scala:1334)
>     at org.apache.spark.rdd.RDD$$anonfun$collect$1$$anonfun$15.apply(RDD.scala:990)
>     at org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:2101)
>     at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:90)
>     at org.apache.spark.scheduler.Task.run(Task.scala:123)
>     at org.apache.spark.executor.Executor$TaskRunner$$anonfun$10.apply(Executor.scala:408)
>     at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1360)
>     at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:414)
>     at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
>     at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
>     at java.lang.Thread.run(Thread.java:748)
> 22/04/10 22:35:41 ERROR Executor: Exception in task 0.0 in stage 24.0 (TID 78)
> java.io.FileNotFoundException: File file:/tmp/hudi_trips_cow/americas/brazil/sao_paulo/.82cffa84-6218-4544-aedc-066f87b39a3e-0_20220410223029836.log.2_0-271-542 does not exist
>     at org.apache.hadoop.fs.RawLocalFileSystem.listStatus(RawLocalFileSystem.java:431)
>     at org.apache.hadoop.fs.FileSystem.listStatus(FileSystem.java:1517)
>     at org.apache.hadoop.fs.FileSystem.listStatus(FileSystem.java:1557)
>     at org.apache.hadoop.fs.ChecksumFileSystem.listStatus(ChecksumFileSystem.java:674)
>     at org.apache.hadoop.fs.FileSystem.listStatus(FileSystem.java:1517)
>     at org.apache.hadoop.fs.FileSystem.listStatus(FileSystem.java:1594)
>     at org.apache.hudi.common.fs.HoodieWrapperFileSystem.lambda$listStatus$21(HoodieWrapperFileSystem.java:595)
>     at org.apache.hudi.common.fs.HoodieWrapperFileSystem.executeFuncWithTimeMetrics(HoodieWrapperFileSystem.java:101)
>     at org.apache.hudi.common.fs.HoodieWrapperFileSystem.listStatus(HoodieWrapperFileSystem.java:594)
>     at org.apache.hudi.table.action.rollback.ListingBasedRollbackStrategy.fetchFilesFromCommitMetadata(ListingBasedRollbackStrategy.java:242)
>     at org.apache.hudi.table.action.rollback.ListingBasedRollbackStrategy.fetchFilesFromInstant(ListingBasedRollbackStrategy.java:228)
>     at org.apache.hudi.table.action.rollback.ListingBasedRollbackStrategy.lambda$getRollbackRequests$762d0ff4$1(ListingBasedRollbackStrategy.java:101)
>     at org.apache.hudi.client.common.HoodieSparkEngineContext.lambda$flatMap$7d470b86$1(HoodieSparkEngineContext.java:137)
>     at org.apache.spark.api.java.JavaRDDLike$$anonfun$fn$1$1.apply(JavaRDDLike.scala:125)
>     at scala.collection.Iterator$$anon$12.nextCur(Iterator.scala:435)
>     at scala.collection.Iterator$$anon$12.hasNext(Iterator.scala:441)
>     at scala.collection.Iterator$class.foreach(Iterator.scala:891)
>     at scala.collection.AbstractIterator.foreach(Iterator.scala:1334)
>     at scala.collection.generic.Growable$class.$plus$plus$eq(Growable.scala:59)
>     at scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:104)
>     at scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:48)
>     at scala.collection.TraversableOnce$class.to(TraversableOnce.scala:310)
>     at scala.collection.AbstractIterator.to(Iterator.scala:1334)
>     at scala.collection.TraversableOnce$class.toBuffer(TraversableOnce.scala:302)
>     at scala.collection.AbstractIterator.toBuffer(Iterator.scala:1334)
>     at scala.collection.TraversableOnce$class.toArray(TraversableOnce.scala:289)
>     at scala.collection.AbstractIterator.toArray(Iterator.scala:1334)
>     at org.apache.spark.rdd.RDD$$anonfun$collect$1$$anonfun$15.apply(RDD.scala:990)
>     at org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:2101)
>     at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:90)
>     at org.apache.spark.scheduler.Task.run(Task.scala:123)
>     at org.apache.spark.executor.Executor$TaskRunner$$anonfun$10.apply(Executor.scala:408)
>     at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1360)
>     at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:414)
>     at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
>     at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
>     at java.lang.Thread.run(Thread.java:748)
> 22/04/10 22:35:41 INFO TaskSetManager: Starting task 2.0 in stage 24.0 (TID 80, localhost, executor driver, partition 2, PROCESS_LOCAL, 7759 bytes)
> 22/04/10 22:35:41 INFO Executor: Running task 2.0 in stage 24.0 (TID 80)
> 22/04/10 22:35:41 WARN TaskSetManager: Lost task 1.0 in stage 24.0 (TID 79, localhost, executor driver): java.io.FileNotFoundException: File file:/tmp/hudi_trips_cow/asia/india/chennai/.4bd46734-6490-4efd-806f-eb7a2b8c36f6-0_20220410223029836.log.2_2-271-544 does not exist
>     at org.apache.hadoop.fs.RawLocalFileSystem.listStatus(RawLocalFileSystem.java:431)
>     at org.apache.hadoop.fs.FileSystem.listStatus(FileSystem.java:1517)
>     at org.apache.hadoop.fs.FileSystem.listStatus(FileSystem.java:1557)
>     at org.apache.hadoop.fs.ChecksumFileSystem.listStatus(ChecksumFileSystem.java:674)
>     at org.apache.hadoop.fs.FileSystem.listStatus(FileSystem.java:1517)
>     at org.apache.hadoop.fs.FileSystem.listStatus(FileSystem.java:1594)
>     at org.apache.hudi.common.fs.HoodieWrapperFileSystem.lambda$listStatus$21(HoodieWrapperFileSystem.java:595)
>     at org.apache.hudi.common.fs.HoodieWrapperFileSystem.executeFuncWithTimeMetrics(HoodieWrapperFileSystem.java:101)
>     at org.apache.hudi.common.fs.HoodieWrapperFileSystem.listStatus(HoodieWrapperFileSystem.java:594)
>     at org.apache.hudi.table.action.rollback.ListingBasedRollbackStrategy.fetchFilesFromCommitMetadata(ListingBasedRollbackStrategy.java:242)
>     at org.apache.hudi.table.action.rollback.ListingBasedRollbackStrategy.fetchFilesFromInstant(ListingBasedRollbackStrategy.java:228)
>     at org.apache.hudi.table.action.rollback.ListingBasedRollbackStrategy.lambda$getRollbackRequests$762d0ff4$1(ListingBasedRollbackStrategy.java:101)
>     at org.apache.hudi.client.common.HoodieSparkEngineContext.lambda$flatMap$7d470b86$1(HoodieSparkEngineContext.java:137)
>     at org.apache.spark.api.java.JavaRDDLike$$anonfun$fn$1$1.apply(JavaRDDLike.scala:125)
>     at scala.collection.Iterator$$anon$12.nextCur(Iterator.scala:435)
>     at scala.collection.Iterator$$anon$12.hasNext(Iterator.scala:441)
>     at scala.collection.Iterator$class.foreach(Iterator.scala:891)
>     at scala.collection.AbstractIterator.foreach(Iterator.scala:1334)
>     at scala.collection.generic.Growable$class.$plus$plus$eq(Growable.scala:59)
>     at scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:104)
>     at scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:48)
>     at scala.collection.TraversableOnce$class.to(TraversableOnce.scala:310)
>     at scala.collection.AbstractIterator.to(Iterator.scala:1334)
>     at scala.collection.TraversableOnce$class.toBuffer(TraversableOnce.scala:302)
>     at scala.collection.AbstractIterator.toBuffer(Iterator.scala:1334)
>     at scala.collection.TraversableOnce$class.toArray(TraversableOnce.scala:289)
>     at scala.collection.AbstractIterator.toArray(Iterator.scala:1334)
>     at org.apache.spark.rdd.RDD$$anonfun$collect$1$$anonfun$15.apply(RDD.scala:990)
>     at org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:2101)
>     at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:90)
>     at org.apache.spark.scheduler.Task.run(Task.scala:123)
>     at org.apache.spark.executor.Executor$TaskRunner$$anonfun$10.apply(Executor.scala:408)
>     at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1360)
>     at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:414)
>     at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
>     at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
>     at java.lang.Thread.run(Thread.java:748)
> hudi:hudi_trips_cow-> {code}
>  
>  
>  
>  



--
This message was sent by Atlassian Jira
(v8.20.1#820001)