You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hudi.apache.org by "clp007 (via GitHub)" <gi...@apache.org> on 2023/02/16 10:07:26 UTC

[GitHub] [hudi] clp007 commented on issue #7960: [SUPPORT]

clp007 commented on issue #7960:
URL: https://github.com/apache/hudi/issues/7960#issuecomment-1432834115

   thanks!
    I got a new execption message after add  `--source-class org.apache.hudi.utilities.sources.ParquetDFSSource` ;
    i dont known why!
   
   ** Stacktrace **
   
   WARN org.apache.spark.scheduler.TaskSetManager: Lost task 0.0 in stage 3.0 (TID
    346) (cluster-855d-m.us-central1-c.c.transferred.internal executor 2): org.apache.hudi.exception
   .HoodieException: ts(Part -ts) field not found in record. Acceptable fields were :[_hoodie_commit
   _time, _hoodie_commit_seqno, _hoodie_record_key, _hoodie_partition_path, _hoodie_file_name, user_
   pseudo_id, advertising_id, event_timestamp, event_name, country, version, id, user_first_touch_ti
   mestamp, event_params, extension, user_properties]
           at org.apache.hudi.avro.HoodieAvroUtils.getNestedFieldVal(HoodieAvroUtils.java:530)      
           at org.apache.hudi.utilities.deltastreamer.DeltaSync.lambda$fetchFromSource$950260a0$1(De
   ltaSync.java:494)
           at org.apache.spark.api.java.JavaPairRDD$.$anonfun$toScalaFunction$1(JavaPairRDD.scala:10
   70)
           at scala.collection.Iterator$$anon$10.next(Iterator.scala:461)
           at scala.collection.Iterator$SliceIterator.next(Iterator.scala:273)
           at scala.collection.Iterator.foreach(Iterator.scala:943)
           at scala.collection.Iterator.foreach$(Iterator.scala:943)
           at scala.collection.AbstractIterator.foreach(Iterator.scala:1431)
           at scala.collection.generic.Growable.$plus$plus$eq(Growable.scala:62)
           at scala.collection.generic.Growable.$plus$plus$eq$(Growable.scala:53)
           at scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:105)
           at scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:49)
           at scala.collection.TraversableOnce.to(TraversableOnce.scala:366)
           at scala.collection.TraversableOnce.to$(TraversableOnce.scala:364)
           at scala.collection.AbstractIterator.to(Iterator.scala:1431)
           at scala.collection.TraversableOnce.toBuffer(TraversableOnce.scala:358)
           at scala.collection.TraversableOnce.toBuffer$(TraversableOnce.scala:358)
           at scala.collection.AbstractIterator.toBuffer(Iterator.scala:1431)
           at scala.collection.TraversableOnce.toArray(TraversableOnce.scala:345)
           at scala.collection.TraversableOnce.toArray$(TraversableOnce.scala:339)
           at scala.collection.AbstractIterator.toArray(Iterator.scala:1431)
           at org.apache.spark.rdd.RDD.$anonfun$take$2(RDD.scala:1449)
           at org.apache.spark.SparkContext.$anonfun$runJob$5(SparkContext.scala:2244)
           at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:90)
           at org.apache.spark.scheduler.Task.run(Task.scala:131)
           at org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$3(Executor.scala:498)      
           at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1439)
           at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:501)
           at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)       
           at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)       
           at java.lang.Thread.run(Thread.java:750)
   
   23/02/16 09:59:36 ERROR org.apache.spark.scheduler.TaskSetManager: Task 0 in stage 3.0 failed 4 t
   imes; aborting job
   23/02/16 09:59:36 ERROR org.apache.hudi.utilities.deltastreamer.HoodieDeltaStreamer: Got error ru
   nning delta sync once. Shutting down
   org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in stage 3.0 failed 4 t
   imes, most recent failure: Lost task 0.3 in stage 3.0 (TID 349) (cluster-855d-m.us-central1-c.c.t
   ransferred.internal executor 1): org.apache.hudi.exception.HoodieException: ts(Part -ts) field no
   t found in record. Acceptable fields were :[_hoodie_commit_time, _hoodie_commit_seqno, _hoodie_re
   cord_key, _hoodie_partition_path, _hoodie_file_name, user_pseudo_id, advertising_id, event_timest
   amp, event_name, country, version, id, user_first_touch_timestamp, event_params, extension, user_
   properties]
           at org.apache.hudi.avro.HoodieAvroUtils.getNestedFieldVal(HoodieAvroUtils.java:530)      
           at org.apache.hudi.utilities.deltastreamer.DeltaSync.lambda$fetchFromSource$950260a0$1(De
   ltaSync.java:494)
           at org.apache.spark.api.java.JavaPairRDD$.$anonfun$toScalaFunction$1(JavaPairRDD.scala:10
   70)
           at scala.collection.Iterator$$anon$10.next(Iterator.scala:461)
           at scala.collection.Iterator$SliceIterator.next(Iterator.scala:273)
           at scala.collection.Iterator.foreach(Iterator.scala:943)
           at scala.collection.Iterator.foreach$(Iterator.scala:943)
           at scala.collection.AbstractIterator.foreach(Iterator.scala:1431)
           at scala.collection.generic.Growable.$plus$plus$eq(Growable.scala:62)
           at scala.collection.generic.Growable.$plus$plus$eq$(Growable.scala:53)
           at scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:105)
           at scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:49)
           at scala.collection.TraversableOnce.to(TraversableOnce.scala:366)
           at scala.collection.TraversableOnce.to$(TraversableOnce.scala:364)
           at scala.collection.AbstractIterator.to(Iterator.scala:1431)
           at scala.collection.TraversableOnce.toBuffer(TraversableOnce.scala:358)
           at scala.collection.TraversableOnce.toBuffer$(TraversableOnce.scala:358)
           at scala.collection.AbstractIterator.toBuffer(Iterator.scala:1431)
           at scala.collection.TraversableOnce.toArray(TraversableOnce.scala:345)
           at scala.collection.TraversableOnce.toArray$(TraversableOnce.scala:339)
           at scala.collection.AbstractIterator.toArray(Iterator.scala:1431)
           at org.apache.spark.rdd.RDD.$anonfun$take$2(RDD.scala:1449)
           at org.apache.spark.SparkContext.$anonfun$runJob$5(SparkContext.scala:2244)
           at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:90)
           at org.apache.spark.scheduler.Task.run(Task.scala:131)
           at org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$3(Executor.scala:498)      
           at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1439)
           at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:501)
           at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)       
           at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)       
           at java.lang.Thread.run(Thread.java:750)


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: commits-unsubscribe@hudi.apache.org

For queries about this service, please contact Infrastructure at:
users@infra.apache.org