You are viewing a plain text version of this content. The canonical link for it is here.
Posted to user@spark.apache.org by KhajaAsmath Mohammed <md...@gmail.com> on 2017/11/22 16:06:23 UTC

Spark Stremaing Hive Dynamic Partitions Issue

Hi,

I am able to wirte data into hive tables from spark stremaing. Job ran
successfully for 37 hours and I started getting errors in task failure as
below. Hive table has data too untill tasks are failed.

Job aborted due to stage failure: Task 0 in stage 691.0 failed 4 times,
most recent failure: Lost task 0.3 in stage 691.0 (TID 10884,
brksvl171.brk.navistar.com, executor 2): org.apache.spark.SparkException:
Task failed while writing rows.+details

Job aborted due to stage failure: Task 0 in stage 691.0 failed 4 times,
most recent failure: Lost task 0.3 in stage 691.0 (TID 10884,
brksvl171.brk.navistar.com, executor 2): org.apache.spark.SparkException:
Task failed while writing rows.

         at
org.apache.spark.sql.hive.SparkHiveDynamicPartitionWriterContainer.writeToFile(hiveWriterContainers.scala:328)

         at
org.apache.spark.sql.hive.execution.InsertIntoHiveTable$$anonfun$saveAsHiveFile$3.apply(InsertIntoHiveTable.scala:210)

         at
org.apache.spark.sql.hive.execution.InsertIntoHiveTable$$anonfun$saveAsHiveFile$3.apply(InsertIntoHiveTable.scala:210)

         at
org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:87)

         at org.apache.spark.scheduler.Task.run(Task.scala:99)

         at
org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:322)

         at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)

         at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)

         at java.lang.Thread.run(Thread.java:745)

Caused by: java.lang.NullPointerException

         at
parquet.hadoop.InternalParquetRecordWriter.flushRowGroupToStore(InternalParquetRecordWriter.java:152)

         at
parquet.hadoop.InternalParquetRecordWriter.close(InternalParquetRecordWriter.java:111)

         at
parquet.hadoop.ParquetRecordWriter.close(ParquetRecordWriter.java:112)

         at
org.apache.hadoop.hive.ql.io.parquet.write.ParquetRecordWriterWrapper.close(ParquetRecordWriterWrapper.java:102)

         at
org.apache.hadoop.hive.ql.io.parquet.write.ParquetRecordWriterWrapper.close(ParquetRecordWriterWrapper.java:119)

         at
org.apache.spark.sql.hive.SparkHiveDynamicPartitionWriterContainer.writeToFile(hiveWriterContainers.scala:320)

         ... 8 more



Driver stacktrace:


any solution for this please?


Thanks,

Asmath