Spark Stremaing Hive Dynamic Partitions Issue

Previous Topic Next Topic
 
classic Classic list List threaded Threaded
1 message Options
Reply | Threaded
Open this post in threaded view
|

Spark Stremaing Hive Dynamic Partitions Issue

khajaasmath786
Hi,

I am able to wirte data into hive tables from spark stremaing. Job ran successfully for 37 hours and I started getting errors in task failure as below. Hive table has data too untill tasks are failed.

Job aborted due to stage failure: Task 0 in stage 691.0 failed 4 times, most recent failure: Lost task 0.3 in stage 691.0 (TID 10884, brksvl171.brk.navistar.com, executor 2): org.apache.spark.SparkException: Task failed while writing rows.+details

Job aborted due to stage failure: Task 0 in stage 691.0 failed 4 times, most recent failure: Lost task 0.3 in stage 691.0 (TID 10884, brksvl171.brk.navistar.com, executor 2): org.apache.spark.SparkException: Task failed while writing rows.

         at org.apache.spark.sql.hive.SparkHiveDynamicPartitionWriterContainer.writeToFile(hiveWriterContainers.scala:328)

         at org.apache.spark.sql.hive.execution.InsertIntoHiveTable$$anonfun$saveAsHiveFile$3.apply(InsertIntoHiveTable.scala:210)

         at org.apache.spark.sql.hive.execution.InsertIntoHiveTable$$anonfun$saveAsHiveFile$3.apply(InsertIntoHiveTable.scala:210)

         at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:87)

         at org.apache.spark.scheduler.Task.run(Task.scala:99)

         at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:322)

         at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)

         at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)

         at java.lang.Thread.run(Thread.java:745)

Caused by: java.lang.NullPointerException

         at parquet.hadoop.InternalParquetRecordWriter.flushRowGroupToStore(InternalParquetRecordWriter.java:152)

         at parquet.hadoop.InternalParquetRecordWriter.close(InternalParquetRecordWriter.java:111)

         at parquet.hadoop.ParquetRecordWriter.close(ParquetRecordWriter.java:112)

         at org.apache.hadoop.hive.ql.io.parquet.write.ParquetRecordWriterWrapper.close(ParquetRecordWriterWrapper.java:102)

         at org.apache.hadoop.hive.ql.io.parquet.write.ParquetRecordWriterWrapper.close(ParquetRecordWriterWrapper.java:119)

         at org.apache.spark.sql.hive.SparkHiveDynamicPartitionWriterContainer.writeToFile(hiveWriterContainers.scala:320)

         ... 8 more

 

Driver stacktrace:


any solution for this please?


Thanks,

Asmath