You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@hive.apache.org by "ASF GitHub Bot (Jira)" <ji...@apache.org> on 2020/06/06 00:26:01 UTC

[jira] [Work logged] (HIVE-22757) NullPointerException when executing SQLs

     [ https://issues.apache.org/jira/browse/HIVE-22757?focusedWorklogId=442124&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-442124 ]

ASF GitHub Bot logged work on HIVE-22757:
-----------------------------------------

                Author: ASF GitHub Bot
            Created on: 06/Jun/20 00:25
            Start Date: 06/Jun/20 00:25
    Worklog Time Spent: 10m 
      Work Description: github-actions[bot] commented on pull request #885:
URL: https://github.com/apache/hive/pull/885#issuecomment-639914342


   This pull request has been automatically marked as stale because it has not had recent activity. It will be closed if no further activity occurs.
   Feel free to reach out on the dev@hive.apache.org list if the patch is in need of reviews.


----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
users@infra.apache.org


Issue Time Tracking
-------------------

            Worklog Id:     (was: 442124)
    Remaining Estimate: 0h
            Time Spent: 10m

> NullPointerException when executing SQLs
> ----------------------------------------
>
>                 Key: HIVE-22757
>                 URL: https://issues.apache.org/jira/browse/HIVE-22757
>             Project: Hive
>          Issue Type: Bug
>    Affects Versions: 2.3.6
>            Reporter: Deegue
>            Priority: Major
>         Attachments: HIVE-22757.patch
>
>          Time Spent: 10m
>  Remaining Estimate: 0h
>
> When executing SQL:
> {code:java}
> insert overwrite table ods.ods_1 partition(stat_day='20191209')
> select
> id
> ,user_id
> ,teacher_user_id
> ,partner_user_id
> ,order_id
> ,barcode
> ,sub_order_id
> ,item_id
> ,sales
> ,refund
> ,teacher_profit
> ,partner_profit
> ,teacher_refund_profit
> ,partner_refund_profit
> ,teacher_commission_value
> ,partner_commission_value
> ,biz_type
> ,pay_time
> ,item_profit_type
> ,black_mark
> ,is_deleted
> ,create_time
> ,modify_time
> from src.src_1
> where partition_date='20191209'
> union all
> select
> t1.id
> ,t1.user_id
> ,t1.teacher_user_id
> ,t1.partner_user_id
> ,t1.order_id
> ,t1.barcode
> ,t1.sub_order_id
> ,t1.item_id
> ,t1.sales
> ,t1.refund
> ,t1.teacher_profit
> ,t1.partner_profit
> ,t1.teacher_refund_profit
> ,t1.partner_refund_profit
> ,t1.teacher_commission_value
> ,t1.partner_commission_value
> ,t1.biz_type
> ,t1.pay_time
> ,t1.item_profit_type
> ,t1.black_mark
> ,t1.is_deleted
> ,t1.create_time
> ,t1.modify_time
> from
> (select *
> from ods.ods_1
> where stat_day='20191208'
> ) t1
> left join
> ( select order_id
> ,sub_order_id
> from src.src_1
> where partition_date='20191209'
> ) t2
> on t1.order_id=t2.order_id
> and t1.sub_order_id=t2.sub_order_id
> where t2.order_id is null
> {code}
> `java.lang.NullPointerException` thrown because the array list `neededNestedColumnPaths` haven't been inited when `addAll` method is invoked.
> {code:java}
> Launching Job 5 out of 5
> Number of reduce tasks is set to 0 since there's no reduce operator
> Starting Job = job_1566481621886_4925755, Tracking URL = http://TXIDC65-bigdata-resourcemanager1:8042/proxy/application_1566481621886_4925755/
> Kill Command = /usr/local/yunji/hadoop/bin/hadoop job  -kill job_1566481621886_4925755
> Hadoop job information for Stage-4: number of mappers: 1; number of reducers: 0
> 2019-12-24 16:00:40,584 Stage-4 map = 0%,  reduce = 0%
> 2019-12-24 16:01:40,956 Stage-4 map = 0%,  reduce = 0%
> 2019-12-24 16:02:41,451 Stage-4 map = 0%,  reduce = 0%
> 2019-12-24 16:02:45,550 Stage-4 map = 100%,  reduce = 0%
> Ended Job = job_1566481621886_4925755 with errors
> Error during job, obtaining debugging information...
> Examining task ID: task_1566481621886_4925755_m_000000 (and more) from job job_1566481621886_4925755
> Task with the most failures(4):
> -----
> Task ID:
>   task_1566481621886_4925755_m_000000
> URL:
>   http://TXIDC65-bigdata-resourcemanager1:8088/taskdetails.jsp?jobid=job_1566481621886_4925755&tipid=task_1566481621886_4925755_m_000000
> -----
> Diagnostic Messages for this Task:
> Error: java.io.IOException: java.lang.reflect.InvocationTargetException
> 	at org.apache.hadoop.hive.io.HiveIOExceptionHandlerChain.handleRecordReaderCreationException(HiveIOExceptionHandlerChain.java:97)
> 	at org.apache.hadoop.hive.io.HiveIOExceptionHandlerUtil.handleRecordReaderCreationException(HiveIOExceptionHandlerUtil.java:57)
> 	at org.apache.hadoop.hive.shims.HadoopShimsSecure$CombineFileRecordReader.initNextRecordReader(HadoopShimsSecure.java:271)
> 	at org.apache.hadoop.hive.shims.HadoopShimsSecure$CombineFileRecordReader.<init>(HadoopShimsSecure.java:217)
> 	at org.apache.hadoop.hive.shims.HadoopShimsSecure$CombineFileInputFormatShim.getRecordReader(HadoopShimsSecure.java:345)
> 	at org.apache.hadoop.hive.ql.io.CombineHiveInputFormat.getRecordReader(CombineHiveInputFormat.java:695)
> 	at org.apache.hadoop.mapred.MapTask$TrackedRecordReader.<init>(MapTask.java:169)
> 	at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:438)
> 	at org.apache.hadoop.mapred.MapTask.run(MapTask.java:343)
> 	at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:164)
> 	at java.security.AccessController.doPrivileged(Native Method)
> 	at javax.security.auth.Subject.doAs(Subject.java:422)
> 	at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1924)
> 	at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:158)
> Caused by: java.lang.reflect.InvocationTargetException
> 	at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
> 	at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
> 	at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
> 	at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
> 	at org.apache.hadoop.hive.shims.HadoopShimsSecure$CombineFileRecordReader.initNextRecordReader(HadoopShimsSecure.java:257)
> 	... 11 more
> Caused by: java.lang.NullPointerException
> 	at java.util.AbstractCollection.addAll(AbstractCollection.java:343)
> 	at org.apache.hadoop.hive.ql.io.parquet.ProjectionPusher.pushProjectionsAndFilters(ProjectionPusher.java:118)
> 	at org.apache.hadoop.hive.ql.io.parquet.ProjectionPusher.pushProjectionsAndFilters(ProjectionPusher.java:189)
> 	at org.apache.hadoop.hive.ql.io.parquet.ParquetRecordReaderBase.getSplit(ParquetRecordReaderBase.java:75)
> 	at org.apache.hadoop.hive.ql.io.parquet.read.ParquetRecordReaderWrapper.<init>(ParquetRecordReaderWrapper.java:75)
> 	at org.apache.hadoop.hive.ql.io.parquet.read.ParquetRecordReaderWrapper.<init>(ParquetRecordReaderWrapper.java:60)
> 	at org.apache.hadoop.hive.ql.io.parquet.MapredParquetInputFormat.getRecordReader(MapredParquetInputFormat.java:75)
> 	at org.apache.hadoop.hive.ql.io.CombineHiveRecordReader.<init>(CombineHiveRecordReader.java:99)
> 	... 16 more
> FAILED: Execution Error, return code 2 from org.apache.hadoop.hive.ql.exec.mr.MapRedTask
> MapReduce Jobs Launched:
> Stage-Stage-11: Map: 2   Cumulative CPU: 8.06 sec   HDFS Read: 30968 HDFS Write: 192 SUCCESS
> Stage-Stage-2: Map: 3   Cumulative CPU: 11.67 sec   HDFS Read: 38059 HDFS Write: 2079 SUCCESS
> Stage-Stage-4: Map: 1   HDFS Read: 0 HDFS Write: 0 FAIL
> Total MapReduce CPU Time Spent: 19 seconds 730 msec
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)