[ https://issues.apache.org/jira/browse/HIVE-22757?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Deegue updated HIVE-22757: -------------------------- Attachment: HIVE-22757.patch > NullPointerException when executing SQLs > ---------------------------------------- > > Key: HIVE-22757 > URL: https://issues.apache.org/jira/browse/HIVE-22757 > Project: Hive > Issue Type: Bug > Affects Versions: 2.3.6 > Reporter: Deegue > Priority: Major > Attachments: HIVE-22757.patch > > > When executing SQL: > {code:java} > insert overwrite table ods.ods_1 partition(stat_day='20191209') > select > id > ,user_id > ,teacher_user_id > ,partner_user_id > ,order_id > ,barcode > ,sub_order_id > ,item_id > ,sales > ,refund > ,teacher_profit > ,partner_profit > ,teacher_refund_profit > ,partner_refund_profit > ,teacher_commission_value > ,partner_commission_value > ,biz_type > ,pay_time > ,item_profit_type > ,black_mark > ,is_deleted > ,create_time > ,modify_time > from src.src_1 > where partition_date='20191209' > union all > select > t1.id > ,t1.user_id > ,t1.teacher_user_id > ,t1.partner_user_id > ,t1.order_id > ,t1.barcode > ,t1.sub_order_id > ,t1.item_id > ,t1.sales > ,t1.refund > ,t1.teacher_profit > ,t1.partner_profit > ,t1.teacher_refund_profit > ,t1.partner_refund_profit > ,t1.teacher_commission_value > ,t1.partner_commission_value > ,t1.biz_type > ,t1.pay_time > ,t1.item_profit_type > ,t1.black_mark > ,t1.is_deleted > ,t1.create_time > ,t1.modify_time > from > (select * > from ods.ods_1 > where stat_day='20191208' > ) t1 > left join > ( select order_id > ,sub_order_id > from src.src_1 > where partition_date='20191209' > ) t2 > on t1.order_id=t2.order_id > and t1.sub_order_id=t2.sub_order_id > where t2.order_id is null > {code} > `java.lang.NullPointerException` thrown because the array list > `neededNestedColumnPaths` haven't been inited when `addAll` method is invoked. > {code:java} > Launching Job 5 out of 5 > Number of reduce tasks is set to 0 since there's no reduce operator > Starting Job = job_1566481621886_4925755, Tracking URL = > http://TXIDC65-bigdata-resourcemanager1:8042/proxy/application_1566481621886_4925755/ > Kill Command = /usr/local/yunji/hadoop/bin/hadoop job -kill > job_1566481621886_4925755 > Hadoop job information for Stage-4: number of mappers: 1; number of reducers: > 0 > 2019-12-24 16:00:40,584 Stage-4 map = 0%, reduce = 0% > 2019-12-24 16:01:40,956 Stage-4 map = 0%, reduce = 0% > 2019-12-24 16:02:41,451 Stage-4 map = 0%, reduce = 0% > 2019-12-24 16:02:45,550 Stage-4 map = 100%, reduce = 0% > Ended Job = job_1566481621886_4925755 with errors > Error during job, obtaining debugging information... > Examining task ID: task_1566481621886_4925755_m_000000 (and more) from job > job_1566481621886_4925755 > Task with the most failures(4): > ----- > Task ID: > task_1566481621886_4925755_m_000000 > URL: > > http://TXIDC65-bigdata-resourcemanager1:8088/taskdetails.jsp?jobid=job_1566481621886_4925755&tipid=task_1566481621886_4925755_m_000000 > ----- > Diagnostic Messages for this Task: > Error: java.io.IOException: java.lang.reflect.InvocationTargetException > at > org.apache.hadoop.hive.io.HiveIOExceptionHandlerChain.handleRecordReaderCreationException(HiveIOExceptionHandlerChain.java:97) > at > org.apache.hadoop.hive.io.HiveIOExceptionHandlerUtil.handleRecordReaderCreationException(HiveIOExceptionHandlerUtil.java:57) > at > org.apache.hadoop.hive.shims.HadoopShimsSecure$CombineFileRecordReader.initNextRecordReader(HadoopShimsSecure.java:271) > at > org.apache.hadoop.hive.shims.HadoopShimsSecure$CombineFileRecordReader.<init>(HadoopShimsSecure.java:217) > at > org.apache.hadoop.hive.shims.HadoopShimsSecure$CombineFileInputFormatShim.getRecordReader(HadoopShimsSecure.java:345) > at > org.apache.hadoop.hive.ql.io.CombineHiveInputFormat.getRecordReader(CombineHiveInputFormat.java:695) > at > org.apache.hadoop.mapred.MapTask$TrackedRecordReader.<init>(MapTask.java:169) > at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:438) > at org.apache.hadoop.mapred.MapTask.run(MapTask.java:343) > at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:164) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:422) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1924) > at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:158) > Caused by: java.lang.reflect.InvocationTargetException > at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) > at > sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) > at > sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) > at java.lang.reflect.Constructor.newInstance(Constructor.java:423) > at > org.apache.hadoop.hive.shims.HadoopShimsSecure$CombineFileRecordReader.initNextRecordReader(HadoopShimsSecure.java:257) > ... 11 more > Caused by: java.lang.NullPointerException > at java.util.AbstractCollection.addAll(AbstractCollection.java:343) > at > org.apache.hadoop.hive.ql.io.parquet.ProjectionPusher.pushProjectionsAndFilters(ProjectionPusher.java:118) > at > org.apache.hadoop.hive.ql.io.parquet.ProjectionPusher.pushProjectionsAndFilters(ProjectionPusher.java:189) > at > org.apache.hadoop.hive.ql.io.parquet.ParquetRecordReaderBase.getSplit(ParquetRecordReaderBase.java:75) > at > org.apache.hadoop.hive.ql.io.parquet.read.ParquetRecordReaderWrapper.<init>(ParquetRecordReaderWrapper.java:75) > at > org.apache.hadoop.hive.ql.io.parquet.read.ParquetRecordReaderWrapper.<init>(ParquetRecordReaderWrapper.java:60) > at > org.apache.hadoop.hive.ql.io.parquet.MapredParquetInputFormat.getRecordReader(MapredParquetInputFormat.java:75) > at > org.apache.hadoop.hive.ql.io.CombineHiveRecordReader.<init>(CombineHiveRecordReader.java:99) > ... 16 more > FAILED: Execution Error, return code 2 from > org.apache.hadoop.hive.ql.exec.mr.MapRedTask > MapReduce Jobs Launched: > Stage-Stage-11: Map: 2 Cumulative CPU: 8.06 sec HDFS Read: 30968 HDFS > Write: 192 SUCCESS > Stage-Stage-2: Map: 3 Cumulative CPU: 11.67 sec HDFS Read: 38059 HDFS > Write: 2079 SUCCESS > Stage-Stage-4: Map: 1 HDFS Read: 0 HDFS Write: 0 FAIL > Total MapReduce CPU Time Spent: 19 seconds 730 msec > {code} -- This message was sent by Atlassian Jira (v8.3.4#803005)