[ https://issues.apache.org/jira/browse/HIVE-22595?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16991184#comment-16991184 ]
Hive QA commented on HIVE-22595: -------------------------------- Here are the results of testing the latest attachment: https://issues.apache.org/jira/secure/attachment/12988284/HIVE-22595.1.patch {color:green}SUCCESS:{color} +1 due to 1 test(s) being added or modified. {color:red}ERROR:{color} -1 due to 80 failed/errored test(s), 17764 tests executed *Failed tests:* {noformat} org.apache.hadoop.hive.cli.TestBlobstoreCliDriver.testCliDriver[buckets] (batchId=304) org.apache.hadoop.hive.cli.TestBlobstoreCliDriver.testCliDriver[insert_into_dynamic_partitions] (batchId=304) org.apache.hadoop.hive.cli.TestBlobstoreCliDriver.testCliDriver[insert_overwrite_dynamic_partitions] (batchId=304) org.apache.hadoop.hive.cli.TestBlobstoreCliDriver.testCliDriver[orc_buckets] (batchId=304) org.apache.hadoop.hive.cli.TestBlobstoreCliDriver.testCliDriver[parquet_buckets] (batchId=304) org.apache.hadoop.hive.cli.TestBlobstoreCliDriver.testCliDriver[rcfile_buckets] (batchId=304) org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[acid_subquery] (batchId=44) org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[acid_table_stats] (batchId=59) org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[avro_extschema_insert] (batchId=5) org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[create_transactional_full_acid] (batchId=84) org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[create_transactional_insert_only] (batchId=66) org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[delete_all_partitioned] (batchId=31) org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[delete_where_partitioned] (batchId=44) org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[dynamic_partition_insert] (batchId=62) org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[explain_locks] (batchId=49) org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[insert_acid_dynamic_partition] (batchId=22) org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[insert_into_with_schema2] (batchId=43) org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[insert_with_move_files_from_source_dir] (batchId=96) org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[load_dyn_part2] (batchId=65) org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[load_static_ptn_into_bucketed_table] (batchId=22) org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[mm_buckets] (batchId=67) org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[update_all_partitioned] (batchId=58) org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[update_where_partitioned] (batchId=69) org.apache.hadoop.hive.cli.TestEncryptedHDFSCliDriver.testCliDriver[encryption_insert_partition_dynamic] (batchId=193) org.apache.hadoop.hive.cli.TestMiniLlapCliDriver.testCliDriver[mm_dp] (batchId=160) org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[auto_sortmerge_join_16] (batchId=178) org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[dynpart_sort_opt_vectorization] (batchId=176) org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[dynpart_sort_optimization] (batchId=177) org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[dynpart_sort_optimization_acid] (batchId=175) org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[llap_smb] (batchId=189) org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[load_data_using_job] (batchId=165) org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[murmur_hash_migration] (batchId=179) org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[orc_analyze] (batchId=165) org.apache.hadoop.hive.cli.TestMiniSparkOnYarnCliDriver.testCliDriver[auto_sortmerge_join_16] (batchId=196) org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver[auto_sortmerge_join_16] (batchId=139) org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver[dynpart_sort_optimization] (batchId=138) org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver[load_dyn_part2] (batchId=141) org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver[sample10] (batchId=136) org.apache.hadoop.hive.ql.TestTxnCommands2.testDynamicPartitionsMerge (batchId=337) org.apache.hadoop.hive.ql.TestTxnCommands2.testDynamicPartitionsMerge2 (batchId=337) org.apache.hadoop.hive.ql.TestTxnCommands2.testMultiInsert (batchId=337) org.apache.hadoop.hive.ql.TestTxnCommands2.updateDeletePartitioned (batchId=337) org.apache.hadoop.hive.ql.TestTxnCommands2WithSplitUpdateAndVectorization.testDynamicPartitionsMerge (batchId=351) org.apache.hadoop.hive.ql.TestTxnCommands2WithSplitUpdateAndVectorization.testDynamicPartitionsMerge2 (batchId=351) org.apache.hadoop.hive.ql.TestTxnCommands2WithSplitUpdateAndVectorization.testMultiInsert (batchId=351) org.apache.hadoop.hive.ql.TestTxnCommands2WithSplitUpdateAndVectorization.updateDeletePartitioned (batchId=351) org.apache.hadoop.hive.ql.TestTxnCommands3.testSdpoBucketed (batchId=352) org.apache.hadoop.hive.ql.TestTxnConcatenate.testConcatenatePart (batchId=322) org.apache.hadoop.hive.ql.lockmgr.TestDbTxnManager2.checkExpectedLocks2 (batchId=355) org.apache.hadoop.hive.ql.lockmgr.TestDbTxnManager2.testMerge3Way01 (batchId=355) org.apache.hadoop.hive.ql.lockmgr.TestDbTxnManager2.testMerge3Way02 (batchId=355) org.apache.hadoop.hive.ql.lockmgr.TestDbTxnManager2.testMergePartitioned01 (batchId=355) org.apache.hadoop.hive.ql.lockmgr.TestDbTxnManager2.testMergePartitioned02 (batchId=355) org.apache.hadoop.hive.ql.lockmgr.TestDbTxnManager2.testWriteSetTracking10 (batchId=355) org.apache.hadoop.hive.ql.lockmgr.TestDbTxnManager2.testWriteSetTracking11 (batchId=355) org.apache.hadoop.hive.ql.lockmgr.TestDbTxnManager2.testWriteSetTracking7 (batchId=355) org.apache.hadoop.hive.ql.lockmgr.TestDbTxnManager2.testWriteSetTracking8 (batchId=355) org.apache.hadoop.hive.ql.lockmgr.TestDbTxnManager2.testWriteSetTracking9 (batchId=355) org.apache.hadoop.hive.ql.parse.TestReplAcidTablesBootstrapWithJsonMessage.testAcidTablesBootstrapDuringIncremental (batchId=254) org.apache.hadoop.hive.ql.parse.TestReplAcidTablesBootstrapWithJsonMessage.testAcidTablesBootstrapDuringIncrementalWithOpenTxnsTimeout (batchId=254) org.apache.hadoop.hive.ql.parse.TestReplAcidTablesBootstrapWithJsonMessage.testBootstrapAcidTablesDuringIncrementalWithConcurrentWrites (batchId=254) org.apache.hadoop.hive.ql.parse.TestReplAcidTablesBootstrapWithJsonMessage.testRetryAcidTablesBootstrapFromDifferentDump (batchId=254) org.apache.hadoop.hive.ql.parse.TestReplAcidTablesWithJsonMessage.testAcidTablesBootstrap (batchId=271) org.apache.hadoop.hive.ql.parse.TestReplTableMigrationWithJsonFormat.testIncrementalLoadMigrationManagedToAcidAllOp (batchId=269) org.apache.hadoop.hive.ql.parse.TestReplicationScenariosAcidTables.testAcidTablesBootstrap (batchId=273) org.apache.hadoop.hive.ql.parse.TestReplicationScenariosAcidTablesBootstrap.testAcidTablesBootstrapDuringIncremental (batchId=252) org.apache.hadoop.hive.ql.parse.TestReplicationScenariosAcidTablesBootstrap.testAcidTablesBootstrapDuringIncrementalWithOpenTxnsTimeout (batchId=252) org.apache.hadoop.hive.ql.parse.TestReplicationScenariosAcidTablesBootstrap.testBootstrapAcidTablesDuringIncrementalWithConcurrentWrites (batchId=252) org.apache.hadoop.hive.ql.parse.TestReplicationScenariosAcidTablesBootstrap.testRetryAcidTablesBootstrapFromDifferentDump (batchId=252) org.apache.hadoop.hive.ql.parse.TestReplicationScenariosIncrementalLoadAcidTables.testAcidTableIncrementalReplication (batchId=274) org.apache.hadoop.hive.ql.parse.TestReplicationWithTableMigration.testIncrementalLoadMigrationManagedToAcidAllOp (batchId=258) org.apache.hadoop.hive.ql.txn.compactor.TestCompactor.dynamicPartitioningDelete (batchId=250) org.apache.hadoop.hive.ql.txn.compactor.TestCompactor.dynamicPartitioningInsert (batchId=250) org.apache.hadoop.hive.ql.txn.compactor.TestCompactor.dynamicPartitioningUpdate (batchId=250) org.apache.hadoop.hive.ql.txn.compactor.TestCompactor.schemaEvolutionAddColDynamicPartitioningInsert (batchId=250) org.apache.hadoop.hive.ql.txn.compactor.TestCompactor.schemaEvolutionAddColDynamicPartitioningUpdate (batchId=250) org.apache.hadoop.hive.ql.txn.compactor.TestCompactor.testDisableCompactionDuringReplLoad (batchId=250) org.apache.hadoop.hive.ql.txn.compactor.TestCrudCompactorOnTez.testCompactionWithSchemaEvolutionAndBuckets (batchId=277) org.apache.hive.service.server.TestHS2HttpServerPamConfiguration.testPamCorrectConfiguration (batchId=241) org.apache.hive.streaming.TestStreamingDynamicPartitioning.testDPStreamBucketingMatchesRegularBucketing (batchId=363) {noformat} Test results: https://builds.apache.org/job/PreCommit-HIVE-Build/19827/testReport Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/19827/console Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-19827/ Messages: {noformat} Executing org.apache.hive.ptest.execution.TestCheckPhase Executing org.apache.hive.ptest.execution.PrepPhase Executing org.apache.hive.ptest.execution.YetusPhase Executing org.apache.hive.ptest.execution.ExecutionPhase Executing org.apache.hive.ptest.execution.ReportingPhase Tests exited with: TestsFailedException: 80 tests failed {noformat} This message is automatically generated. ATTACHMENT ID: 12988284 - PreCommit-HIVE-Build > Dynamic partition inserts fail on Avro table table with external schema > ----------------------------------------------------------------------- > > Key: HIVE-22595 > URL: https://issues.apache.org/jira/browse/HIVE-22595 > Project: Hive > Issue Type: Bug > Components: Avro, Serializers/Deserializers > Reporter: Jason Dere > Assignee: Jason Dere > Priority: Major > Attachments: HIVE-22595.1.patch > > > Example qfile test: > {noformat} > create external table avro_extschema_insert1 (name string) partitioned by (p1 > string) > stored as avro tblproperties > ('avro.schema.url'='${system:test.tmp.dir}/table1.avsc'); > create external table avro_extschema_insert2 like avro_extschema_insert1; > insert overwrite table avro_extschema_insert1 partition (p1='part1') values > ('col1_value', 1, 'col3_value'); > insert overwrite table avro_extschema_insert2 partition (p1) select * from > avro_extschema_insert1; > {noformat} > The last statement fails with the following error: > {noformat} > ], TaskAttempt 3 failed, info=[Error: Error while running task ( failure ) : > attempt_1575484789169_0003_4_00_000000_3:java.lang.RuntimeException: > java.lang.RuntimeException: org.apache.hadoop.hive.ql.metadata.HiveException: > Hive Runtime Error while processing row > at > org.apache.hadoop.hive.ql.exec.tez.TezProcessor.initializeAndRunProcessor(TezProcessor.java:296) > at > org.apache.hadoop.hive.ql.exec.tez.TezProcessor.run(TezProcessor.java:250) > at > org.apache.tez.runtime.LogicalIOProcessorRuntimeTask.run(LogicalIOProcessorRuntimeTask.java:374) > at > org.apache.tez.runtime.task.TaskRunner2Callable$1.run(TaskRunner2Callable.java:73) > at > org.apache.tez.runtime.task.TaskRunner2Callable$1.run(TaskRunner2Callable.java:61) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:422) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1730) > at > org.apache.tez.runtime.task.TaskRunner2Callable.callInternal(TaskRunner2Callable.java:61) > at > org.apache.tez.runtime.task.TaskRunner2Callable.callInternal(TaskRunner2Callable.java:37) > at org.apache.tez.common.CallableWithNdc.call(CallableWithNdc.java:36) > at > com.google.common.util.concurrent.TrustedListenableFutureTask$TrustedFutureInterruptibleTask.runInterruptibly(TrustedListenableFutureTask.java:125) > at > com.google.common.util.concurrent.InterruptibleTask.run(InterruptibleTask.java:69) > at > com.google.common.util.concurrent.TrustedListenableFutureTask.run(TrustedListenableFutureTask.java:78) > at > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) > at java.lang.Thread.run(Thread.java:745) > Caused by: java.lang.RuntimeException: > org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime Error while > processing row > at > org.apache.hadoop.hive.ql.exec.tez.MapRecordSource.processRow(MapRecordSource.java:101) > at > org.apache.hadoop.hive.ql.exec.tez.MapRecordSource.pushRecord(MapRecordSource.java:76) > at > org.apache.hadoop.hive.ql.exec.tez.MapRecordProcessor.run(MapRecordProcessor.java:426) > at > org.apache.hadoop.hive.ql.exec.tez.TezProcessor.initializeAndRunProcessor(TezProcessor.java:267) > ... 16 more > Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime > Error while processing row > at > org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:576) > at > org.apache.hadoop.hive.ql.exec.tez.MapRecordSource.processRow(MapRecordSource.java:92) > ... 19 more > Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: > org.apache.hadoop.hive.serde2.avro.AvroSerdeException: Number of input > columns was different than output columns (in = 2 vs out = 1 > at > org.apache.hadoop.hive.ql.exec.FileSinkOperator.process(FileSinkOperator.java:1047) > at > org.apache.hadoop.hive.ql.exec.Operator.baseForward(Operator.java:994) > at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:940) > at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:927) > at > org.apache.hadoop.hive.ql.exec.SelectOperator.process(SelectOperator.java:95) > at > org.apache.hadoop.hive.ql.exec.Operator.baseForward(Operator.java:994) > at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:940) > at > org.apache.hadoop.hive.ql.exec.TableScanOperator.process(TableScanOperator.java:125) > at > org.apache.hadoop.hive.ql.exec.MapOperator$MapOpCtx.forward(MapOperator.java:153) > at > org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:555) > ... 20 more > Caused by: org.apache.hadoop.hive.serde2.avro.AvroSerdeException: Number of > input columns was different than output columns (in = 2 vs out = 1 > at > org.apache.hadoop.hive.serde2.avro.AvroSerializer.serialize(AvroSerializer.java:77) > at > org.apache.hadoop.hive.serde2.avro.AvroSerDe.serialize(AvroSerDe.java:223) > at > org.apache.hadoop.hive.ql.exec.FileSinkOperator.process(FileSinkOperator.java:951) > ... 29 more > {noformat} -- This message was sent by Atlassian Jira (v8.3.4#803005)