[ https://issues.apache.org/jira/browse/HIVE-17680?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16993624#comment-16993624 ]
László Bodor edited comment on HIVE-17680 at 12/11/19 3:20 PM: --------------------------------------------------------------- I think this test is not 100% valid on branch-2 this way the original exception was {code} The table must be bucketed and stored using an ACID compliant format (such as ORC) {code} while creating this table {code} CREATE TABLE srcpart_acid (key STRING, value STRING) PARTITIONED BY (ds STRING, hr STRING) stored as ORC TBLPROPERTIES ('transactional'='true', 'transactional_properties'='default'); {code} seems like bucketing should be explicitly told while creating a transactional table after changing the CREATE TABLE ddl to use bucketing, the problem disappears, but it also changes the negative output message from {code} FAILED: SemanticException org.apache.hadoop.hive.ql.parse.SemanticException: Concatenate/Merge can not be performed on transactional tables {code} to {code} FAILED: SemanticException org.apache.hadoop.hive.ql.parse.SemanticException: Concatenate/Merge can not be performed on bucketed tables {code} this q.out was only affected on branch-2 by HIVE-17403, so it should have been failing since that if we accept that expected negative output becomes "bucketed", this test could remain valid, but a bit similar to merge_negative_4.q,merge_negative_5.q (which was about bucketed table on master) was (Author: abstractdog): I thinks test is not 100% valid on branch-2 the original exception was {code} The table must be bucketed and stored using an ACID compliant format (such as ORC) {code} while creating this table {code} CREATE TABLE srcpart_acid (key STRING, value STRING) PARTITIONED BY (ds STRING, hr STRING) stored as ORC TBLPROPERTIES ('transactional'='true', 'transactional_properties'='default'); {code} seems like bucketing should be explicitly told while creating a transactional table after changing the CREATE TABLE ddl to use bucketing, the probles disappears, but it also changes the negative output message from {code} FAILED: SemanticException org.apache.hadoop.hive.ql.parse.SemanticException: Concatenate/Merge can not be performed on transactional tables {code} to {code} FAILED: SemanticException org.apache.hadoop.hive.ql.parse.SemanticException: Concatenate/Merge can not be performed on bucketed tables {code} this q.out was only affected on branch-2 by HIVE-17403, so it should have been failing since that if we accept that expected negative output becomes "bucketed", this test could remain valid, but a bit similar or merge_negative_4.q,merge_negative_5.q (which was about bucketed table on master) > TestNegativeCliDriver[merge_negative_5.q] > ----------------------------------------- > > Key: HIVE-17680 > URL: https://issues.apache.org/jira/browse/HIVE-17680 > Project: Hive > Issue Type: Sub-task > Reporter: Vihang Karajgaonkar > Assignee: László Bodor > Priority: Major > Attachments: HIVE-17680.01.branch-2.patch > > > Seeing this error stack in the hive.log when I run it locally > Test failed in > https://builds.apache.org/job/PreCommit-HIVE-Build/7083/testReport > {noformat} > 2017-10-03T10:34:59,335 ERROR [d6419a0a-1fc9-4454-9575-f085c7f2cb55 main] > metastore.RetryingHMSHandler: MetaException(message:The table must be > bucketed and stored using an ACID compliant format (such as ORC)) > at > org.apache.hadoop.hive.metastore.TransactionalValidationListener.handleCreateTableTransactionalProp(TransactionalValidationListener.java:199) > at > org.apache.hadoop.hive.metastore.TransactionalValidationListener.handle(TransactionalValidationListener.java:68) > at > org.apache.hadoop.hive.metastore.TransactionalValidationListener.onEvent(TransactionalValidationListener.java:53) > at > org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.firePreEvent(HiveMetaStore.java:2290) > at > org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.create_table_core(HiveMetaStore.java:1413) > at > org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.create_table_with_environment_context(HiveMetaStore.java:1511) > at sun.reflect.GeneratedMethodAccessor56.invoke(Unknown Source) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:498) > at > org.apache.hadoop.hive.metastore.RetryingHMSHandler.invokeInternal(RetryingHMSHandler.java:148) > at > org.apache.hadoop.hive.metastore.RetryingHMSHandler.invoke(RetryingHMSHandler.java:107) > at > com.sun.proxy.$Proxy34.create_table_with_environment_context(Unknown Source) > at > org.apache.hadoop.hive.metastore.HiveMetaStoreClient.create_table_with_environment_context(HiveMetaStoreClient.java:2407) > at > org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient.create_table_with_environment_context(SessionHiveMetaStoreClient.java:93) > at > org.apache.hadoop.hive.metastore.HiveMetaStoreClient.createTable(HiveMetaStoreClient.java:760) > at > org.apache.hadoop.hive.metastore.HiveMetaStoreClient.createTable(HiveMetaStoreClient.java:748) > at sun.reflect.GeneratedMethodAccessor55.invoke(Unknown Source) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:498) > at > org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.invoke(RetryingMetaStoreClient.java:173) > at com.sun.proxy.$Proxy35.createTable(Unknown Source) > at org.apache.hadoop.hive.ql.metadata.Hive.createTable(Hive.java:852) > at org.apache.hadoop.hive.ql.metadata.Hive.createTable(Hive.java:867) > at > org.apache.hadoop.hive.ql.exec.DDLTask.createTable(DDLTask.java:4376) > at org.apache.hadoop.hive.ql.exec.DDLTask.execute(DDLTask.java:354) > at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:199) > at > org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:100) > at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:2183) > at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1839) > at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1526) > at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1237) > at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1227) > at > org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:233) > at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:184) > at > org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:403) > at > org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:336) > at > org.apache.hadoop.hive.ql.QTestUtil.executeClientInternal(QTestUtil.java:1337) > at > org.apache.hadoop.hive.ql.QTestUtil.executeClient(QTestUtil.java:1311) > at > org.apache.hadoop.hive.cli.control.CoreNegativeCliDriver.runTest(CoreNegativeCliDriver.java:121) > at > org.apache.hadoop.hive.cli.control.CliAdapter.runTest(CliAdapter.java:104) > at > org.apache.hadoop.hive.cli.TestNegativeCliDriver.testCliDriver(TestNegativeCliDriver.java:59) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:498) > at > org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:47) > at > org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) > at > org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:44) > at > org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) > at > org.apache.hadoop.hive.cli.control.CliAdapter$2$1.evaluate(CliAdapter.java:92) > at org.junit.rules.RunRules.evaluate(RunRules.java:20) > at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:271) > at > org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:70) > at > org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50) > at org.junit.runners.ParentRunner$3.run(ParentRunner.java:238) > at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:63) > at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:236) > at org.junit.runners.ParentRunner.access$000(ParentRunner.java:53) > at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:229) > at org.junit.runners.ParentRunner.run(ParentRunner.java:309) > at org.junit.runners.Suite.runChild(Suite.java:127) > {noformat} -- This message was sent by Atlassian Jira (v8.3.4#803005)