[ https://issues.apache.org/jira/browse/HIVE-21497?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16801265#comment-16801265 ]
Hive QA commented on HIVE-21497: -------------------------------- Here are the results of testing the latest attachment: https://issues.apache.org/jira/secure/attachment/12963663/HIVE-21497.2.patch {color:red}ERROR:{color} -1 due to no test(s) being added or modified. {color:green}SUCCESS:{color} +1 due to 15840 tests passed Test results: https://builds.apache.org/job/PreCommit-HIVE-Build/16677/testReport Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/16677/console Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-16677/ Messages: {noformat} Executing org.apache.hive.ptest.execution.TestCheckPhase Executing org.apache.hive.ptest.execution.PrepPhase Executing org.apache.hive.ptest.execution.YetusPhase Executing org.apache.hive.ptest.execution.ExecutionPhase Executing org.apache.hive.ptest.execution.ReportingPhase {noformat} This message is automatically generated. ATTACHMENT ID: 12963663 - PreCommit-HIVE-Build > Direct SQL exception thrown by PartitionManagementTask > ------------------------------------------------------ > > Key: HIVE-21497 > URL: https://issues.apache.org/jira/browse/HIVE-21497 > Project: Hive > Issue Type: Bug > Components: Standalone Metastore > Affects Versions: 4.0.0 > Reporter: Prasanth Jayachandran > Assignee: Jesus Camacho Rodriguez > Priority: Major > Attachments: HIVE-21497.1.patch, HIVE-21497.2.patch > > > Metastore runs background thread out of which one is partition discovery. > While removing expired partitions following exception is thrown > {code:java} > 2019-03-24 04:24:59.583 WARN [PartitionDiscoveryTask-0] > metastore.MetaStoreDirectSql: Failed to execute [select > "PARTITIONS"."PART_ID" from "PARTITIONS" inner join "TBLS" on > "PARTITIONS"."TBL_ID" = "TBLS"."TBL_ID" and "TBLS"."TBL_NAME" = ? inner join > "DBS" on "TBLS"."DB_ID" = "DBS"."DB_ID" and "DBS"."NAME" = ? inner join > "PARTITION_KEY_VALS" "FILTER0" on "FILTER0"."PART_ID" = > "PARTITIONS"."PART_ID" and "FILTER0"."INTEGER_IDX" = 0 inner join > "PARTITION_KEY_VALS" "FILTER1" on "FILTER1"."PART_ID" = > "PARTITIONS"."PART_ID" and "FILTER1"."INTEGER_IDX" = 1 inner join > "PARTITION_KEY_VALS" "FILTER2" on "FILTER2"."PART_ID" = > "PARTITIONS"."PART_ID" and "FILTER2"."INTEGER_IDX" = 2 where > "DBS"."CTLG_NAME" = ? and ( ( (((case when "FILTER0"."PART_KEY_VAL" <> ? and > "TBLS"."TBL_NAME" = ? and "DBS"."NAME" = ? and "DBS"."CTLG_NAME" = ? and > "FILTER0"."PART_ID" = "PARTITIONS"."PART_ID" and "FILTER0"."INTEGER_IDX" = 0 > then cast("FILTER0"."PART_KEY_VAL" as date) else null end) = ?) and > ("FILTER1"."PART_KEY_VAL" = ?)) and ("FILTER2"."PART_KEY_VAL" = ?)) )] with > parameters [logs, sys, hive, __HIVE_DEFAULT_PARTITION__, logs, sys, hive, > 2019-03-23, warehouse-1553300821-692w, metastore-db-create-job] > javax.jdo.JDODataStoreException: Error executing SQL query "select > "PARTITIONS"."PART_ID" from "PARTITIONS" inner join "TBLS" on > "PARTITIONS"."TBL_ID" = "TBLS"."TBL_ID" and "TBLS"."TBL_NAME" = ? inner join > "DBS" on "TBLS"."DB_ID" = "DBS"."DB_ID" and "DBS"."NAME" = ? inner join > "PARTITION_KEY_VALS" "FILTER0" on "FILTER0"."PART_ID" = > "PARTITIONS"."PART_ID" and "FILTER0"."INTEGER_IDX" = 0 inner join > "PARTITION_KEY_VALS" "FILTER1" on "FILTER1"."PART_ID" = > "PARTITIONS"."PART_ID" and "FILTER1"."INTEGER_IDX" = 1 inner join > "PARTITION_KEY_VALS" "FILTER2" on "FILTER2"."PART_ID" = > "PARTITIONS"."PART_ID" and "FILTER2"."INTEGER_IDX" = 2 where > "DBS"."CTLG_NAME" = ? and ( ( (((case when "FILTER0"."PART_KEY_VAL" <> ? and > "TBLS"."TBL_NAME" = ? and "DBS"."NAME" = ? and "DBS"."CTLG_NAME" = ? and > "FILTER0"."PART_ID" = "PARTITIONS"."PART_ID" and "FILTER0"."INTEGER_IDX" = 0 > then cast("FILTER0"."PART_KEY_VAL" as date) else null end) = ?) and > ("FILTER1"."PART_KEY_VAL" = ?)) and ("FILTER2"."PART_KEY_VAL" = ?)) )". > at > org.datanucleus.api.jdo.NucleusJDOHelper.getJDOExceptionForNucleusException(NucleusJDOHelper.java:543) > at org.datanucleus.api.jdo.JDOQuery.executeInternal(JDOQuery.java:391) > at org.datanucleus.api.jdo.JDOQuery.executeWithArray(JDOQuery.java:267) > at > org.apache.hadoop.hive.metastore.MetaStoreDirectSql.executeWithArray(MetaStoreDirectSql.java:2042) > at > org.apache.hadoop.hive.metastore.MetaStoreDirectSql.getPartitionIdsViaSqlFilter(MetaStoreDirectSql.java:621) > at > org.apache.hadoop.hive.metastore.MetaStoreDirectSql.getPartitionsViaSqlFilter(MetaStoreDirectSql.java:487) > at > org.apache.hadoop.hive.metastore.ObjectStore$9.getSqlResult(ObjectStore.java:3426) > at > org.apache.hadoop.hive.metastore.ObjectStore$9.getSqlResult(ObjectStore.java:3418) > at > org.apache.hadoop.hive.metastore.ObjectStore$GetHelper.run(ObjectStore.java:3702) > at > org.apache.hadoop.hive.metastore.ObjectStore.getPartitionsByExprInternal(ObjectStore.java:3453) > at > org.apache.hadoop.hive.metastore.ObjectStore.getPartitionsByExpr(ObjectStore.java:3406) > at sun.reflect.GeneratedMethodAccessor82.invoke(Unknown Source) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:498) > at > org.apache.hadoop.hive.metastore.RawStoreProxy.invoke(RawStoreProxy.java:97) > at com.sun.proxy.$Proxy33.getPartitionsByExpr(Unknown Source) > at > org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.drop_partitions_req(HiveMetaStore.java:4521) > at sun.reflect.GeneratedMethodAccessor84.invoke(Unknown Source) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:498) > at > org.apache.hadoop.hive.metastore.RetryingHMSHandler.invokeInternal(RetryingHMSHandler.java:147) > at > org.apache.hadoop.hive.metastore.RetryingHMSHandler.invoke(RetryingHMSHandler.java:108) > at com.sun.proxy.$Proxy34.drop_partitions_req(Unknown Source) > at > org.apache.hadoop.hive.metastore.HiveMetaStoreClient.dropPartitions(HiveMetaStoreClient.java:1288) > at org.apache.hadoop.hive.metastore.Msck$2.execute(Msck.java:474) > at org.apache.hadoop.hive.metastore.Msck$2.execute(Msck.java:435) > at > org.apache.hadoop.hive.metastore.utils.RetryUtilities$ExponentiallyDecayingBatchWork.run(RetryUtilities.java:91) > at > org.apache.hadoop.hive.metastore.Msck.dropPartitionsInBatches(Msck.java:499) > at org.apache.hadoop.hive.metastore.Msck.repair(Msck.java:226) > at > org.apache.hadoop.hive.metastore.PartitionManagementTask$MsckThread.run(PartitionManagementTask.java:213) > at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) > at java.util.concurrent.FutureTask.run(FutureTask.java:266) > at > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) > at java.lang.Thread.run(Thread.java:748) > NestedThrowablesStackTrace: > org.postgresql.util.PSQLException: ERROR: operator does not exist: date = > character varying > Hint: No operator matches the given name and argument type(s). You might need > to add explicit type casts. > Position: 886 > at > org.postgresql.core.v3.QueryExecutorImpl.receiveErrorResponse(QueryExecutorImpl.java:2284) > at > org.postgresql.core.v3.QueryExecutorImpl.processResults(QueryExecutorImpl.java:2003) > at > org.postgresql.core.v3.QueryExecutorImpl.execute(QueryExecutorImpl.java:200) > at org.postgresql.jdbc.PgStatement.execute(PgStatement.java:424) > at > org.postgresql.jdbc.PgPreparedStatement.executeWithFlags(PgPreparedStatement.java:161) > at > org.postgresql.jdbc.PgPreparedStatement.executeQuery(PgPreparedStatement.java:114) > at > com.zaxxer.hikari.pool.ProxyPreparedStatement.executeQuery(ProxyPreparedStatement.java:52) > at > com.zaxxer.hikari.pool.HikariProxyPreparedStatement.executeQuery(HikariProxyPreparedStatement.java) > at > org.datanucleus.store.rdbms.ParamLoggingPreparedStatement.executeQuery(ParamLoggingPreparedStatement.java:375) > at > org.datanucleus.store.rdbms.SQLController.executeStatementQuery(SQLController.java:552) > at > org.datanucleus.store.rdbms.query.SQLQuery.performExecute(SQLQuery.java:645) > at org.datanucleus.store.query.Query.executeQuery(Query.java:1855) > at > org.datanucleus.store.rdbms.query.SQLQuery.executeWithArray(SQLQuery.java:807) > at org.datanucleus.api.jdo.JDOQuery.executeInternal(JDOQuery.java:368) > at org.datanucleus.api.jdo.JDOQuery.executeWithArray(JDOQuery.java:267) > at > org.apache.hadoop.hive.metastore.MetaStoreDirectSql.executeWithArray(MetaStoreDirectSql.java:2042) > at > org.apache.hadoop.hive.metastore.MetaStoreDirectSql.getPartitionIdsViaSqlFilter(MetaStoreDirectSql.java:621) > at > org.apache.hadoop.hive.metastore.MetaStoreDirectSql.getPartitionsViaSqlFilter(MetaStoreDirectSql.java:487) > at > org.apache.hadoop.hive.metastore.ObjectStore$9.getSqlResult(ObjectStore.java:3426) > at > org.apache.hadoop.hive.metastore.ObjectStore$9.getSqlResult(ObjectStore.java:3418) > at > org.apache.hadoop.hive.metastore.ObjectStore$GetHelper.run(ObjectStore.java:3702) > at > org.apache.hadoop.hive.metastore.ObjectStore.getPartitionsByExprInternal(ObjectStore.java:3453) > at > org.apache.hadoop.hive.metastore.ObjectStore.getPartitionsByExpr(ObjectStore.java:3406) > at sun.reflect.GeneratedMethodAccessor82.invoke(Unknown Source) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:498) > at > org.apache.hadoop.hive.metastore.RawStoreProxy.invoke(RawStoreProxy.java:97) > at com.sun.proxy.$Proxy33.getPartitionsByExpr(Unknown Source) > at > org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.drop_partitions_req(HiveMetaStore.java:4521) > at sun.reflect.GeneratedMethodAccessor84.invoke(Unknown Source) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:498) > at > org.apache.hadoop.hive.metastore.RetryingHMSHandler.invokeInternal(RetryingHMSHandler.java:147) > at > org.apache.hadoop.hive.metastore.RetryingHMSHandler.invoke(RetryingHMSHandler.java:108) > at com.sun.proxy.$Proxy34.drop_partitions_req(Unknown Source) > at > org.apache.hadoop.hive.metastore.HiveMetaStoreClient.dropPartitions(HiveMetaStoreClient.java:1288) > at org.apache.hadoop.hive.metastore.Msck$2.execute(Msck.java:474) > at org.apache.hadoop.hive.metastore.Msck$2.execute(Msck.java:435) > at > org.apache.hadoop.hive.metastore.utils.RetryUtilities$ExponentiallyDecayingBatchWork.run(RetryUtilities.java:91) > at > org.apache.hadoop.hive.metastore.Msck.dropPartitionsInBatches(Msck.java:499) > at org.apache.hadoop.hive.metastore.Msck.repair(Msck.java:226) > at > org.apache.hadoop.hive.metastore.PartitionManagementTask$MsckThread.run(PartitionManagementTask.java:213) > at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) > at java.util.concurrent.FutureTask.run(FutureTask.java:266) > at > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) > at java.lang.Thread.run(Thread.java:748) > {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)