brucewoo created HIVE-16988: ------------------------------- Summary: if partition column type is boolean, Streaming api AbstractRecordWriter.getPathForEndPoint NoSuchObjectException: partition values=[Y, 2017-06-29 14:32:36.508] Key: HIVE-16988 URL: https://issues.apache.org/jira/browse/HIVE-16988 Project: Hive Issue Type: Bug Reporter: brucewoo
org.apache.nifi.util.hive.HiveWriter$ConnectFailure: Failed connecting to EndPoint {metaStoreUri='thrift://localhost:9083', database='dw_subject', table='alls', partitionVals=[Y, 2017-06-29 14:32:36.508] } at org.apache.nifi.util.hive.HiveWriter.<init>(HiveWriter.java:53) ~[nifi-hive-processors-1.1.2.jar:1.1.2] at org.apache.nifi.processors.hive.InsertTdfHive2.getOrCreateWriter(InsertTdfHive2.java:971) [nifi-hive-processors-1.1.2.jar:1.1.2] at org.apache.nifi.processors.hive.InsertTdfHive2.putStreamingHive(InsertTdfHive2.java:872) [nifi-hive-processors-1.1.2.jar:1.1.2] at org.apache.nifi.processors.hive.InsertTdfHive2.onTrigger(InsertTdfHive2.java:411) [nifi-hive-processors-1.1.2.jar:1.1.2] at org.apache.nifi.processor.AbstractProcessor.onTrigger(AbstractProcessor.java:27) [nifi-api-1.1.2.jar:1.1.2] at org.apache.nifi.controller.StandardProcessorNode.onTrigger(StandardProcessorNode.java:1099) [nifi-framework-core-1.1.2.jar:1.1.2] at org.apache.nifi.controller.tasks.ContinuallyRunProcessorTask.call(ContinuallyRunProcessorTask.java:136) [nifi-framework-core-1.1.2.jar:1.1.2] at org.apache.nifi.controller.tasks.ContinuallyRunProcessorTask.call(ContinuallyRunProcessorTask.java:1) [nifi-framework-core-1.1.2.jar:1.1.2] at org.apache.nifi.controller.scheduling.TimerDrivenSchedulingAgent$1.run(TimerDrivenSchedulingAgent.java:132) [nifi-framework-core-1.1.2.jar:1.1.2] at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) [na:1.8.0_131] at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308) [na:1.8.0_131] at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180) [na:1.8.0_131] at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294) [na:1.8.0_131] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) [na:1.8.0_131] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) [na:1.8.0_131] at java.lang.Thread.run(Thread.java:748) [na:1.8.0_131] Caused by: org.apache.hive.hcatalog.streaming.StreamingException: partition values=[Y, 2017-06-29 14:32:36.508]. Unable to get path for end point: [Y, 2017-06-29 14:32:36.508] at org.apache.hive.hcatalog.streaming.AbstractRecordWriter.getPathForEndPoint(AbstractRecordWriter.java:268) ~[hive-hcatalog-streaming-2.0.0.jar:2.0.0] at org.apache.hive.hcatalog.streaming.AbstractRecordWriter.<init>(AbstractRecordWriter.java:79) ~[hive-hcatalog-streaming-2.0.0.jar:2.0.0] at org.apache.hive.hcatalog.streaming.DelimitedInputWriter.<init>(DelimitedInputWriter.java:121) ~[hive-hcatalog-streaming-2.0.0.jar:2.0.0] at org.apache.hive.hcatalog.streaming.DelimitedInputWriter.<init>(DelimitedInputWriter.java:98) ~[hive-hcatalog-streaming-2.0.0.jar:2.0.0] at org.apache.hive.hcatalog.streaming.DelimitedInputWriter.<init>(DelimitedInputWriter.java:79) ~[hive-hcatalog-streaming-2.0.0.jar:2.0.0] at org.apache.nifi.util.hive.HiveWriter.getDelimitedInputWriter(HiveWriter.java:60) ~[nifi-hive-processors-1.1.2.jar:1.1.2] at org.apache.nifi.util.hive.HiveWriter.<init>(HiveWriter.java:46) ~[nifi-hive-processors-1.1.2.jar:1.1.2] ... 15 common frames omitted Caused by: org.apache.hadoop.hive.metastore.api.NoSuchObjectException: partition values=[Y, 2017-06-29 14:32:36.508] at org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$get_partition_result$get_partition_resultStandardScheme.read(ThriftHiveMetastore.java) ~[hive-exec-2.0.0.jar:2.0.0] at org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$get_partition_result$get_partition_resultStandardScheme.read(ThriftHiveMetastore.java) ~[hive-exec-2.0.0.jar:2.0.0] at org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$get_partition_result.read(ThriftHiveMetastore.java) ~[hive-exec-2.0.0.jar:2.0.0] at org.apache.thrift.TServiceClient.receiveBase(TServiceClient.java:86) ~[libthrift-0.9.3.jar:0.9.3] at org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.recv_get_partition(ThriftHiveMetastore.java:1924) ~[hive-exec-2.0.0.jar:2.0.0] at org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.get_partition(ThriftHiveMetastore.java:1909) ~[hive-exec-2.0.0.jar:2.0.0] at org.apache.hadoop.hive.metastore.HiveMetaStoreClient.getPartition(HiveMetaStoreClient.java:1231) ~[hive-exec-2.0.0.jar:2.0.0] at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[na:1.8.0_131] at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) ~[na:1.8.0_131] at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[na:1.8.0_131] at java.lang.reflect.Method.invoke(Method.java:498) ~[na:1.8.0_131] at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.invoke(RetryingMetaStoreClient.java:152) ~[hive-exec-2.0.0.jar:2.0.0] at com.sun.proxy.$Proxy112.getPartition(Unknown Source) ~[na:na] at org.apache.hive.hcatalog.streaming.AbstractRecordWriter.getPathForEndPoint(AbstractRecordWriter.java:263) ~[hive-hcatalog-streaming-2.0.0.jar:2.0.0] ... 21 common frames omitted -- This message was sent by Atlassian JIRA (v6.4.14#64029)