You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@hive.apache.org by "Zoltan Haindrich (JIRA)" <ji...@apache.org> on 2017/06/29 07:30:00 UTC

[jira] [Commented] (HIVE-16988) if partition column type is boolean, Streaming api AbstractRecordWriter.getPathForEndPoint NoSuchObjectException: partition values=[Y, 2017-06-29 14:32:36.508]

    [ https://issues.apache.org/jira/browse/HIVE-16988?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16067910#comment-16067910 ] 

Zoltan Haindrich commented on HIVE-16988:
-----------------------------------------

from the stacktrace it seems like you are using hive 2.0.0; but anyway I suspect that this might be related to HIVE-6590.
I'll try to reproduce this...but, could you give some more detail about how to reproduce your issue?

> if partition column type is boolean, Streaming api AbstractRecordWriter.getPathForEndPoint NoSuchObjectException: partition values=[Y, 2017-06-29 14:32:36.508]
> ---------------------------------------------------------------------------------------------------------------------------------------------------------------
>
>                 Key: HIVE-16988
>                 URL: https://issues.apache.org/jira/browse/HIVE-16988
>             Project: Hive
>          Issue Type: Bug
>            Reporter: brucewoo
>
> org.apache.nifi.util.hive.HiveWriter$ConnectFailure: Failed connecting to EndPoint {metaStoreUri='thrift://localhost:9083', database='dw_subject', table='alls', partitionVals=[Y, 2017-06-29 14:32:36.508] }
> 	at org.apache.nifi.util.hive.HiveWriter.<init>(HiveWriter.java:53) ~[nifi-hive-processors-1.1.2.jar:1.1.2]
> 	at org.apache.nifi.processors.hive.InsertTdfHive2.getOrCreateWriter(InsertTdfHive2.java:971) [nifi-hive-processors-1.1.2.jar:1.1.2]
> 	at org.apache.nifi.processors.hive.InsertTdfHive2.putStreamingHive(InsertTdfHive2.java:872) [nifi-hive-processors-1.1.2.jar:1.1.2]
> 	at org.apache.nifi.processors.hive.InsertTdfHive2.onTrigger(InsertTdfHive2.java:411) [nifi-hive-processors-1.1.2.jar:1.1.2]
> 	at org.apache.nifi.processor.AbstractProcessor.onTrigger(AbstractProcessor.java:27) [nifi-api-1.1.2.jar:1.1.2]
> 	at org.apache.nifi.controller.StandardProcessorNode.onTrigger(StandardProcessorNode.java:1099) [nifi-framework-core-1.1.2.jar:1.1.2]
> 	at org.apache.nifi.controller.tasks.ContinuallyRunProcessorTask.call(ContinuallyRunProcessorTask.java:136) [nifi-framework-core-1.1.2.jar:1.1.2]
> 	at org.apache.nifi.controller.tasks.ContinuallyRunProcessorTask.call(ContinuallyRunProcessorTask.java:1) [nifi-framework-core-1.1.2.jar:1.1.2]
> 	at org.apache.nifi.controller.scheduling.TimerDrivenSchedulingAgent$1.run(TimerDrivenSchedulingAgent.java:132) [nifi-framework-core-1.1.2.jar:1.1.2]
> 	at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) [na:1.8.0_131]
> 	at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308) [na:1.8.0_131]
> 	at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180) [na:1.8.0_131]
> 	at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294) [na:1.8.0_131]
> 	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) [na:1.8.0_131]
> 	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) [na:1.8.0_131]
> 	at java.lang.Thread.run(Thread.java:748) [na:1.8.0_131]
> Caused by: org.apache.hive.hcatalog.streaming.StreamingException: partition values=[Y, 2017-06-29 14:32:36.508]. Unable to get path for end point: [Y, 2017-06-29 14:32:36.508]
> 	at org.apache.hive.hcatalog.streaming.AbstractRecordWriter.getPathForEndPoint(AbstractRecordWriter.java:268) ~[hive-hcatalog-streaming-2.0.0.jar:2.0.0]
> 	at org.apache.hive.hcatalog.streaming.AbstractRecordWriter.<init>(AbstractRecordWriter.java:79) ~[hive-hcatalog-streaming-2.0.0.jar:2.0.0]
> 	at org.apache.hive.hcatalog.streaming.DelimitedInputWriter.<init>(DelimitedInputWriter.java:121) ~[hive-hcatalog-streaming-2.0.0.jar:2.0.0]
> 	at org.apache.hive.hcatalog.streaming.DelimitedInputWriter.<init>(DelimitedInputWriter.java:98) ~[hive-hcatalog-streaming-2.0.0.jar:2.0.0]
> 	at org.apache.hive.hcatalog.streaming.DelimitedInputWriter.<init>(DelimitedInputWriter.java:79) ~[hive-hcatalog-streaming-2.0.0.jar:2.0.0]
> 	at org.apache.nifi.util.hive.HiveWriter.getDelimitedInputWriter(HiveWriter.java:60) ~[nifi-hive-processors-1.1.2.jar:1.1.2]
> 	at org.apache.nifi.util.hive.HiveWriter.<init>(HiveWriter.java:46) ~[nifi-hive-processors-1.1.2.jar:1.1.2]
> 	... 15 common frames omitted
> Caused by: org.apache.hadoop.hive.metastore.api.NoSuchObjectException: partition values=[Y, 2017-06-29 14:32:36.508]
> 	at org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$get_partition_result$get_partition_resultStandardScheme.read(ThriftHiveMetastore.java) ~[hive-exec-2.0.0.jar:2.0.0]
> 	at org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$get_partition_result$get_partition_resultStandardScheme.read(ThriftHiveMetastore.java) ~[hive-exec-2.0.0.jar:2.0.0]
> 	at org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$get_partition_result.read(ThriftHiveMetastore.java) ~[hive-exec-2.0.0.jar:2.0.0]
> 	at org.apache.thrift.TServiceClient.receiveBase(TServiceClient.java:86) ~[libthrift-0.9.3.jar:0.9.3]
> 	at org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.recv_get_partition(ThriftHiveMetastore.java:1924) ~[hive-exec-2.0.0.jar:2.0.0]
> 	at org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.get_partition(ThriftHiveMetastore.java:1909) ~[hive-exec-2.0.0.jar:2.0.0]
> 	at org.apache.hadoop.hive.metastore.HiveMetaStoreClient.getPartition(HiveMetaStoreClient.java:1231) ~[hive-exec-2.0.0.jar:2.0.0]
> 	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[na:1.8.0_131]
> 	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) ~[na:1.8.0_131]
> 	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[na:1.8.0_131]
> 	at java.lang.reflect.Method.invoke(Method.java:498) ~[na:1.8.0_131]
> 	at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.invoke(RetryingMetaStoreClient.java:152) ~[hive-exec-2.0.0.jar:2.0.0]
> 	at com.sun.proxy.$Proxy112.getPartition(Unknown Source) ~[na:na]
> 	at org.apache.hive.hcatalog.streaming.AbstractRecordWriter.getPathForEndPoint(AbstractRecordWriter.java:263) ~[hive-hcatalog-streaming-2.0.0.jar:2.0.0]
> 	... 21 common frames omitted



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)