You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@nifi.apache.org by "Dmitro-Vasilenko (JIRA)" <ji...@apache.org> on 2016/12/06 13:02:58 UTC

[jira] [Comment Edited] (NIFI-2873) PutHiveStreaming throws UnknownHostException with HA NameNode

    [ https://issues.apache.org/jira/browse/NIFI-2873?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15725302#comment-15725302 ] 

Dmitro-Vasilenko edited comment on NIFI-2873 at 12/6/16 1:02 PM:
-----------------------------------------------------------------

Hello!

PutHiveStreaming failed to process due to java.net.UnknownHostException   in version   nifi-1.1.0  ,  HDP-2.5


was (Author: dmitro-vasilenko):
Hello!

PutHiveStreaming failed to process due to java.net.UnknownHostException   in version   nifi-1.1.0-RC2  ,  HDP-2.5

> PutHiveStreaming throws UnknownHostException with HA NameNode
> -------------------------------------------------------------
>
>                 Key: NIFI-2873
>                 URL: https://issues.apache.org/jira/browse/NIFI-2873
>             Project: Apache NiFi
>          Issue Type: Bug
>          Components: Extensions
>    Affects Versions: 1.0.0
>            Reporter: Franco
>             Fix For: 1.1.0
>
>         Attachments: nifi-app_2016-12-06_11.0.txt
>
>
> This is the same issue that previously affected Spark:
> https://github.com/Jianfeng-chs/spark/commit/9f2b2bf001262215742be418f24d5093c92ff10f
> We are experiencing this issue consistently when trying to use PutHiveStreaming. In theory this should be a problem with GetHDFS but for whatever reason it is not.
> The fix is identical namely preloading the Hadoop configuration during the processor setup phase. Pull request forthcoming.
> {code:title=Stack Trace|borderStyle=solid}
> 2016-10-06 16:07:59,225 ERROR [Timer-Driven Process Thread-9] o.a.n.processors.hive.PutHiveStreaming
> java.lang.IllegalArgumentException: java.net.UnknownHostException: tdcdv2
>         at org.apache.hadoop.security.SecurityUtil.buildTokenService(SecurityUtil.java:374) ~[hadoop-common-2.6.2.jar:na]
>         at org.apache.hadoop.hdfs.NameNodeProxies.createNonHAProxy(NameNodeProxies.java:310) ~[hadoop-hdfs-2.6.2.jar:na]
>         at org.apache.hadoop.hdfs.NameNodeProxies.createProxy(NameNodeProxies.java:176) ~[hadoop-hdfs-2.6.2.jar:na]
>         at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:668) ~[hadoop-hdfs-2.6.2.jar:na]
>         at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:604) ~[hadoop-hdfs-2.6.2.jar:na]
>         at org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:148) ~[hadoop-hdfs-2.6.2.jar:na]
>         at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2596) ~[hadoop-common-2.6.2.jar:na]
>         at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:91) ~[hadoop-common-2.6.2.jar:na]
>         at org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2630) ~[hadoop-common-2.6.2.jar:na]
>         at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2612) ~[hadoop-common-2.6.2.jar:na]
>         at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:370) ~[hadoop-common-2.6.2.jar:na]
>         at org.apache.hadoop.fs.Path.getFileSystem(Path.java:296) ~[hadoop-common-2.6.2.jar:na]
>         at org.apache.hadoop.hive.ql.io.orc.OrcRecordUpdater.<init>(OrcRecordUpdater.java:221) ~[hive-exec-1.2.1.jar:1.2.1]
>         at org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat.getRecordUpdater(OrcOutputFormat.java:292) ~[hive-exec-1.2.1.jar:1.2.1]
>         at org.apache.hive.hcatalog.streaming.AbstractRecordWriter.createRecordUpdater(AbstractRecordWriter.java:141) ~[hive-hcatalog-streaming-1.2.1.jar:1.2.1]
>         at org.apache.hive.hcatalog.streaming.AbstractRecordWriter.newBatch(AbstractRecordWriter.java:121) ~[hive-hcatalog-streaming-1.2.1.jar:1.2.1]
>         at org.apache.hive.hcatalog.streaming.StrictJsonWriter.newBatch(StrictJsonWriter.java:37) ~[hive-hcatalog-streaming-1.2.1.jar:1.2.1]
>         at org.apache.hive.hcatalog.streaming.HiveEndPoint$TransactionBatchImpl.<init>(HiveEndPoint.java:509) ~[hive-hcatalog-streaming-1.2.1.jar:1.2.1]
>         at org.apache.hive.hcatalog.streaming.HiveEndPoint$TransactionBatchImpl.<init>(HiveEndPoint.java:461) ~[hive-hcatalog-streaming-1.2.1.jar:1.2.1]
>         at org.apache.hive.hcatalog.streaming.HiveEndPoint$ConnectionImpl.fetchTransactionBatchImpl(HiveEndPoint.java:345) ~[hive-hcatalog-streaming-1.2.1.jar:1.2.1]
>         at org.apache.hive.hcatalog.streaming.HiveEndPoint$ConnectionImpl.fetchTransactionBatch(HiveEndPoint.java:325) ~[hive-hcatalog-streaming-1.2.1.jar:1.2.1]
>         at org.apache.nifi.util.hive.HiveWriter.lambda$nextTxnBatch$1(HiveWriter.java:250) ~[nifi-hive-processors-1.0.0.jar:1.0.0]
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)