You are viewing a plain text version of this content. The canonical link for it is here.
Posted to user@flink.apache.org by 邓俊华 <de...@startdt.com> on 2017/10/20 04:07:02 UTC

flink can't read hdfs namenode logical url

hi,
I start yarn-ssession.sh on yarn, but it can't read hdfs logical url. It always connect to hdfs://master:8020, it should be 9000, my hdfs defaultfs is hdfs://master.I have config the YARN_CONF_DIR and HADOOP_CONF_DIR, it didn't work.Is it a bug? i use flink-1.3.0-bin-hadoop27-scala_2.10
2017-10-20 11:00:05,395 DEBUG org.apache.hadoop.ipc.Client                                  - IPC Client (1035144464) connection to startdt/173.16.5.215:8020 from admin: closed
2017-10-20 11:00:05,398 ERROR org.apache.flink.yarn.YarnApplicationMasterRunner             - YARN Application Master initialization failed
java.net.ConnectException: Call From spark3/173.16.5.216 to master:8020 failed on connection exception: java.net.ConnectException: Connection refused; For more details see:  http://wiki.apache.org/hadoop/ConnectionRefused
	at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
	at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57)
	at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
	at java.lang.reflect.Constructor.newInstance(Constructor.java:526)
	at org.apache.hadoop.net.NetUtils.wrapWithMessage(NetUtils.java:792)
	at org.apache.hadoop.net.NetUtils.wrapException(NetUtils.java:732)
	at org.apache.hadoop.ipc.Client.call(Client.java:1479)
	at org.apache.hadoop.ipc.Client.call(Client.java:1412)
	at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:229)
	at com.sun.proxy.$Proxy9.getFileInfo(Unknown Source)
	at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.getFileInfo(ClientNamenodeProtocolTranslatorPB.java:771)
	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:606)
	at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:191)
	at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:102)
	at com.sun.proxy.$Proxy10.getFileInfo(Unknown Source)
	at org.apache.hadoop.hdfs.DFSClient.getFileInfo(DFSClient.java:2108)
	at org.apache.hadoop.hdfs.DistributedFileSystem$22.doCall(DistributedFileSystem.java:1305)
	at org.apache.hadoop.hdfs.DistributedFileSystem$22.doCall(DistributedFileSystem.java:1301)
	at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)


Re: flink can't read hdfs namenode logical url

Posted by Piotr Nowojski <pi...@data-artisans.com>.
Hi,

Why in this new message there is a different host? Previously code was trying to connect to “master:8020” and now it is “startdt”? If you were able to change this host somehow between runs, I guess you should be also able to set it to correct one.

Piotrek

> On 23 Oct 2017, at 09:11, 邓俊华 <de...@startdt.com> wrote:
> 
> Hi,
> 
> Thanks for your replay! I have been block in this for several days.
> And I have double checked that there are hdfs-site.xml,core-site.xml,yarn-site.xml  in YARN_CONF_DIR. But it is still can't read hdfs namenode logical url.
> 
> 2017-10-23 14:35:17,750 DEBUG org.apache.flink.api.java.hadoop.mapred.utils.HadoopUtils     - Cannot find hdfs-default configuration file
> 2017-10-23 14:35:17,750 DEBUG org.apache.flink.api.java.hadoop.mapred.utils.HadoopUtils     - Cannot find hdfs-site configuration file
> 2017-10-23 14:35:17,751 DEBUG org.apache.flink.api.java.hadoop.mapred.utils.HadoopUtils     - Adding /data/hadoop-2.7.3/etc/hadoop/core-site.xml to hadoop configuration
> 2017-10-23 14:35:17,751 DEBUG org.apache.flink.api.java.hadoop.mapred.utils.HadoopUtils     - Adding /data/hadoop-2.7.3/etc/hadoop/hdfs-site.xml to hadoop configuration
> 
> 2017-10-23 14:35:19,887 DEBUG org.apache.hadoop.hdfs.BlockReaderLocal                       - dfs.domain.socket.path = 
> 2017-10-23 14:35:19,952 ERROR org.apache.flink.yarn.YarnApplicationMasterRunner             - YARN Application Master initialization failed
> java.lang.IllegalArgumentException: java.net.UnknownHostException: startdt
> 	at org.apache.hadoop.security.SecurityUtil.buildTokenService(SecurityUtil.java:378)
> 	at org.apache.hadoop.hdfs.NameNodeProxies.createNonHAProxy(NameNodeProxies.java:310)
> 	at org.apache.hadoop.hdfs.NameNodeProxies.createProxy(NameNodeProxies.java:176)
> 	at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:678)
> 	at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:619)
> 	at org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:149)
> ------------------------------------------------------------------
> 发件人:Piotr Nowojski <pi...@data-artisans.com>
> 发送时间:2017年10月20日(星期五) 21:39
> 收件人:邓俊华 <de...@startdt.com>
> 抄 送:user <us...@flink.apache.org>
> 主 题:Re: flink can't read hdfs namenode logical url
> 
> Hi,
> 
> Please double check the content of config files in YARN_CONF_DIR and HADOOP_CONF_DIR (the first one has a priority over the latter one) and that they are pointing to correct files.
> 
> Also check logs (WARN and INFO) for any relevant entries.
> 
> Piotrek
> 
> On 20 Oct 2017, at 06:07, 邓俊华 <dengjunhua@startdt.com <ma...@startdt.com>> wrote:
> 
> hi,
> 
> I start yarn-ssession.sh on yarn, but it can't read hdfs logical url. It always connect to hdfs://master:8020 <hdfs://master:8020>, it should be 9000, my hdfs defaultfs is hdfs://master <hdfs://master>.
> I have config the YARN_CONF_DIR and HADOOP_CONF_DIR, it didn't work.
> Is it a bug? i use flink-1.3.0-bin-hadoop27-scala_2.10
> 
> 2017-10-20 11:00:05,395 DEBUG org.apache.hadoop.ipc.Client                                  - IPC Client (1035144464) connection to startdt/173.16.5.215:8020 from admin: closed2017-10-20 11:00:05,398 ERROR org.apache.flink.yarn.YarnApplicationMasterRunner             - YARN Application Master initialization failedjava.net.ConnectException: Call From spark3/173.16.5.216 to master:8020 failed on connection exception: java.net.ConnectException: Connection refused; For more details see:  http://wiki.apache.org/hadoop/ConnectionRefused <http://wiki.apache.org/hadoop/ConnectionRefused>	at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)	at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57)	at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)	at java.lang.reflect.Constructor.newInstance(Constructor.java:526)	at org.apache.hadoop.net.NetUtils.wrapWithMessage(NetUtils.java:792)	at org.apache.hadoop.net.NetUtils.wrapException(NetUtils.java:732)	at org.apache.hadoop.ipc.Client.call(Client.java:1479)	at org.apache.hadoop.ipc.Client.call(Client.java:1412)	at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:229)	at com.sun.proxy.$Proxy9.getFileInfo(Unknown Source)	at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.getFileInfo(ClientNamenodeProtocolTranslatorPB.java:771)	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)	at java.lang.reflect.Method.invoke(Method.java:606)	at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:191)	at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:102)	at com.sun.proxy.$Proxy10.getFileInfo(Unknown Source)	at org.apache.hadoop.hdfs.DFSClient.getFileInfo(DFSClient.java:2108)	at org.apache.hadoop.hdfs.DistributedFileSystem$22.doCall(DistributedFileSystem.java:1305)	at org.apache.hadoop.hdfs.DistributedFileSystem$22.doCall(DistributedFileSystem.java:1301)	at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
> 
> 
> 
> 


回复:flink can't read hdfs namenode logical url

Posted by 邓俊华 <de...@startdt.com>.
Hi,
Thanks for your replay! I have been block in this for several days.And I have double checked that there are hdfs-site.xml,core-site.xml,yarn-site.xml  in YARN_CONF_DIR. But it is still can't read hdfs namenode logical url.
2017-10-23 14:35:17,750 DEBUG org.apache.flink.api.java.hadoop.mapred.utils.HadoopUtils     - Cannot find hdfs-default configuration file
2017-10-23 14:35:17,750 DEBUG org.apache.flink.api.java.hadoop.mapred.utils.HadoopUtils     - Cannot find hdfs-site configuration file
2017-10-23 14:35:17,751 DEBUG org.apache.flink.api.java.hadoop.mapred.utils.HadoopUtils     - Adding /data/hadoop-2.7.3/etc/hadoop/core-site.xml to hadoop configuration
2017-10-23 14:35:17,751 DEBUG org.apache.flink.api.java.hadoop.mapred.utils.HadoopUtils     - Adding /data/hadoop-2.7.3/etc/hadoop/hdfs-site.xml to hadoop configuration
2017-10-23 14:35:19,887 DEBUG org.apache.hadoop.hdfs.BlockReaderLocal                       - dfs.domain.socket.path = 
2017-10-23 14:35:19,952 ERROR org.apache.flink.yarn.YarnApplicationMasterRunner             - YARN Application Master initialization failed
java.lang.IllegalArgumentException: java.net.UnknownHostException: startdt
	at org.apache.hadoop.security.SecurityUtil.buildTokenService(SecurityUtil.java:378)
	at org.apache.hadoop.hdfs.NameNodeProxies.createNonHAProxy(NameNodeProxies.java:310)
	at org.apache.hadoop.hdfs.NameNodeProxies.createProxy(NameNodeProxies.java:176)
	at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:678)
	at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:619)
	at org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:149)------------------------------------------------------------------发件人:Piotr Nowojski <pi...@data-artisans.com>发送时间:2017年10月20日(星期五) 21:39收件人:邓俊华 <de...@startdt.com>抄 送:user <us...@flink.apache.org>主 题:Re: flink can't read hdfs namenode  logical url
Hi,
Please double check the content of config files in YARN_CONF_DIR and HADOOP_CONF_DIR (the first one has a priority over the latter one) and that they are pointing to correct files.
Also check logs (WARN and INFO) for any relevant entries.
Piotrek
On 20 Oct 2017, at 06:07, 邓俊华 <de...@startdt.com> wrote:
hi,
I start yarn-ssession.sh on yarn, but it can't read hdfs logical url. It always connect to hdfs://master:8020, it should be 9000, my hdfs defaultfs is hdfs://master.I have config the YARN_CONF_DIR and HADOOP_CONF_DIR, it didn't work.Is it a bug? i use flink-1.3.0-bin-hadoop27-scala_2.10
2017-10-20 11:00:05,395 DEBUG org.apache.hadoop.ipc.Client                                  - IPC Client (1035144464) connection to startdt/173.16.5.215:8020 from admin: closed2017-10-20 11:00:05,398 ERROR org.apache.flink.yarn.YarnApplicationMasterRunner             - YARN Application Master initialization failedjava.net.ConnectException: Call From spark3/173.16.5.216 to master:8020 failed on connection exception: java.net.ConnectException: Connection refused; For more details see:  http://wiki.apache.org/hadoop/ConnectionRefused	at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)	at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57)	at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)	at java.lang.reflect.Constructor.newInstance(Constructor.java:526)	at org.apache.hadoop.net.NetUtils.wrapWithMessage(NetUtils.java:792)	at org.apache.hadoop.net.NetUtils.wrapException(NetUtils.java:732)	at org.apache.hadoop.ipc.Client.call(Client.java:1479)	at org.apache.hadoop.ipc.Client.call(Client.java:1412)	at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:229)	at com.sun.proxy.$Proxy9.getFileInfo(Unknown Source)	at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.getFileInfo(ClientNamenodeProtocolTranslatorPB.java:771)	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)	at java.lang.reflect.Method.invoke(Method.java:606)	at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:191)	at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:102)	at com.sun.proxy.$Proxy10.getFileInfo(Unknown Source)	at org.apache.hadoop.hdfs.DFSClient.getFileInfo(DFSClient.java:2108)	at org.apache.hadoop.hdfs.DistributedFileSystem$22.doCall(DistributedFileSystem.java:1305)	at org.apache.hadoop.hdfs.DistributedFileSystem$22.doCall(DistributedFileSystem.java:1301)	at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)




Re: flink can't read hdfs namenode logical url

Posted by Piotr Nowojski <pi...@data-artisans.com>.
Hi,

Please double check the content of config files in YARN_CONF_DIR and HADOOP_CONF_DIR (the first one has a priority over the latter one) and that they are pointing to correct files.

Also check logs (WARN and INFO) for any relevant entries.

Piotrek

> On 20 Oct 2017, at 06:07, 邓俊华 <de...@startdt.com> wrote:
> 
> hi,
> 
> I start yarn-ssession.sh on yarn, but it can't read hdfs logical url. It always connect to hdfs://master:8020, it should be 9000, my hdfs defaultfs is hdfs://master.
> I have config the YARN_CONF_DIR and HADOOP_CONF_DIR, it didn't work.
> Is it a bug? i use flink-1.3.0-bin-hadoop27-scala_2.10
> 
> 2017-10-20 11:00:05,395 DEBUG org.apache.hadoop.ipc.Client                                  - IPC Client (1035144464) connection to startdt/173.16.5.215:8020 from admin: closed
> 2017-10-20 11:00:05,398 ERROR org.apache.flink.yarn.YarnApplicationMasterRunner             - YARN Application Master initialization failed
> java.net.ConnectException: Call From spark3/173.16.5.216 to master:8020 failed on connection exception: java.net.ConnectException: Connection refused; For more details see:  http://wiki.apache.org/hadoop/ConnectionRefused
> 	at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
> 	at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57)
> 	at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
> 	at java.lang.reflect.Constructor.newInstance(Constructor.java:526)
> 	at org.apache.hadoop.net.NetUtils.wrapWithMessage(NetUtils.java:792)
> 	at org.apache.hadoop.net.NetUtils.wrapException(NetUtils.java:732)
> 	at org.apache.hadoop.ipc.Client.call(Client.java:1479)
> 	at org.apache.hadoop.ipc.Client.call(Client.java:1412)
> 	at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:229)
> 	at com.sun.proxy.$Proxy9.getFileInfo(Unknown Source)
> 	at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.getFileInfo(ClientNamenodeProtocolTranslatorPB.java:771)
> 	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> 	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
> 	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> 	at java.lang.reflect.Method.invoke(Method.java:606)
> 	at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:191)
> 	at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:102)
> 	at com.sun.proxy.$Proxy10.getFileInfo(Unknown Source)
> 	at org.apache.hadoop.hdfs.DFSClient.getFileInfo(DFSClient.java:2108)
> 	at org.apache.hadoop.hdfs.DistributedFileSystem$22.doCall(DistributedFileSystem.java:1305)
> 	at org.apache.hadoop.hdfs.DistributedFileSystem$22.doCall(DistributedFileSystem.java:1301)
> 	at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
> 
>