You are viewing a plain text version of this content. The canonical link for it is here.
Posted to hdfs-dev@hadoop.apache.org by "Frantisek Vacek (JIRA)" <ji...@apache.org> on 2014/11/12 13:55:33 UTC

[jira] [Created] (HDFS-7392) org.apache.hadoop.hdfs.DistributedFileSystem open invalid URI forever

Frantisek Vacek created HDFS-7392:
-------------------------------------

             Summary: org.apache.hadoop.hdfs.DistributedFileSystem open invalid URI forever
                 Key: HDFS-7392
                 URL: https://issues.apache.org/jira/browse/HDFS-7392
             Project: Hadoop HDFS
          Issue Type: Bug
          Components: hdfs-client
            Reporter: Frantisek Vacek
            Priority: Critical


In some specific circumstances, org.apache.hadoop.hdfs.DistributedFileSystem.open(invalid URI) never timeouts and last forever.

What are specific circumstances:
1) HDFS URI (hdfs://share.merck.com:8020/someDir/someFile.txt) should point to valid IP address but without name node service running on it.
2) There should be at least 2 IP addresses for such a URI. See output below:

[~/proj/quickbox]$ nslookup share.merck.com
Server:         127.0.1.1
Address:        127.0.1.1#53

share.merck.com canonical name = internal-gicprg-share-merck-com-1538706884.us-east-1.elb.amazonaws.com.
Name:   internal-gicprg-share-merck-com-1538706884.us-east-1.elb.amazonaws.com
Address: 54.40.29.223
Name:   internal-gicprg-share-merck-com-1538706884.us-east-1.elb.amazonaws.com
Address: 54.40.29.65

In such a case the org.apache.hadoop.ipc.Client.Connection.updateAddress() returns sometimes true (even if address didn't actually changed see img. 1) and the timeoutFailures counter is set to 0 (see img. 2). The maxRetriesOnSocketTimeouts (45) is never reached and connection attempt is repeated forever.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)