You are viewing a plain text version of this content. The canonical link for it is here.
Posted to common-commits@hadoop.apache.org by Apache Wiki <wi...@apache.org> on 2011/09/14 11:58:17 UTC

[Hadoop Wiki] Update of "TroubleShooting" by SteveLoughran

Dear Wiki user,

You have subscribed to a wiki page or wiki category on "Hadoop Wiki" for change notification.

The "TroubleShooting" page has been changed by SteveLoughran:
http://wiki.apache.org/hadoop/TroubleShooting?action=diff&rev1=13&rev2=14

Comment:
x-ref DataNode text to the page, mention resolv.conf. 

  
  There are a number of possible of causes for this.
   * The NameNode may be overloaded. Check the logs for messages that say "discarding calls..."
-  * There may not be enough (any) DataNodes for the data to be written. Again, check the logs.
+  * There may not be enough (any) DataNode nodes running for the data to be written. Again, check the logs.
-  * The DataNodes on which the blocks were stored might be down.
+  * Every DataNode on which the blocks were stored might be down (or not connected to the NameNode; it is impossible to distinguish the two).
  
  === Error message: Could not obtain block ===
  
@@ -62, +62 @@

   java.io.IOException: No live nodes contain current block 
  }}}
  
- There are no live DataNodes containing a copy of the block of the file you are looking for. Bring up any nodes that are down, or skip that block.
+ There are no live DataNode nodes containing a copy of the block of the file you are looking for. Bring up any nodes that are down, or skip that block.
  
  == Reduce hangs ==
  
  This can be a DNS issue. Two problems which have been encountered in practice are:
   * Machines with multiple NICs. In this case, set {{{ dfs.datanode.dns.interface }}} (in {{{ hdfs-site.xml }}}) and {{{ mapred.datanode.dns.interface }}} (in {{{ mapred-site.xml }}}) to the name of the network interface used by Hadoop (something like {{{ eth0 }}} under Linux),
-  * Badly formatted or incorrect hosts files ({{{ /etc/hosts }}} under Linux) can wreak havoc. Any DNS problem will hobble Hadoop, so ensure that names can be resolved correctly.
+  * Badly formatted or incorrect hosts and DNS files ({{{ /etc/hosts }}} and {{{{ /etc/resolv.conf }}}under Linux) can wreak havoc. Any DNS problem will hobble Hadoop, so ensure that names can be resolved correctly.
  
  == Error message saying a file "Could only be replicated to 0 nodes instead of 1" ==