You are viewing a plain text version of this content. The canonical link for it is here.
Posted to general@hadoop.apache.org by Terrence Martin <tm...@physics.ucsd.edu> on 2009/03/20 18:19:28 UTC

Java exception when starting the namenode

I am seeing the following exception when I attempt to start my Hadoop namenode. This is a namenode that was working and the system had a few TB of files in it. Yesterday we were attempting to upgrade it to a more recent version and things appeared to be working. 

2009-03-20 10:14:02,667 INFO org.apache.hadoop.hdfs.server.namenode.NameNode: STARTUP_MSG:  
/************************************************************                               
STARTUP_MSG: Starting NameNode                                                              
STARTUP_MSG:   host = proxy-1.t2.ucsd.edu/169.228.130.63                                    
STARTUP_MSG:   args = []                                                                    
STARTUP_MSG:   version = 0.19.2-dev                                                         
STARTUP_MSG:   build = http://svn.apache.org/repos/asf/hadoop/core/tags/release-0.19.1 -r 748415; compiled by 'sl1-user' on Tue Mar 17 10:49:06 CDT 2009                                                                                                                                
************************************************************/                                                                               
2009-03-20 10:14:02,772 INFO org.apache.hadoop.ipc.metrics.RpcMetrics: Initializing RPC Metrics with hostName=NameNode, port=9000           
2009-03-20 10:14:02,776 INFO org.apache.hadoop.hdfs.server.namenode.NameNode: Namenode up at: proxy-1.t2.ucsd.edu/169.228.130.63:9000       
2009-03-20 10:14:02,778 INFO org.apache.hadoop.metrics.jvm.JvmMetrics: Initializing JVM Metrics with processName=NameNode, sessionId=Hadoop 
2009-03-20 10:14:02,783 INFO org.apache.hadoop.hdfs.server.namenode.metrics.NameNodeMetrics: Initializing NameNodeMeterics using context object:org.apache.hadoop.metrics.spi.NullContext                                                                                               
2009-03-20 10:14:02,832 INFO org.apache.hadoop.hdfs.server.namenode.FSNamesystem: fsOwner=root,root,bin,daemon,sys,adm,disk,wheel           
2009-03-20 10:14:02,832 INFO org.apache.hadoop.hdfs.server.namenode.FSNamesystem: supergroup=supergroup                                     
2009-03-20 10:14:02,832 INFO org.apache.hadoop.hdfs.server.namenode.FSNamesystem: isPermissionEnabled=true                                  
2009-03-20 10:14:02,839 INFO org.apache.hadoop.hdfs.server.namenode.metrics.FSNamesystemMetrics: Initializing FSNamesystemMetrics using context object:org.apache.hadoop.metrics.spi.NullContext                                                                                        
2009-03-20 10:14:02,840 INFO org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Registered FSNamesystemStatusMBean                        
2009-03-20 10:14:02,866 INFO org.apache.hadoop.hdfs.server.common.Storage: Number of files = 1435                                           
2009-03-20 10:14:03,061 INFO org.apache.hadoop.hdfs.server.common.Storage: Number of files under construction = 3                           
2009-03-20 10:14:03,065 INFO org.apache.hadoop.hdfs.server.common.Storage: Image file of size 361956 loaded in 0 seconds.
2009-03-20 10:14:03,074 ERROR org.apache.hadoop.hdfs.server.namenode.NameNode: java.lang.NullPointerException
        at org.apache.hadoop.hdfs.server.namenode.FSDirectory.addChild(FSDirectory.java:1006)
        at org.apache.hadoop.hdfs.server.namenode.FSDirectory.addNode(FSDirectory.java:982)
        at org.apache.hadoop.hdfs.server.namenode.FSDirectory.unprotectedAddFile(FSDirectory.java:194)
        at org.apache.hadoop.hdfs.server.namenode.FSEditLog.loadFSEdits(FSEditLog.java:613)
        at org.apache.hadoop.hdfs.server.namenode.FSImage.loadFSEdits(FSImage.java:973)
        at org.apache.hadoop.hdfs.server.namenode.FSImage.loadFSImage(FSImage.java:793)
        at org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:352)
        at org.apache.hadoop.hdfs.server.namenode.FSDirectory.loadFSImage(FSDirectory.java:87)
        at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.initialize(FSNamesystem.java:309)
        at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.<init>(FSNamesystem.java:288)
        at org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:163)
        at org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:208)
        at org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:194)
        at org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:859)
        at org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:868)

2009-03-20 10:14:03,075 INFO org.apache.hadoop.hdfs.server.namenode.NameNode: SHUTDOWN_MSG:
/************************************************************
SHUTDOWN_MSG: Shutting down NameNode at proxy-1.t2.ucsd.edu/169.228.130.63
************************************************************/

Could this be some sort of corruption on loading the image? 

Terrence