You are viewing a plain text version of this content. The canonical link for it is here.
Posted to hdfs-dev@hadoop.apache.org by "Daniel Einspanjer (JIRA)" <ji...@apache.org> on 2011/02/17 23:17:12 UTC

[jira] Created: (HDFS-1631) DFSClient errors when a block replica disappears could be more informative and less scary.

DFSClient errors when a block replica disappears could be more informative and less scary.
------------------------------------------------------------------------------------------

                 Key: HDFS-1631
                 URL: https://issues.apache.org/jira/browse/HDFS-1631
             Project: Hadoop HDFS
          Issue Type: Improvement
          Components: hdfs client
    Affects Versions: 0.20-append
            Reporter: Daniel Einspanjer
            Priority: Minor


When HBase starts up, it opens a ton of files from HDFS.  If you are closely monitoring the startup, you will see lots of the following example of logs on the client and remote DN.  The exception is logged at an INFO level on the client side, but at a WARN and ERROR on the DN side.  In both cases, it would be nice to put a little more context into the error and reassure the user that it isn't anything to be worried about if the error doesn't persist.  Also, why bother with a complete stack trace?

2011-02-17 13:46:30,995 INFO org.apache.hadoop.hbase.regionserver.HRegionServer: Worker: MSG_REGION_OPEN: crash_reports,81101
138606ab7a-85e4-416f-8cd9-bb4002110113,1294954735047.184604411                                                               
2011-02-17 13:46:31,056 INFO org.apache.hadoop.hdfs.DFSClient: Failed to connect to /10.8.100.16:50010, add to deadNodes and 
continue                                                                                                                     
java.io.IOException: Got error in response to OP_READ_BLOCK self=/10.8.100.13:54349, remote=/10.8.100.16:50010 for file /hbas
e/crash_reports/184604411/meta_data/8999005319667206647 for block -3256062388303907746_2989166                               
at org.apache.hadoop.hdfs.DFSClient$BlockReader.newBlockReader(DFSClient.java:1471)                                          
at org.apache.hadoop.hdfs.DFSClient$DFSInputStream.blockSeekTo(DFSClient.java:1794)                                          
at org.apache.hadoop.hdfs.DFSClient$DFSInputStream.read(DFSClient.java:1931)                                                 
at java.io.DataInputStream.readFully(DataInputStream.java:178)                                                               
at java.io.DataInputStream.readFully(DataInputStream.java:152)                                                               
at org.apache.hadoop.hbase.io.hfile.HFile$FixedFileTrailer.deserialize(HFile.java:1492)                                      
at org.apache.hadoop.hbase.io.hfile.HFile$Reader.readTrailer(HFile.java:860)                                                 
at org.apache.hadoop.hbase.io.hfile.HFile$Reader.loadFileInfo(HFile.java:805)                                                
at org.apache.hadoop.hbase.regionserver.StoreFile$Reader.loadFileInfo(StoreFile.java:977)                                    
at org.apache.hadoop.hbase.regionserver.StoreFile.open(StoreFile.java:387)                                                   
at org.apache.hadoop.hbase.regionserver.StoreFile.createReader(StoreFile.java:438)                                           
at org.apache.hadoop.hbase.regionserver.Store.loadStoreFiles(Store.java:245)                                                 
at org.apache.hadoop.hbase.regionserver.Store.<init>(Store.java:187)                                                         
at org.apache.hadoop.hbase.regionserver.HRegion.instantiateHStore(HRegion.java:1923)                                         
at org.apache.hadoop.hbase.regionserver.HRegion.initialize(HRegion.java:334)                                                 
at org.apache.hadoop.hbase.regionserver.HRegionServer.instantiateRegion(HRegionServer.java:1588)                             
at org.apache.hadoop.hbase.regionserver.HRegionServer.openRegion(HRegionServer.java:1553)                                    
at org.apache.hadoop.hbase.regionserver.HRegionServer$Worker.run(HRegionServer.java:1465)                                    
at java.lang.Thread.run(Thread.java:619)                                                                                     
2011-02-17 13:46:31,456 INFO org.apache.hadoop.hbase.regionserver.HRegion: Onlined crash_reports,81101138606ab7a-85e4-416f-8c
d9-bb4002110113,1294954735047.184604411; next sequenceid=4358482563



On node .16:
2011-02-17 13:46:31,056 WARN org.apache.hadoop.hdfs.server.datanode.DataNode: DatanodeRegistration(10.8.100.16:50010, storageID=DS-2053327189-10.8.100.16-50010-1291655260617, infoPort=50075, ipcPort=50020):Got exception while serving blk_-3256062388303907746_2989166 to /10.8.100.13:
java.io.IOException: Block blk_-3256062388303907746_2989166 is not valid.
        at org.apache.hadoop.hdfs.server.datanode.FSDataset.getBlockFile(FSDataset.java:976)
        at org.apache.hadoop.hdfs.server.datanode.FSDataset.getLength(FSDataset.java:939)
        at org.apache.hadoop.hdfs.server.datanode.FSDataset.getVisibleLength(FSDataset.java:949)
        at org.apache.hadoop.hdfs.server.datanode.BlockSender.<init>(BlockSender.java:94)
        at org.apache.hadoop.hdfs.server.datanode.DataXceiver.readBlock(DataXceiver.java:206)
        at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:114)

2011-02-17 13:46:31,056 ERROR org.apache.hadoop.hdfs.server.datanode.DataNode: DatanodeRegistration(10.8.100.16:50010, storageID=DS-2053327189-10.8.100.16-50010-1291655260617, infoPort=50075, ipcPort=50020):DataXceiver
java.io.IOException: Block blk_-3256062388303907746_2989166 is not valid.
        at org.apache.hadoop.hdfs.server.datanode.FSDataset.getBlockFile(FSDataset.java:976)
        at org.apache.hadoop.hdfs.server.datanode.FSDataset.getLength(FSDataset.java:939)
        at org.apache.hadoop.hdfs.server.datanode.FSDataset.getVisibleLength(FSDataset.java:949)
        at org.apache.hadoop.hdfs.server.datanode.BlockSender.<init>(BlockSender.java:94)
        at org.apache.hadoop.hdfs.server.datanode.DataXceiver.readBlock(DataXceiver.java:206)
        at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:114)

-- 
This message is automatically generated by JIRA.
-
For more information on JIRA, see: http://www.atlassian.com/software/jira