You are viewing a plain text version of this content. The canonical link for it is here.
Posted to common-user@hadoop.apache.org by Krzysztof Szlapinski <kr...@starline.hk> on 2008/11/19 15:45:24 UTC
crashes at a heavy load
hi
im running some test of hbase on hadoop
when namenode is heavily overloaded it crushes with the following error
java.io.IOException: Connection reset by peer
at sun.nio.ch.FileDispatcher.write0(Native Method)
at sun.nio.ch.SocketDispatcher.write(SocketDispatcher.java:29)
at sun.nio.ch.IOUtil.writeFromNativeBuffer(IOUtil.java:104)
at sun.nio.ch.IOUtil.write(IOUtil.java:75)
at sun.nio.ch.SocketChannelImpl.write(SocketChannelImpl.java:334)
at org.apache.hadoop.net.SocketOutputStream$Writer.performIO(SocketOutputStream.java:55)
at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:140)
at org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:146)
at org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:107)
at java.io.BufferedOutputStream.flushBuffer(BufferedOutputStream.java:65)
at java.io.BufferedOutputStream.flush(BufferedOutputStream.java:123)
at java.io.DataOutputStream.flush(DataOutputStream.java:106)
at org.apache.hadoop.dfs.DataNode$BlockSender.sendBlock(DataNode.java:2019)
at org.apache.hadoop.dfs.DataNode$DataXceiver.readBlock(DataNode.java:1140)
at org.apache.hadoop.dfs.DataNode$DataXceiver.run(DataNode.java:1068)
at java.lang.Thread.run(Thread.java:619)
2008-11-19 14:52:42,007 ERROR org.apache.hadoop.dfs.DataNode: DatanodeRegistration(192.168.1.109:50010, storageID=DS-1340454997-192.168.1.109-50010-122709122
1665, infoPort=50075, ipcPort=50020):DataXceiver: java.io.IOException: Connection reset by peer
at sun.nio.ch.FileDispatcher.write0(Native Method)
at sun.nio.ch.SocketDispatcher.write(SocketDispatcher.java:29)
at sun.nio.ch.IOUtil.writeFromNativeBuffer(IOUtil.java:104)
at sun.nio.ch.IOUtil.write(IOUtil.java:75)
at sun.nio.ch.SocketChannelImpl.write(SocketChannelImpl.java:334)
at org.apache.hadoop.net.SocketOutputStream$Writer.performIO(SocketOutputStream.java:55)
at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:140)
at org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:146)
at org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:107)
at java.io.BufferedOutputStream.flushBuffer(BufferedOutputStream.java:65)
at java.io.BufferedOutputStream.flush(BufferedOutputStream.java:123)
at java.io.DataOutputStream.flush(DataOutputStream.java:106)
at org.apache.hadoop.dfs.DataNode$BlockSender.sendBlock(DataNode.java:2019)
at org.apache.hadoop.dfs.DataNode$DataXceiver.readBlock(DataNode.java:1140)
at org.apache.hadoop.dfs.DataNode$DataXceiver.run(DataNode.java:1068)
at java.lang.Thread.run(Thread.java:619)
and throws Could not obtain blk as a result
Can I tune hadoop in some way to prevent this error?
Im using the trunk version of hadoop
krzysiek