You are viewing a plain text version of this content. The canonical link for it is here.
Posted to user@hadoop.apache.org by rauljin <li...@sina.com> on 2013/04/16 05:24:19 UTC
threads quota is exceeded question
HI:
The hadoop cluster is running balance.
And one datannode 172.16.80.72 is :
Datanode :Not able to copy block -507744952197054725 to /172.16.80.73:51658 because threads quota is exceeded.
ERROR org.apache.hadoop.hdfs.server.datanode.DataNode: DatanodeRegistration(172.16.80.72:50010, storageID=DS-1202844662-172.16
.80.72-50010-1330656432004, infoPort=50075, ipcPort=50020):DataXceiver
java.io.IOException: Block blk_8443528692263789109_8159545 is not valid.
at org.apache.hadoop.hdfs.server.datanode.FSDataset.getBlockFile(FSDataset.java:734)
at org.apache.hadoop.hdfs.server.datanode.FSDataset.getLength(FSDataset.java:722)
at org.apache.hadoop.hdfs.server.datanode.BlockSender.<init>(BlockSender.java:92)
at org.apache.hadoop.hdfs.server.datanode.DataXceiver.readBlock(DataXceiver.java:172)
at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:95)
at java.lang.Thread.run(Thread.java:636)
And other datanode:
ERROR org.apache.hadoop.hdfs.server.datanode.DataNode: DatanodeRegistration(172.16.80.73:50010, storageID=DS-1771394657-172.16.80.73-50010-1362474580654, infoPort=50075, ipcPort=50020):DataXceiver
java.io.EOFException
at java.io.DataInputStream.readByte(DataInputStream.java:267)
at org.apache.hadoop.util.DataChecksum.newDataChecksum(DataChecksum.java:84)
at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.<init>(BlockReceiver.java:92)
at org.apache.hadoop.hdfs.server.datanode.DataXceiver.replaceBlock(DataXceiver.java:580)
at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:115)
at java.lang.Thread.run(Thread.java:636)
At that moment , the hdfs is not avalible.
I restart the 172.16.80.72 datanode service ,and the service is ok.
what causes the problem.
Any ideas?
Thanks!
rauljin
Re: threads quota is exceeded question
Posted by Thanh Do <th...@cs.wisc.edu>.
Hadoop by default limit 5 concurrent threads per node for balancing
purpose. That causes your problem.
On Mon, Apr 15, 2013 at 10:24 PM, rauljin <li...@sina.com> wrote:
> **
> HI:
> The hadoop cluster is running balance.
>
> And one datannode 172.16.80.72 is :
>
> Datanode :Not able to copy block -507744952197054725 to /
> 172.16.80.73:51658 because threads quota is exceeded.
>
>
> ERROR org.apache.hadoop.hdfs.server.datanode.DataNode: DatanodeRegistration(
> 172.16.80.72:50010, storageID=DS-1202844662-172.16
> .80.72-50010-1330656432004, infoPort=50075, ipcPort=50020):DataXceiver
> java.io.IOException: Block blk_8443528692263789109_8159545 is not valid.
>
> at org.apache.hadoop.hdfs.server.datanode.FSDataset.getBlockFile(FSDataset.java:734)
>
> at org.apache.hadoop.hdfs.server.datanode.FSDataset.getLength(FSDataset.java:722)
>
> at org.apache.hadoop.hdfs.server.datanode.BlockSender.<init>(BlockSender.java:92)
>
> at org.apache.hadoop.hdfs.server.datanode.DataXceiver.readBlock(DataXceiver.java:172)
>
> at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:95)
> at java.lang.Thread.run(Thread.java:636)
>
>
> And other datanode:
>
>
>
> ERROR org.apache.hadoop.hdfs.server.datanode.DataNode: DatanodeRegistration(
> 172.16.80.73:50010
> , storageID=DS-1771394657-172.16.80.73-50010-1362474580654, infoPort=50075, ipcPort=50020):DataXceiver
> java.io.EOFException
> at java.io.DataInputStream.readByte(DataInputStream.java:267)
>
> at org.apache.hadoop.util.DataChecksum.newDataChecksum(DataChecksum.java:84)
>
> at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.<init>(BlockReceiver.java:92)
>
> at org.apache.hadoop.hdfs.server.datanode.DataXceiver.replaceBlock(DataXceiver.java:580)
>
> at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:115)
> at java.lang.Thread.run(Thread.java:636)
>
> At that moment , the hdfs is not avalible.
>
>
> I restart the 172.16.80.72 datanode service ,and the service is ok.
>
>
>
> what causes the problem.
>
> Any ideas?
> Thanks!
>
>
>
>
>
>
>
>
>
>
> ------------------------------
> rauljin
>
Re: threads quota is exceeded question
Posted by Thanh Do <th...@cs.wisc.edu>.
Hadoop by default limit 5 concurrent threads per node for balancing
purpose. That causes your problem.
On Mon, Apr 15, 2013 at 10:24 PM, rauljin <li...@sina.com> wrote:
> **
> HI:
> The hadoop cluster is running balance.
>
> And one datannode 172.16.80.72 is :
>
> Datanode :Not able to copy block -507744952197054725 to /
> 172.16.80.73:51658 because threads quota is exceeded.
>
>
> ERROR org.apache.hadoop.hdfs.server.datanode.DataNode: DatanodeRegistration(
> 172.16.80.72:50010, storageID=DS-1202844662-172.16
> .80.72-50010-1330656432004, infoPort=50075, ipcPort=50020):DataXceiver
> java.io.IOException: Block blk_8443528692263789109_8159545 is not valid.
>
> at org.apache.hadoop.hdfs.server.datanode.FSDataset.getBlockFile(FSDataset.java:734)
>
> at org.apache.hadoop.hdfs.server.datanode.FSDataset.getLength(FSDataset.java:722)
>
> at org.apache.hadoop.hdfs.server.datanode.BlockSender.<init>(BlockSender.java:92)
>
> at org.apache.hadoop.hdfs.server.datanode.DataXceiver.readBlock(DataXceiver.java:172)
>
> at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:95)
> at java.lang.Thread.run(Thread.java:636)
>
>
> And other datanode:
>
>
>
> ERROR org.apache.hadoop.hdfs.server.datanode.DataNode: DatanodeRegistration(
> 172.16.80.73:50010
> , storageID=DS-1771394657-172.16.80.73-50010-1362474580654, infoPort=50075, ipcPort=50020):DataXceiver
> java.io.EOFException
> at java.io.DataInputStream.readByte(DataInputStream.java:267)
>
> at org.apache.hadoop.util.DataChecksum.newDataChecksum(DataChecksum.java:84)
>
> at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.<init>(BlockReceiver.java:92)
>
> at org.apache.hadoop.hdfs.server.datanode.DataXceiver.replaceBlock(DataXceiver.java:580)
>
> at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:115)
> at java.lang.Thread.run(Thread.java:636)
>
> At that moment , the hdfs is not avalible.
>
>
> I restart the 172.16.80.72 datanode service ,and the service is ok.
>
>
>
> what causes the problem.
>
> Any ideas?
> Thanks!
>
>
>
>
>
>
>
>
>
>
> ------------------------------
> rauljin
>
Re: threads quota is exceeded question
Posted by Thanh Do <th...@cs.wisc.edu>.
Hadoop by default limit 5 concurrent threads per node for balancing
purpose. That causes your problem.
On Mon, Apr 15, 2013 at 10:24 PM, rauljin <li...@sina.com> wrote:
> **
> HI:
> The hadoop cluster is running balance.
>
> And one datannode 172.16.80.72 is :
>
> Datanode :Not able to copy block -507744952197054725 to /
> 172.16.80.73:51658 because threads quota is exceeded.
>
>
> ERROR org.apache.hadoop.hdfs.server.datanode.DataNode: DatanodeRegistration(
> 172.16.80.72:50010, storageID=DS-1202844662-172.16
> .80.72-50010-1330656432004, infoPort=50075, ipcPort=50020):DataXceiver
> java.io.IOException: Block blk_8443528692263789109_8159545 is not valid.
>
> at org.apache.hadoop.hdfs.server.datanode.FSDataset.getBlockFile(FSDataset.java:734)
>
> at org.apache.hadoop.hdfs.server.datanode.FSDataset.getLength(FSDataset.java:722)
>
> at org.apache.hadoop.hdfs.server.datanode.BlockSender.<init>(BlockSender.java:92)
>
> at org.apache.hadoop.hdfs.server.datanode.DataXceiver.readBlock(DataXceiver.java:172)
>
> at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:95)
> at java.lang.Thread.run(Thread.java:636)
>
>
> And other datanode:
>
>
>
> ERROR org.apache.hadoop.hdfs.server.datanode.DataNode: DatanodeRegistration(
> 172.16.80.73:50010
> , storageID=DS-1771394657-172.16.80.73-50010-1362474580654, infoPort=50075, ipcPort=50020):DataXceiver
> java.io.EOFException
> at java.io.DataInputStream.readByte(DataInputStream.java:267)
>
> at org.apache.hadoop.util.DataChecksum.newDataChecksum(DataChecksum.java:84)
>
> at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.<init>(BlockReceiver.java:92)
>
> at org.apache.hadoop.hdfs.server.datanode.DataXceiver.replaceBlock(DataXceiver.java:580)
>
> at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:115)
> at java.lang.Thread.run(Thread.java:636)
>
> At that moment , the hdfs is not avalible.
>
>
> I restart the 172.16.80.72 datanode service ,and the service is ok.
>
>
>
> what causes the problem.
>
> Any ideas?
> Thanks!
>
>
>
>
>
>
>
>
>
>
> ------------------------------
> rauljin
>
Re: threads quota is exceeded question
Posted by Thanh Do <th...@cs.wisc.edu>.
Hadoop by default limit 5 concurrent threads per node for balancing
purpose. That causes your problem.
On Mon, Apr 15, 2013 at 10:24 PM, rauljin <li...@sina.com> wrote:
> **
> HI:
> The hadoop cluster is running balance.
>
> And one datannode 172.16.80.72 is :
>
> Datanode :Not able to copy block -507744952197054725 to /
> 172.16.80.73:51658 because threads quota is exceeded.
>
>
> ERROR org.apache.hadoop.hdfs.server.datanode.DataNode: DatanodeRegistration(
> 172.16.80.72:50010, storageID=DS-1202844662-172.16
> .80.72-50010-1330656432004, infoPort=50075, ipcPort=50020):DataXceiver
> java.io.IOException: Block blk_8443528692263789109_8159545 is not valid.
>
> at org.apache.hadoop.hdfs.server.datanode.FSDataset.getBlockFile(FSDataset.java:734)
>
> at org.apache.hadoop.hdfs.server.datanode.FSDataset.getLength(FSDataset.java:722)
>
> at org.apache.hadoop.hdfs.server.datanode.BlockSender.<init>(BlockSender.java:92)
>
> at org.apache.hadoop.hdfs.server.datanode.DataXceiver.readBlock(DataXceiver.java:172)
>
> at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:95)
> at java.lang.Thread.run(Thread.java:636)
>
>
> And other datanode:
>
>
>
> ERROR org.apache.hadoop.hdfs.server.datanode.DataNode: DatanodeRegistration(
> 172.16.80.73:50010
> , storageID=DS-1771394657-172.16.80.73-50010-1362474580654, infoPort=50075, ipcPort=50020):DataXceiver
> java.io.EOFException
> at java.io.DataInputStream.readByte(DataInputStream.java:267)
>
> at org.apache.hadoop.util.DataChecksum.newDataChecksum(DataChecksum.java:84)
>
> at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.<init>(BlockReceiver.java:92)
>
> at org.apache.hadoop.hdfs.server.datanode.DataXceiver.replaceBlock(DataXceiver.java:580)
>
> at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:115)
> at java.lang.Thread.run(Thread.java:636)
>
> At that moment , the hdfs is not avalible.
>
>
> I restart the 172.16.80.72 datanode service ,and the service is ok.
>
>
>
> what causes the problem.
>
> Any ideas?
> Thanks!
>
>
>
>
>
>
>
>
>
>
> ------------------------------
> rauljin
>