You are viewing a plain text version of this content. The canonical link for it is here.
Posted to hdfs-dev@hadoop.apache.org by "Yongjun Zhang (JIRA)" <ji...@apache.org> on 2014/06/14 18:11:02 UTC
[jira] [Created] (HDFS-6532) Intermittent test failure
org.apache.hadoop.hdfs.TestCrcCorruption.testCorruptionDuringWrt
Yongjun Zhang created HDFS-6532:
-----------------------------------
Summary: Intermittent test failure org.apache.hadoop.hdfs.TestCrcCorruption.testCorruptionDuringWrt
Key: HDFS-6532
URL: https://issues.apache.org/jira/browse/HDFS-6532
Project: Hadoop HDFS
Issue Type: Bug
Components: datanode, hdfs-client
Affects Versions: 2.4.0
Reporter: Yongjun Zhang
Per https://builds.apache.org/job/Hadoop-Hdfs-trunk/1774/testReport
{code}
Regression
org.apache.hadoop.hdfs.TestCrcCorruption.testCorruptionDuringWrt
Failing for the past 1 build (Since Failed#1774 )
Took 50 sec.
Error Message
test timed out after 50000 milliseconds
Stacktrace
java.lang.Exception: test timed out after 50000 milliseconds
at java.lang.Object.wait(Native Method)
at org.apache.hadoop.hdfs.DFSOutputStream.waitForAckedSeqno(DFSOutputStream.java:2024)
at org.apache.hadoop.hdfs.DFSOutputStream.flushInternal(DFSOutputStream.java:2008)
at org.apache.hadoop.hdfs.DFSOutputStream.close(DFSOutputStream.java:2107)
at org.apache.hadoop.fs.FSDataOutputStream$PositionCache.close(FSDataOutputStream.java:70)
at org.apache.hadoop.fs.FSDataOutputStream.close(FSDataOutputStream.java:98)
at org.apache.hadoop.hdfs.TestCrcCorruption.testCorruptionDuringWrt(TestCrcCorruption.java:133)
{code}
See relevant exceptions in log
{code}
2014-06-14 11:56:15,283 WARN datanode.DataNode (BlockReceiver.java:verifyChunks(404)) - Checksum error in block BP-1675558312-67.195.138.30-1402746971712:blk_1073741825_1001 from /127.0.0.1:41708
org.apache.hadoop.fs.ChecksumException: Checksum error: DFSClient_NONMAPREDUCE_-1139495951_8 at 64512 exp: 1379611785 got: -12163112
at org.apache.hadoop.util.DataChecksum.verifyChunkedSums(DataChecksum.java:353)
at org.apache.hadoop.util.DataChecksum.verifyChunkedSums(DataChecksum.java:284)
at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.verifyChunks(BlockReceiver.java:402)
at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:537)
at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:734)
at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:741)
at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:124)
at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:71)
at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:234)
at java.lang.Thread.run(Thread.java:662)
2014-06-14 11:56:15,285 WARN datanode.DataNode (BlockReceiver.java:run(1207)) - IOException in BlockReceiver.run():
java.io.IOException: Shutting down writer and responder due to a checksum error in received data. The error response has been sent upstream.
at org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.sendAckUpstreamUnprotected(BlockReceiver.java:1352)
at org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.sendAckUpstream(BlockReceiver.java:1278)
at org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.run(BlockReceiver.java:1199)
at java.lang.Thread.run(Thread.java:662)
...
{code}
--
This message was sent by Atlassian JIRA
(v6.2#6252)