You are viewing a plain text version of this content. The canonical link for it is here.
Posted to hdfs-dev@hadoop.apache.org by "Yongjun Zhang (JIRA)" <ji...@apache.org> on 2014/06/14 18:11:02 UTC

[jira] [Created] (HDFS-6532) Intermittent test failure org.apache.hadoop.hdfs.TestCrcCorruption.testCorruptionDuringWrt

Yongjun Zhang created HDFS-6532:
-----------------------------------

             Summary: Intermittent test failure org.apache.hadoop.hdfs.TestCrcCorruption.testCorruptionDuringWrt
                 Key: HDFS-6532
                 URL: https://issues.apache.org/jira/browse/HDFS-6532
             Project: Hadoop HDFS
          Issue Type: Bug
          Components: datanode, hdfs-client
    Affects Versions: 2.4.0
            Reporter: Yongjun Zhang


Per https://builds.apache.org/job/Hadoop-Hdfs-trunk/1774/testReport

{code}
Regression

org.apache.hadoop.hdfs.TestCrcCorruption.testCorruptionDuringWrt

Failing for the past 1 build (Since Failed#1774 )
Took 50 sec.
Error Message

test timed out after 50000 milliseconds
Stacktrace

java.lang.Exception: test timed out after 50000 milliseconds
	at java.lang.Object.wait(Native Method)
	at org.apache.hadoop.hdfs.DFSOutputStream.waitForAckedSeqno(DFSOutputStream.java:2024)
	at org.apache.hadoop.hdfs.DFSOutputStream.flushInternal(DFSOutputStream.java:2008)
	at org.apache.hadoop.hdfs.DFSOutputStream.close(DFSOutputStream.java:2107)
	at org.apache.hadoop.fs.FSDataOutputStream$PositionCache.close(FSDataOutputStream.java:70)
	at org.apache.hadoop.fs.FSDataOutputStream.close(FSDataOutputStream.java:98)
	at org.apache.hadoop.hdfs.TestCrcCorruption.testCorruptionDuringWrt(TestCrcCorruption.java:133)
{code}

See relevant exceptions in log
{code}
2014-06-14 11:56:15,283 WARN  datanode.DataNode (BlockReceiver.java:verifyChunks(404)) - Checksum error in block BP-1675558312-67.195.138.30-1402746971712:blk_1073741825_1001 from /127.0.0.1:41708
org.apache.hadoop.fs.ChecksumException: Checksum error: DFSClient_NONMAPREDUCE_-1139495951_8 at 64512 exp: 1379611785 got: -12163112
	at org.apache.hadoop.util.DataChecksum.verifyChunkedSums(DataChecksum.java:353)
	at org.apache.hadoop.util.DataChecksum.verifyChunkedSums(DataChecksum.java:284)
	at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.verifyChunks(BlockReceiver.java:402)
	at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:537)
	at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:734)
	at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:741)
	at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:124)
	at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:71)
	at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:234)
	at java.lang.Thread.run(Thread.java:662)
2014-06-14 11:56:15,285 WARN  datanode.DataNode (BlockReceiver.java:run(1207)) - IOException in BlockReceiver.run(): 
java.io.IOException: Shutting down writer and responder due to a checksum error in received data. The error response has been sent upstream.
	at org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.sendAckUpstreamUnprotected(BlockReceiver.java:1352)
	at org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.sendAckUpstream(BlockReceiver.java:1278)
	at org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.run(BlockReceiver.java:1199)
	at java.lang.Thread.run(Thread.java:662)
...
{code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)