You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@slider.apache.org by "Steve Loughran (JIRA)" <ji...@apache.org> on 2014/08/29 12:27:53 UTC
[jira] [Commented] (SLIDER-377) slider MiniHDFSCluster tests
failing on windows+branch2
[ https://issues.apache.org/jira/browse/SLIDER-377?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14115087#comment-14115087 ]
Steve Loughran commented on SLIDER-377:
---------------------------------------
Test {{mvn test -Dtest=TestConfPersisterLocksHDFS#testAcqAcqRelReadlock }}
{code}
/Work/slider/slider-core/target/hdfs/TestConfPersister/data/data2/]] heartbeating to /127.0.0.1:52834] INFO util.GSet (LightWeightG
Set.java:computeCapacity(361)) - capacity = 2^19 = 524288 entries
2014-08-29 11:24:15,800 [DataXceiver for client DFSClient_NONMAPREDUCE_-1927566207_1 at /127.0.0.1:52858 [Receiving block BP-12632692
4-192.168.1.138-1409307851902:blk_1073741825_1001]] ERROR datanode.DataNode (DataXceiver.java:run(243)) - 127.0.0.1:52841:DataXceiver
error processing WRITE_BLOCK operation src: /127.0.0.1:52858 dst: /127.0.0.1:52841
java.lang.UnsatisfiedLinkError: org.apache.hadoop.util.NativeCrc32.nativeComputeChunkedSumsByteArray(II[BI[BIILjava/lang/String;JZ)V
at org.apache.hadoop.util.NativeCrc32.nativeComputeChunkedSumsByteArray(Native Method)
at org.apache.hadoop.util.NativeCrc32.verifyChunkedSumsByteArray(NativeCrc32.java:67)
at org.apache.hadoop.util.DataChecksum.verifyChunkedSums(DataChecksum.java:344)
at org.apache.hadoop.util.DataChecksum.verifyChunkedSums(DataChecksum.java:292)
at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.verifyChunks(BlockReceiver.java:416)
at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:551)
at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:771)
at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:718)
at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:126)
at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:72)
at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:225)
at java.lang.Thread.run(Thread.java:745)
2014-08-29 11:24:15,800 [ResponseProcessor for block BP-126326924-192.168.1.138-1409307851902:blk_1073741825_1001] WARN hdfs.DFSClie
nt (DFSOutputStream.java:run(880)) - DFSOutputStream ResponseProcessor exception for block BP-126326924-192.168.1.138-1409307851902:
blk_1073741825_1001
java.io.EOFException: Premature EOF: no length prefix available
at org.apache.hadoop.hdfs.protocolPB.PBHelper.vintPrefixed(PBHelper.java:2081)
at org.apache.hadoop.hdfs.protocol.datatransfer.PipelineAck.readFields(PipelineAck.java:176)
at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer$ResponseProcessor.run(DFSOutputStream.java:795)
2014-08-29 11:24:15,808 [Thread-63] INFO test.SliderTestUtils (SliderTestUtils.groovy:describe(71)) -
2014-08-29 11:24:15,808 [Thread-63] INFO test.SliderTestUtils (SliderTestUtils.groovy:describe(72)) - ==============================
=
2014-08-29 11:24:15,809 [Thread-63] INFO test.SliderTestUtils (SliderTestUtils.groovy:describe(73)) - teardown
2014-08-29 11:24:15,809 [Thread-63] INFO test.SliderTestUtils (SliderTestUtils.groovy:describe(74)) - ==============================
=
2014-08-29 11:24:15,809 [Thread-63] INFO test.SliderTestUtils (SliderTestUtils.groovy:describe(75)) -
2014-08-29 11:24:15,856 [JUnit] WARN datanode.DirectoryScanner (DirectoryScanner.java:shutdown(375)) - DirectoryScanner: shutdown ha
s been called
2014-08-29 11:24:15,940 [JUnit] INFO mortbay.log (Slf4jLog.java:info(67)) - Stopped HttpServer2$SelectChannelConnectorWithSafeStartu
p@127.0.0.1:0
2014-08-29 11:24:15,942 [JUnit] INFO ipc.Server (Server.java:stop(2398)) - Stopping server on 52845
2014-08-29 11:24:15,946 [IPC Server listener on 52845] INFO ipc.Server (Server.java:run(694)) - Stopping IPC Server listener on 5284
5
2014-08-29 11:24:15,947 [IPC Server Responder] INFO ipc.Server (Server.java:run(820)) - Stopping IPC Server Responder
2014-08-29 11:24:15,949 [DataNode: [[[DISK]file:/C:/Work/slider/slider-core/target/hdfs/TestConfPersister/data/data1/, [DISK]file:/C:
/Work/slider/slider-core/target/hdfs/TestConfPersister/data/data2/]] heartbeating to /127.0.0.1:52834] WARN datanode.DataNode (BPSe
rviceActor.java:offerService(734)) - BPOfferService for Block pool BP-126326924-192.168.1.138-1409307851902 (Datanode Uuid 1011a713-1
bc1-40d7-b652-f95dc4aa0b27) service to /127.0.0.1:52834 interrupted
2014-08-29 11:24:15,950 [DataNode: [[[DISK]file:/C:/Work/slider/slider-core/target/hdfs/TestConfPersister/data/data1/, [DISK]file:/C:
/Work/slider/slider-core/target/hdfs/TestConfPersister/data/data2/]] heartbeating to /127.0.0.1:52834] WARN datanode.DataNode (BPSe
rviceActor.java:run(857)) - Ending block pool service for: Block pool BP-126326924-192.168.1.138-1409307851902 (Datanode Uuid 1011a71
3-1bc1-40d7-b652-f95dc4aa0b27) service to /127.0.0.1:52834
2014-08-29 11:24:15,999 [JUnit] INFO ipc.Server (Server.java:stop(2398)) - Stopping server on 52834
2014-08-29 11:24:16,003 [IPC Server listener on 52834] INFO ipc.Server (Server.java:run(694)) - Stopping IPC Server listener on 5283
4
2014-08-29 11:24:16,004 [IPC Server Responder] INFO ipc.Server (Server.java:run(820)) - Stopping IPC Server Responder
2014-08-29 11:24:16,005 [org.apache.hadoop.hdfs.server.blockmanagement.DecommissionManager$Monitor@2247c9cd] WARN blockmanagement.De
commissionManager (DecommissionManager.java:run(78)) - Monitor interrupted: java.lang.InterruptedException: sleep interrupted
2014-08-29 11:24:16,032 [JUnit] INFO mortbay.log (Slf4jLog.java:info(67)) - Stopped HttpServer2$SelectChannelConnectorWithSafeStartu
p@127.0.0.1:0
Tests run: 1, Failures: 1, Errors: 0, Skipped: 0, Time elapsed: 7.801 sec <<< FAILURE! - in org.apache.slider.core.persist.TestConfPe
rsisterLocksHDFS
testAcqAcqRelReadlock(org.apache.slider.core.persist.TestConfPersisterLocksHDFS) Time elapsed: 0.545 sec <<< FAILURE!
org.codehaus.groovy.runtime.powerassert.PowerAssertionError: assert persister.acquireReadLock()
| |
| false
Persister to hdfs://localhost:52834/user/administrator/.slider/cluster/testAcqRelReadlock
at org.codehaus.groovy.runtime.InvokerHelper.assertFailed(InvokerHelper.java:398)
at org.codehaus.groovy.runtime.ScriptBytecodeAdapter.assertFailed(ScriptBytecodeAdapter.java:646)
at org.apache.slider.core.persist.TestConfPersisterLocksHDFS.testAcqAcqRelReadlock(TestConfPersisterLocksHDFS.groovy:135)
2014-08-29 11:24:16,144 [Thread-32] ERROR hdfs.DFSClient (DFSClient.java:closeAllFilesBeingWritten(888)) - Failed to close inode 1639
1
java.io.IOException: All datanodes 127.0.0.1:52841 are bad. Aborting...
at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.setupPipelineForAppendOrRecovery(DFSOutputStream.java:1132)
at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.processDatanodeError(DFSOutputStream.java:930)
at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.run(DFSOutputStream.java:481)
{code}
> slider MiniHDFSCluster tests failing on windows+branch2
> -------------------------------------------------------
>
> Key: SLIDER-377
> URL: https://issues.apache.org/jira/browse/SLIDER-377
> Project: Slider
> Issue Type: Sub-task
> Components: test, windows
> Affects Versions: Slider 0.60
> Reporter: Steve Loughran
>
> Tests that use the MiniHDFSCluster are failing on windows with link errors -datanodes are failing on JNI linkage errors calculating CRC32 checksums
--
This message was sent by Atlassian JIRA
(v6.2#6252)