You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@kafka.apache.org by Apache Jenkins Server <je...@builds.apache.org> on 2012/09/24 19:07:36 UTC

Build failed in Jenkins: Kafka-0.8 #49

See <https://builds.apache.org/job/Kafka-0.8/49/changes>

Changes:

[junrao] migration tool from 0.7 to 0.8; patched by Yang Ye; reviewed by Jun Rao; kafka-327

------------------------------------------
[...truncated 1596 lines...]
[info] == core-kafka / kafka.server.ReplicaFetchTest ==
[info] Test Starting: testReplicaFetcherThread(kafka.server.ReplicaFetchTest)
[info] Test Passed: testReplicaFetcherThread(kafka.server.ReplicaFetchTest)
[info] == core-kafka / kafka.server.ReplicaFetchTest ==
[info] 
[info] == core-kafka / kafka.server.LogRecoveryTest ==
[info] Test Starting: testHWCheckpointNoFailuresSingleLogSegment(kafka.server.LogRecoveryTest)
[info] Test Passed: testHWCheckpointNoFailuresSingleLogSegment(kafka.server.LogRecoveryTest)
[info] Test Starting: testHWCheckpointWithFailuresSingleLogSegment(kafka.server.LogRecoveryTest)
[error] Test Failed: testHWCheckpointWithFailuresSingleLogSegment(kafka.server.LogRecoveryTest)
java.lang.AssertionError: expected:<60> but was:<30>
	at org.junit.Assert.fail(Assert.java:69)
	at org.junit.Assert.failNotEquals(Assert.java:314)
	at org.junit.Assert.assertEquals(Assert.java:94)
	at org.junit.Assert.assertEquals(Assert.java:104)
	at kafka.server.LogRecoveryTest.testHWCheckpointWithFailuresSingleLogSegment(LogRecoveryTest.scala:150)
	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
	at java.lang.reflect.Method.invoke(Method.java:597)
	at junit.framework.TestCase.runTest(TestCase.java:164)
	at junit.framework.TestCase.runBare(TestCase.java:130)
	at junit.framework.TestResult$1.protect(TestResult.java:110)
	at junit.framework.TestResult.runProtected(TestResult.java:128)
	at junit.framework.TestResult.run(TestResult.java:113)
	at junit.framework.TestCase.run(TestCase.java:120)
	at junit.framework.TestSuite.runTest(TestSuite.java:228)
	at junit.framework.TestSuite.run(TestSuite.java:223)
	at junit.framework.TestSuite.runTest(TestSuite.java:228)
	at junit.framework.TestSuite.run(TestSuite.java:223)
	at org.scalatest.junit.JUnit3Suite.run(JUnit3Suite.scala:309)
	at org.scalatest.tools.ScalaTestFramework$ScalaTestRunner.run(ScalaTestFramework.scala:40)
	at sbt.TestRunner.run(TestFramework.scala:53)
	at sbt.TestRunner.runTest$1(TestFramework.scala:67)
	at sbt.TestRunner.run(TestFramework.scala:76)
	at sbt.TestFramework$$anonfun$10$$anonfun$apply$11.runTest$2(TestFramework.scala:194)
	at sbt.TestFramework$$anonfun$10$$anonfun$apply$11$$anonfun$apply$12.apply(TestFramework.scala:205)
	at sbt.TestFramework$$anonfun$10$$anonfun$apply$11$$anonfun$apply$12.apply(TestFramework.scala:205)
	at sbt.NamedTestTask.run(TestFramework.scala:92)
	at sbt.ScalaProject$$anonfun$sbt$ScalaProject$$toTask$1.apply(ScalaProject.scala:193)
	at sbt.ScalaProject$$anonfun$sbt$ScalaProject$$toTask$1.apply(ScalaProject.scala:193)
	at sbt.TaskManager$Task.invoke(TaskManager.scala:62)
	at sbt.impl.RunTask.doRun$1(RunTask.scala:77)
	at sbt.impl.RunTask.runTask(RunTask.scala:85)
	at sbt.impl.RunTask.sbt$impl$RunTask$$runIfNotRoot(RunTask.scala:60)
	at sbt.impl.RunTask$$anonfun$runTasksExceptRoot$2.apply(RunTask.scala:48)
	at sbt.impl.RunTask$$anonfun$runTasksExceptRoot$2.apply(RunTask.scala:48)
	at sbt.Distributor$Run$Worker$$anonfun$2.apply(ParallelRunner.scala:131)
	at sbt.Distributor$Run$Worker$$anonfun$2.apply(ParallelRunner.scala:131)
	at sbt.Control$.trapUnit(Control.scala:19)
	at sbt.Distributor$Run$Worker.run(ParallelRunner.scala:131)
[info] Test Starting: testHWCheckpointNoFailuresMultipleLogSegments(kafka.server.LogRecoveryTest)
[info] Test Passed: testHWCheckpointNoFailuresMultipleLogSegments(kafka.server.LogRecoveryTest)
[info] Test Starting: testHWCheckpointWithFailuresMultipleLogSegments(kafka.server.LogRecoveryTest)
[info] Test Passed: testHWCheckpointWithFailuresMultipleLogSegments(kafka.server.LogRecoveryTest)
[info] == core-kafka / kafka.server.LogRecoveryTest ==
[info] 
[info] == core-kafka / kafka.javaapi.message.ByteBufferMessageSetTest ==
[info] Test Starting: testWrittenEqualsRead
[info] Test Passed: testWrittenEqualsRead
[info] Test Starting: testIteratorIsConsistent
[info] Test Passed: testIteratorIsConsistent
[info] Test Starting: testSizeInBytes
[info] Test Passed: testSizeInBytes
[info] Test Starting: testEquals
[info] Test Passed: testEquals
[info] Test Starting: testIteratorIsConsistentWithCompression
[info] Test Passed: testIteratorIsConsistentWithCompression
[info] Test Starting: testSizeInBytesWithCompression
[info] Test Passed: testSizeInBytesWithCompression
[info] Test Starting: testEqualsWithCompression
[info] Test Passed: testEqualsWithCompression
[info] == core-kafka / kafka.javaapi.message.ByteBufferMessageSetTest ==
[info] 
[info] == core-kafka / kafka.network.SocketServerTest ==
[info] Test Starting: simpleRequest
[info] Test Passed: simpleRequest
[info] Test Starting: tooBigRequestIsRejected
[info] Test Passed: tooBigRequestIsRejected
[info] == core-kafka / kafka.network.SocketServerTest ==
[info] 
[info] == core-kafka / kafka.server.RequestPurgatoryTest ==
[info] Test Starting: testRequestSatisfaction(kafka.server.RequestPurgatoryTest)
[info] Test Passed: testRequestSatisfaction(kafka.server.RequestPurgatoryTest)
[info] Test Starting: testRequestExpiry(kafka.server.RequestPurgatoryTest)
[info] Test Passed: testRequestExpiry(kafka.server.RequestPurgatoryTest)
[info] == core-kafka / kafka.server.RequestPurgatoryTest ==
[info] 
[info] == core-kafka / kafka.zk.ZKEphemeralTest ==
[info] Test Starting: testEphemeralNodeCleanup(kafka.zk.ZKEphemeralTest)
[info] Test Passed: testEphemeralNodeCleanup(kafka.zk.ZKEphemeralTest)
[info] == core-kafka / kafka.zk.ZKEphemeralTest ==
[info] 
[info] == core-kafka / kafka.integration.LazyInitProducerTest ==
[info] Test Starting: testProduceAndMultiFetch(kafka.integration.LazyInitProducerTest)
[info] Test Passed: testProduceAndMultiFetch(kafka.integration.LazyInitProducerTest)
[info] Test Starting: testMultiProduce(kafka.integration.LazyInitProducerTest)
[info] Test Passed: testMultiProduce(kafka.integration.LazyInitProducerTest)
[info] Test Starting: testProduceAndFetch(kafka.integration.LazyInitProducerTest)
[info] Test Passed: testProduceAndFetch(kafka.integration.LazyInitProducerTest)
[info] Test Starting: testMultiProduceResend(kafka.integration.LazyInitProducerTest)
[info] Test Passed: testMultiProduceResend(kafka.integration.LazyInitProducerTest)
[info] == core-kafka / kafka.integration.LazyInitProducerTest ==
[info] 
[info] == core-kafka / kafka.producer.ProducerTest ==
[info] Test Starting: testUpdateBrokerPartitionInfo(kafka.producer.ProducerTest)
[info] Test Passed: testUpdateBrokerPartitionInfo(kafka.producer.ProducerTest)
[info] Test Starting: testSendToNewTopic(kafka.producer.ProducerTest)
[info] Test Passed: testSendToNewTopic(kafka.producer.ProducerTest)
[info] Test Starting: testSendWithDeadBroker(kafka.producer.ProducerTest)
[info] Test Passed: testSendWithDeadBroker(kafka.producer.ProducerTest)
[info] Test Starting: testAsyncSendCanCorrectlyFailWithTimeout(kafka.producer.ProducerTest)
[info] Test Passed: testAsyncSendCanCorrectlyFailWithTimeout(kafka.producer.ProducerTest)
[info] == core-kafka / kafka.producer.ProducerTest ==
[info] 
[info] == core-kafka / kafka.server.ServerShutdownTest ==
[info] Test Starting: testCleanShutdown(kafka.server.ServerShutdownTest)
[info] Test Passed: testCleanShutdown(kafka.server.ServerShutdownTest)
[info] == core-kafka / kafka.server.ServerShutdownTest ==
[info] 
[info] == core-kafka / kafka.consumer.FetcherTest ==
[info] Test Starting: testFetcher(kafka.consumer.FetcherTest)
[info] Test Passed: testFetcher(kafka.consumer.FetcherTest)
[info] == core-kafka / kafka.consumer.FetcherTest ==
[info] 
[info] == core-kafka / kafka.utils.UtilsTest ==
[info] Test Starting: testSwallow
[info] Test Passed: testSwallow
[info] Test Starting: testCircularIterator
[info] Test Passed: testCircularIterator
[info] == core-kafka / kafka.utils.UtilsTest ==
[info] 
[info] == core-kafka / kafka.message.MessageTest ==
[info] Test Starting: testFieldValues
[info] Test Passed: testFieldValues
[info] Test Starting: testChecksum
[info] Test Passed: testChecksum
[info] Test Starting: testEquality
[info] Test Passed: testEquality
[info] Test Starting: testIsHashable
[info] Test Passed: testIsHashable
[info] == core-kafka / kafka.message.MessageTest ==
[info] 
[info] == core-kafka / kafka.producer.AsyncProducerTest ==
[info] Test Starting: testProducerQueueSize(kafka.producer.AsyncProducerTest)
[info] Test Passed: testProducerQueueSize(kafka.producer.AsyncProducerTest)
[info] Test Starting: testProduceAfterClosed(kafka.producer.AsyncProducerTest)
[info] Test Passed: testProduceAfterClosed(kafka.producer.AsyncProducerTest)
[info] Test Starting: testBatchSize(kafka.producer.AsyncProducerTest)
[info] Test Passed: testBatchSize(kafka.producer.AsyncProducerTest)
[info] Test Starting: testQueueTimeExpired(kafka.producer.AsyncProducerTest)
[info] Test Passed: testQueueTimeExpired(kafka.producer.AsyncProducerTest)
[info] Test Starting: testPartitionAndCollateEvents(kafka.producer.AsyncProducerTest)
[info] Test Passed: testPartitionAndCollateEvents(kafka.producer.AsyncProducerTest)
[info] Test Starting: testSerializeEvents(kafka.producer.AsyncProducerTest)
[info] Test Passed: testSerializeEvents(kafka.producer.AsyncProducerTest)
[info] Test Starting: testInvalidPartition(kafka.producer.AsyncProducerTest)
[info] Test Passed: testInvalidPartition(kafka.producer.AsyncProducerTest)
[info] Test Starting: testNoBroker(kafka.producer.AsyncProducerTest)
[info] Test Passed: testNoBroker(kafka.producer.AsyncProducerTest)
[info] Test Starting: testIncompatibleEncoder(kafka.producer.AsyncProducerTest)
[info] Test Passed: testIncompatibleEncoder(kafka.producer.AsyncProducerTest)
[info] Test Starting: testRandomPartitioner(kafka.producer.AsyncProducerTest)
[info] Test Passed: testRandomPartitioner(kafka.producer.AsyncProducerTest)
[info] Test Starting: testBrokerListAndAsync(kafka.producer.AsyncProducerTest)
[info] Test Passed: testBrokerListAndAsync(kafka.producer.AsyncProducerTest)
[info] Test Starting: testFailedSendRetryLogic(kafka.producer.AsyncProducerTest)
[info] Test Passed: testFailedSendRetryLogic(kafka.producer.AsyncProducerTest)
[info] Test Starting: testJavaProducer(kafka.producer.AsyncProducerTest)
[info] Test Passed: testJavaProducer(kafka.producer.AsyncProducerTest)
[info] Test Starting: testInvalidConfiguration(kafka.producer.AsyncProducerTest)
[info] Test Passed: testInvalidConfiguration(kafka.producer.AsyncProducerTest)
[info] == core-kafka / kafka.producer.AsyncProducerTest ==
[info] 
[info] == core-kafka / kafka.consumer.ZookeeperConsumerConnectorTest ==
[info] Test Starting: testBasic(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] Test Passed: testBasic(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] Test Starting: testCompression(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] Test Passed: testCompression(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] Test Starting: testCompressionSetConsumption(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] Test Passed: testCompressionSetConsumption(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] Test Starting: testConsumerDecoder(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] Test Passed: testConsumerDecoder(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] Test Starting: testLeaderSelectionForPartition(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] Test Passed: testLeaderSelectionForPartition(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] == core-kafka / kafka.consumer.ZookeeperConsumerConnectorTest ==
[info] 
[info] == core-kafka / kafka.message.CompressionUtilTest ==
[info] Test Starting: testSimpleCompressDecompress
[info] Test Passed: testSimpleCompressDecompress
[info] Test Starting: testComplexCompressDecompress
[info] Test Passed: testComplexCompressDecompress
[info] Test Starting: testSnappyCompressDecompressExplicit
[info] Test Passed: testSnappyCompressDecompressExplicit
[info] == core-kafka / kafka.message.CompressionUtilTest ==
[info] 
[info] == core-kafka / kafka.javaapi.consumer.ZookeeperConsumerConnectorTest ==
[info] Test Starting: testBasic(kafka.javaapi.consumer.ZookeeperConsumerConnectorTest)
[info] Test Passed: testBasic(kafka.javaapi.consumer.ZookeeperConsumerConnectorTest)
[info] == core-kafka / kafka.javaapi.consumer.ZookeeperConsumerConnectorTest ==
[info] 
[info] == core-kafka / kafka.network.RpcDataSerializationTest ==
[info] Test Starting: testSerializationAndDeserialization
[info] Test Passed: testSerializationAndDeserialization
[info] == core-kafka / kafka.network.RpcDataSerializationTest ==
[info] 
[info] == core-kafka / unit.kafka.metrics.KafkaTimerTest ==
[info] Test Starting: testKafkaTimer(unit.kafka.metrics.KafkaTimerTest)
[info] Test Passed: testKafkaTimer(unit.kafka.metrics.KafkaTimerTest)
[info] == core-kafka / unit.kafka.metrics.KafkaTimerTest ==
[info] 
[info] == core-kafka / kafka.server.SimpleFetchTest ==
[info] Test Starting: testNonReplicaSeesHwWhenFetching(kafka.server.SimpleFetchTest)
[info] Test Passed: testNonReplicaSeesHwWhenFetching(kafka.server.SimpleFetchTest)
[info] Test Starting: testReplicaSeesLeoWhenFetching(kafka.server.SimpleFetchTest)
[info] Test Passed: testReplicaSeesLeoWhenFetching(kafka.server.SimpleFetchTest)
[info] == core-kafka / kafka.server.SimpleFetchTest ==
[info] 
[info] == core-kafka / kafka.log.LogManagerTest ==
[info] Test Starting: testCreateLog(kafka.log.LogManagerTest)
[info] Test Passed: testCreateLog(kafka.log.LogManagerTest)
[info] Test Starting: testGetLog(kafka.log.LogManagerTest)
[info] Test Passed: testGetLog(kafka.log.LogManagerTest)
[info] Test Starting: testCleanupExpiredSegments(kafka.log.LogManagerTest)
[info] Test Passed: testCleanupExpiredSegments(kafka.log.LogManagerTest)
[info] Test Starting: testCleanupSegmentsToMaintainSize(kafka.log.LogManagerTest)
[info] Test Passed: testCleanupSegmentsToMaintainSize(kafka.log.LogManagerTest)
[info] Test Starting: testTimeBasedFlush(kafka.log.LogManagerTest)
[info] Test Passed: testTimeBasedFlush(kafka.log.LogManagerTest)
[info] == core-kafka / kafka.log.LogManagerTest ==
[info] 
[info] == core-kafka / test-finish ==
[error] Failed: : Total 136, Failed 1, Errors 0, Passed 135, Skipped 0
[info] == core-kafka / test-finish ==
[info] 
[info] == core-kafka / Test cleanup 1 ==
[info] Deleting directory /tmp/sbt_28229fb9
[info] == core-kafka / Test cleanup 1 ==
[info] 
[info] == core-kafka / test-cleanup ==
[info] == core-kafka / test-cleanup ==
[error] Error running kafka.server.LogRecoveryTest: Test FAILED
[error] Error running test: One or more subtasks failed
[info] 
[info] Total time: 187 s, completed Sep 24, 2012 5:07:35 PM
[info] 
[info] Total session time: 187 s, completed Sep 24, 2012 5:07:35 PM
[error] Error during build.
Build step 'Execute shell' marked build as failure

Build failed in Jenkins: Kafka-0.8 #51

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/Kafka-0.8/51/changes>

Changes:

[junrao] restrict topic names (reopened); patched by Swapnil Ghike; reviewed by Jun Rao; KAFKA-495

[junrao] Use getMetadata Api in ZookeeperConsumerConnector; patched by Yang Ye; reviewed by Jun Rao; KAFKA-473

[joestein] KAFKA-533 changes to NOTICE and LICENSE related to KAFKA-534 removing client libraries from repo

[joestein] KAFKA-534 remove client library directory

------------------------------------------
[...truncated 5063 lines...]
	at junit.framework.TestSuite.run(TestSuite.java:223)
	at org.scalatest.junit.JUnit3Suite.run(JUnit3Suite.scala:309)
	at org.scalatest.tools.ScalaTestFramework$ScalaTestRunner.run(ScalaTestFramework.scala:40)
	at sbt.TestRunner.run(TestFramework.scala:53)
	at sbt.TestRunner.runTest$1(TestFramework.scala:67)
	at sbt.TestRunner.run(TestFramework.scala:76)
	at sbt.TestFramework$$anonfun$10$$anonfun$apply$11.runTest$2(TestFramework.scala:194)
	at sbt.TestFramework$$anonfun$10$$anonfun$apply$11$$anonfun$apply$12.apply(TestFramework.scala:205)
	at sbt.TestFramework$$anonfun$10$$anonfun$apply$11$$anonfun$apply$12.apply(TestFramework.scala:205)
	at sbt.NamedTestTask.run(TestFramework.scala:92)
	at sbt.ScalaProject$$anonfun$sbt$ScalaProject$$toTask$1.apply(ScalaProject.scala:193)
	at sbt.ScalaProject$$anonfun$sbt$ScalaProject$$toTask$1.apply(ScalaProject.scala:193)
	at sbt.TaskManager$Task.invoke(TaskManager.scala:62)
	at sbt.impl.RunTask.doRun$1(RunTask.scala:77)
	at sbt.impl.RunTask.runTask(RunTask.scala:85)
	at sbt.impl.RunTask.sbt$impl$RunTask$$runIfNotRoot(RunTask.scala:60)
	at sbt.impl.RunTask$$anonfun$runTasksExceptRoot$2.apply(RunTask.scala:48)
	at sbt.impl.RunTask$$anonfun$runTasksExceptRoot$2.apply(RunTask.scala:48)
	at sbt.Distributor$Run$Worker$$anonfun$2.apply(ParallelRunner.scala:131)
	at sbt.Distributor$Run$Worker$$anonfun$2.apply(ParallelRunner.scala:131)
	at sbt.Control$.trapUnit(Control.scala:19)
	at sbt.Distributor$Run$Worker.run(ParallelRunner.scala:131)
[2012-09-27 01:46:39,980] ERROR Connection attempt to localhost:44303 failed, next attempt in 1000 ms (kafka.producer.SyncProducer:102)
java.net.ConnectException: Connection refused
	at sun.nio.ch.Net.connect(Native Method)
	at sun.nio.ch.SocketChannelImpl.connect(SocketChannelImpl.java:507)
	at kafka.network.BlockingChannel.connect(BlockingChannel.scala:57)
	at kafka.producer.SyncProducer.connect(SyncProducer.scala:148)
	at kafka.producer.SyncProducer.getOrMakeConnection(SyncProducer.scala:169)
	at kafka.producer.SyncProducer.kafka$producer$SyncProducer$$doSend(SyncProducer.scala:74)
	at kafka.producer.SyncProducer.send(SyncProducer.scala:112)
	at kafka.utils.Utils$.getTopicMetadata(Utils.scala:689)
	at kafka.producer.BrokerPartitionInfo.updateInfo(BrokerPartitionInfo.scala:74)
	at kafka.producer.async.DefaultEventHandler$$anonfun$handle$1.apply$mcV$sp(DefaultEventHandler.scala:58)
	at kafka.utils.Utils$.swallow(Utils.scala:251)
	at kafka.utils.Logging$class.swallowError(Logging.scala:105)
	at kafka.utils.Utils$.swallowError(Utils.scala:41)
	at kafka.producer.async.DefaultEventHandler.handle(DefaultEventHandler.scala:58)
	at kafka.producer.Producer.send(Producer.scala:66)
	at kafka.producer.ProducerTest.testSendWithDeadBroker(ProducerTest.scala:237)
	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
	at java.lang.reflect.Method.invoke(Method.java:597)
	at junit.framework.TestCase.runTest(TestCase.java:164)
	at junit.framework.TestCase.runBare(TestCase.java:130)
	at junit.framework.TestResult$1.protect(TestResult.java:110)
	at junit.framework.TestResult.runProtected(TestResult.java:128)
	at junit.framework.TestResult.run(TestResult.java:113)
	at junit.framework.TestCase.run(TestCase.java:120)
	at junit.framework.TestSuite.runTest(TestSuite.java:228)
	at junit.framework.TestSuite.run(TestSuite.java:223)
	at junit.framework.TestSuite.runTest(TestSuite.java:228)
	at junit.framework.TestSuite.run(TestSuite.java:223)
	at org.scalatest.junit.JUnit3Suite.run(JUnit3Suite.scala:309)
	at org.scalatest.tools.ScalaTestFramework$ScalaTestRunner.run(ScalaTestFramework.scala:40)
	at sbt.TestRunner.run(TestFramework.scala:53)
	at sbt.TestRunner.runTest$1(TestFramework.scala:67)
	at sbt.TestRunner.run(TestFramework.scala:76)
	at sbt.TestFramework$$anonfun$10$$anonfun$apply$11.runTest$2(TestFramework.scala:194)
	at sbt.TestFramework$$anonfun$10$$anonfun$apply$11$$anonfun$apply$12.apply(TestFramework.scala:205)
	at sbt.TestFramework$$anonfun$10$$anonfun$apply$11$$anonfun$apply$12.apply(TestFramework.scala:205)
	at sbt.NamedTestTask.run(TestFramework.scala:92)
	at sbt.ScalaProject$$anonfun$sbt$ScalaProject$$toTask$1.apply(ScalaProject.scala:193)
	at sbt.ScalaProject$$anonfun$sbt$ScalaProject$$toTask$1.apply(ScalaProject.scala:193)
	at sbt.TaskManager$Task.invoke(TaskManager.scala:62)
	at sbt.impl.RunTask.doRun$1(RunTask.scala:77)
	at sbt.impl.RunTask.runTask(RunTask.scala:85)
	at sbt.impl.RunTask.sbt$impl$RunTask$$runIfNotRoot(RunTask.scala:60)
	at sbt.impl.RunTask$$anonfun$runTasksExceptRoot$2.apply(RunTask.scala:48)
	at sbt.impl.RunTask$$anonfun$runTasksExceptRoot$2.apply(RunTask.scala:48)
	at sbt.Distributor$Run$Worker$$anonfun$2.apply(ParallelRunner.scala:131)
	at sbt.Distributor$Run$Worker$$anonfun$2.apply(ParallelRunner.scala:131)
	at sbt.Control$.trapUnit(Control.scala:19)
	at sbt.Distributor$Run$Worker.run(ParallelRunner.scala:131)
[2012-09-27 01:46:39,980] WARN EndOfStreamException: Unable to read additional data from client sessionid 0x13a0566de200002, likely client has closed socket (org.apache.zookeeper.server.NIOServerCnxn:634)
[2012-09-27 01:46:40,990] ERROR Producer connection to localhost:44303 timing out after 5000 ms (kafka.producer.SyncProducer:102)
java.net.ConnectException: Connection refused
	at sun.nio.ch.Net.connect(Native Method)
	at sun.nio.ch.SocketChannelImpl.connect(SocketChannelImpl.java:507)
	at kafka.network.BlockingChannel.connect(BlockingChannel.scala:57)
	at kafka.producer.SyncProducer.connect(SyncProducer.scala:148)
	at kafka.producer.SyncProducer.getOrMakeConnection(SyncProducer.scala:169)
	at kafka.producer.SyncProducer.kafka$producer$SyncProducer$$doSend(SyncProducer.scala:74)
	at kafka.producer.SyncProducer.send(SyncProducer.scala:112)
	at kafka.utils.Utils$.getTopicMetadata(Utils.scala:689)
	at kafka.producer.BrokerPartitionInfo.updateInfo(BrokerPartitionInfo.scala:74)
	at kafka.producer.async.DefaultEventHandler$$anonfun$handle$1.apply$mcV$sp(DefaultEventHandler.scala:58)
	at kafka.utils.Utils$.swallow(Utils.scala:251)
	at kafka.utils.Logging$class.swallowError(Logging.scala:105)
	at kafka.utils.Utils$.swallowError(Utils.scala:41)
	at kafka.producer.async.DefaultEventHandler.handle(DefaultEventHandler.scala:58)
	at kafka.producer.Producer.send(Producer.scala:66)
	at kafka.producer.ProducerTest.testSendWithDeadBroker(ProducerTest.scala:237)
	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
	at java.lang.reflect.Method.invoke(Method.java:597)
	at junit.framework.TestCase.runTest(TestCase.java:164)
	at junit.framework.TestCase.runBare(TestCase.java:130)
	at junit.framework.TestResult$1.protect(TestResult.java:110)
	at junit.framework.TestResult.runProtected(TestResult.java:128)
	at junit.framework.TestResult.run(TestResult.java:113)
	at junit.framework.TestCase.run(TestCase.java:120)
	at junit.framework.TestSuite.runTest(TestSuite.java:228)
	at junit.framework.TestSuite.run(TestSuite.java:223)
	at junit.framework.TestSuite.runTest(TestSuite.java:228)
	at junit.framework.TestSuite.run(TestSuite.java:223)
	at org.scalatest.junit.JUnit3Suite.run(JUnit3Suite.scala:309)
	at org.scalatest.tools.ScalaTestFramework$ScalaTestRunner.run(ScalaTestFramework.scala:40)
	at sbt.TestRunner.run(TestFramework.scala:53)
	at sbt.TestRunner.runTest$1(TestFramework.scala:67)
	at sbt.TestRunner.run(TestFramework.scala:76)
	at sbt.TestFramework$$anonfun$10$$anonfun$apply$11.runTest$2(TestFramework.scala:194)
	at sbt.TestFramework$$anonfun$10$$anonfun$apply$11$$anonfun$apply$12.apply(TestFramework.scala:205)
	at sbt.TestFramework$$anonfun$10$$anonfun$apply$11$$anonfun$apply$12.apply(TestFramework.scala:205)
	at sbt.NamedTestTask.run(TestFramework.scala:92)
	at sbt.ScalaProject$$anonfun$sbt$ScalaProject$$toTask$1.apply(ScalaProject.scala:193)
	at sbt.ScalaProject$$anonfun$sbt$ScalaProject$$toTask$1.apply(ScalaProject.scala:193)
	at sbt.TaskManager$Task.invoke(TaskManager.scala:62)
	at sbt.impl.RunTask.doRun$1(RunTask.scala:77)
	at sbt.impl.RunTask.runTask(RunTask.scala:85)
	at sbt.impl.RunTask.sbt$impl$RunTask$$runIfNotRoot(RunTask.scala:60)
	at sbt.impl.RunTask$$anonfun$runTasksExceptRoot$2.apply(RunTask.scala:48)
	at sbt.impl.RunTask$$anonfun$runTasksExceptRoot$2.apply(RunTask.scala:48)
	at sbt.Distributor$Run$Worker$$anonfun$2.apply(ParallelRunner.scala:131)
	at sbt.Distributor$Run$Worker$$anonfun$2.apply(ParallelRunner.scala:131)
	at sbt.Control$.trapUnit(Control.scala:19)
	at sbt.Distributor$Run$Worker.run(ParallelRunner.scala:131)
[2012-09-27 01:46:41,003] ERROR Failed to send the following requests: ListBuffer(ProducerData(new-topic,test,ListBuffer(message(magic = 1, attributes = 0, crc = 2326977762, payload = java.nio.HeapByteBuffer[pos=0 lim=5 cap=5])))) (kafka.producer.async.DefaultEventHandler:96)
[info] Test Passed: testSendWithDeadBroker(kafka.producer.ProducerTest)
[info] Test Starting: testAsyncSendCanCorrectlyFailWithTimeout(kafka.producer.ProducerTest)
[2012-09-27 01:46:41,810] WARN EndOfStreamException: Unable to read additional data from client sessionid 0x13a0566de200002, likely client has closed socket (org.apache.zookeeper.server.NIOServerCnxn:634)
[2012-09-27 01:46:42,211] WARN EndOfStreamException: Unable to read additional data from client sessionid 0x13a056731ef0004, likely client has closed socket (org.apache.zookeeper.server.NIOServerCnxn:634)
[2012-09-27 01:46:43,500] WARN EndOfStreamException: Unable to read additional data from client sessionid 0x13a0566de200002, likely client has closed socket (org.apache.zookeeper.server.NIOServerCnxn:634)
[2012-09-27 01:46:43,660] WARN EndOfStreamException: Unable to read additional data from client sessionid 0x13a056731ef0004, likely client has closed socket (org.apache.zookeeper.server.NIOServerCnxn:634)
[2012-09-27 01:46:45,211] WARN EndOfStreamException: Unable to read additional data from client sessionid 0x13a0566de200002, likely client has closed socket (org.apache.zookeeper.server.NIOServerCnxn:634)
[2012-09-27 01:46:45,670] WARN EndOfStreamException: Unable to read additional data from client sessionid 0x13a056731ef0004, likely client has closed socket (org.apache.zookeeper.server.NIOServerCnxn:634)
[2012-09-27 01:46:46,670] ERROR Failed to send the following requests: ListBuffer(ProducerData(new-topic,test,ListBuffer(message(magic = 1, attributes = 0, crc = 3632233996, payload = java.nio.HeapByteBuffer[pos=0 lim=4 cap=4])))) (kafka.producer.async.DefaultEventHandler:96)
[2012-09-27 01:46:46,890] WARN EndOfStreamException: Unable to read additional data from client sessionid 0x13a0566de200002, likely client has closed socket (org.apache.zookeeper.server.NIOServerCnxn:634)
[info] Test Passed: testAsyncSendCanCorrectlyFailWithTimeout(kafka.producer.ProducerTest)
[info] == core-kafka / kafka.producer.ProducerTest ==
[info] 
[info] == core-kafka / kafka.server.RequestPurgatoryTest ==
[info] Test Starting: testRequestSatisfaction(kafka.server.RequestPurgatoryTest)
[info] Test Passed: testRequestSatisfaction(kafka.server.RequestPurgatoryTest)
[info] Test Starting: testRequestExpiry(kafka.server.RequestPurgatoryTest)
[info] Test Passed: testRequestExpiry(kafka.server.RequestPurgatoryTest)
[info] == core-kafka / kafka.server.RequestPurgatoryTest ==
[info] 
[info] == core-kafka / kafka.message.MessageTest ==
[info] Test Starting: testFieldValues
[2012-09-27 01:46:47,550] ERROR Unable to open socket to localhost/127.0.0.1:2182 (org.apache.zookeeper.ClientCnxn:1076)
[2012-09-27 01:46:47,550] WARN Session 0x13a056731ef0004 for server null, unexpected error, closing socket connection and attempting reconnect (org.apache.zookeeper.ClientCnxn:1188)
java.net.ConnectException: Connection refused
	at sun.nio.ch.Net.connect(Native Method)
	at sun.nio.ch.SocketChannelImpl.connect(SocketChannelImpl.java:507)
	at org.apache.zookeeper.ClientCnxn$SendThread.startConnect(ClientCnxn.java:1071)
	at org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1104)
[info] Test Passed: testFieldValues
[info] Test Starting: testChecksum
[info] Test Passed: testChecksum
[info] Test Starting: testEquality
[info] Test Passed: testEquality
[info] Test Starting: testIsHashable
[info] Test Passed: testIsHashable
[info] == core-kafka / kafka.message.MessageTest ==
[info] 
[info] == core-kafka / kafka.log4j.KafkaLog4jAppenderTest ==
[info] Test Starting: testKafkaLog4jConfigs(kafka.log4j.KafkaLog4jAppenderTest)
log4j:WARN No appenders could be found for logger (org.I0Itec.zkclient.ZkEventThread).
log4j:WARN Please initialize the log4j system properly.
[info] Test Passed: testKafkaLog4jConfigs(kafka.log4j.KafkaLog4jAppenderTest)
[info] Test Starting: testLog4jAppends(kafka.log4j.KafkaLog4jAppenderTest)
[info] Test Passed: testLog4jAppends(kafka.log4j.KafkaLog4jAppenderTest)
[info] == core-kafka / kafka.log4j.KafkaLog4jAppenderTest ==
[info] 
[info] == core-kafka / kafka.network.SocketServerTest ==
[info] Test Starting: simpleRequest
[info] Test Passed: simpleRequest
[info] Test Starting: tooBigRequestIsRejected
[info] Test Passed: tooBigRequestIsRejected
[info] == core-kafka / kafka.network.SocketServerTest ==
[info] 
[info] == core-kafka / kafka.integration.TopicMetadataTest ==
[info] Test Starting: testTopicMetadataRequest(kafka.integration.TopicMetadataTest)
[info] Test Passed: testTopicMetadataRequest(kafka.integration.TopicMetadataTest)
[info] Test Starting: testBasicTopicMetadata(kafka.integration.TopicMetadataTest)
[info] Test Passed: testBasicTopicMetadata(kafka.integration.TopicMetadataTest)
[info] Test Starting: testAutoCreateTopic(kafka.integration.TopicMetadataTest)
[info] Test Passed: testAutoCreateTopic(kafka.integration.TopicMetadataTest)
[info] == core-kafka / kafka.integration.TopicMetadataTest ==
[info] 
[info] == core-kafka / kafka.admin.AdminTest ==
[info] Test Starting: testReplicaAssignment(kafka.admin.AdminTest)
[info] Test Passed: testReplicaAssignment(kafka.admin.AdminTest)
[info] Test Starting: testManualReplicaAssignment(kafka.admin.AdminTest)
[info] Test Passed: testManualReplicaAssignment(kafka.admin.AdminTest)
[info] Test Starting: testTopicCreationInZK(kafka.admin.AdminTest)
[info] Test Passed: testTopicCreationInZK(kafka.admin.AdminTest)
[info] Test Starting: testGetTopicMetadata(kafka.admin.AdminTest)
[info] Test Passed: testGetTopicMetadata(kafka.admin.AdminTest)
[info] == core-kafka / kafka.admin.AdminTest ==
[info] 
[info] == core-kafka / kafka.integration.LazyInitProducerTest ==
[info] Test Starting: testProduceAndMultiFetch(kafka.integration.LazyInitProducerTest)
[info] Test Passed: testProduceAndMultiFetch(kafka.integration.LazyInitProducerTest)
[info] Test Starting: testMultiProduce(kafka.integration.LazyInitProducerTest)
[info] Test Passed: testMultiProduce(kafka.integration.LazyInitProducerTest)
[info] Test Starting: testProduceAndFetch(kafka.integration.LazyInitProducerTest)
[info] Test Passed: testProduceAndFetch(kafka.integration.LazyInitProducerTest)
[info] Test Starting: testMultiProduceResend(kafka.integration.LazyInitProducerTest)
[info] Test Passed: testMultiProduceResend(kafka.integration.LazyInitProducerTest)
[info] == core-kafka / kafka.integration.LazyInitProducerTest ==
[info] 
[info] == core-kafka / kafka.integration.AutoOffsetResetTest ==
[info] Test Starting: testResetToEarliestWhenOffsetTooHigh(kafka.integration.AutoOffsetResetTest)
[info] Test Passed: testResetToEarliestWhenOffsetTooHigh(kafka.integration.AutoOffsetResetTest)
[info] Test Starting: testResetToEarliestWhenOffsetTooLow(kafka.integration.AutoOffsetResetTest)
[info] Test Passed: testResetToEarliestWhenOffsetTooLow(kafka.integration.AutoOffsetResetTest)
[info] Test Starting: testResetToLatestWhenOffsetTooHigh(kafka.integration.AutoOffsetResetTest)
[info] Test Passed: testResetToLatestWhenOffsetTooHigh(kafka.integration.AutoOffsetResetTest)
[info] Test Starting: testResetToLatestWhenOffsetTooLow(kafka.integration.AutoOffsetResetTest)
[info] Test Passed: testResetToLatestWhenOffsetTooLow(kafka.integration.AutoOffsetResetTest)
[info] == core-kafka / kafka.integration.AutoOffsetResetTest ==
[info] 
[info] == core-kafka / kafka.consumer.FetcherTest ==
[info] Test Starting: testFetcher(kafka.consumer.FetcherTest)
[info] Test Passed: testFetcher(kafka.consumer.FetcherTest)
[info] == core-kafka / kafka.consumer.FetcherTest ==
[info] 
[info] == core-kafka / Test cleanup 1 ==
[info] Deleting directory /var/tmp/sbt_ab10d986
[info] == core-kafka / Test cleanup 1 ==
[info] 
[info] == core-kafka / test-finish ==
[error] Failed: : Total 136, Failed 1, Errors 0, Passed 135, Skipped 0
[info] == core-kafka / test-finish ==
[info] 
[info] == core-kafka / test-cleanup ==
[info] == core-kafka / test-cleanup ==
[error] Error running kafka.message.CompressionUtilTest: Test FAILED
[error] Error running test: One or more subtasks failed
[info] 
[info] Total time: 269 s, completed Sep 27, 2012 1:47:01 AM
[info] 
[info] Total session time: 270 s, completed Sep 27, 2012 1:47:01 AM
[error] Error during build.
Build step 'Execute shell' marked build as failure

Build failed in Jenkins: Kafka-0.8 #50

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/Kafka-0.8/50/changes>

Changes:

[junrao] System Test should remove the top level data log directory; patched by John Fung; reviewed by Jun Rao; KAFKA-526

------------------------------------------
[...truncated 1253 lines...]
	at sbt.impl.RunTask$$anonfun$runTasksExceptRoot$2.apply(RunTask.scala:48)
	at sbt.Distributor$Run$Worker$$anonfun$2.apply(ParallelRunner.scala:131)
	at sbt.Distributor$Run$Worker$$anonfun$2.apply(ParallelRunner.scala:131)
	at sbt.Control$.trapUnit(Control.scala:19)
	at sbt.Distributor$Run$Worker.run(ParallelRunner.scala:131)
Caused by: java.lang.UnsatisfiedLinkError: no snappyjava in java.library.path
	at java.lang.ClassLoader.loadLibrary(ClassLoader.java:1734)
	at java.lang.Runtime.loadLibrary0(Runtime.java:823)
	at java.lang.System.loadLibrary(System.java:1028)
	at org.xerial.snappy.SnappyNativeLoader.loadLibrary(SnappyNativeLoader.java:52)
	... 53 more
[error] Test Failed: testSnappyCompressDecompressExplicit
org.xerial.snappy.SnappyError: [FAILED_TO_LOAD_NATIVE_LIBRARY] null
	at org.xerial.snappy.SnappyLoader.load(SnappyLoader.java:229)
	at org.xerial.snappy.Snappy.<clinit>(Snappy.java:44)
	at org.xerial.snappy.SnappyOutputStream.<init>(SnappyOutputStream.java:79)
	at org.xerial.snappy.SnappyOutputStream.<init>(SnappyOutputStream.java:66)
	at kafka.message.SnappyCompression.<init>(CompressionUtils.scala:61)
	at kafka.message.CompressionFactory$.apply(CompressionUtils.scala:82)
	at kafka.message.CompressionUtils$.compress(CompressionUtils.scala:109)
	at kafka.message.CompressionUtilTest.testSnappyCompressDecompressExplicit(CompressionUtilsTest.scala:65)
	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
	at java.lang.reflect.Method.invoke(Method.java:597)
	at org.junit.internal.runners.TestMethodRunner.executeMethodBody(TestMethodRunner.java:99)
	at org.junit.internal.runners.TestMethodRunner.runUnprotected(TestMethodRunner.java:81)
	at org.junit.internal.runners.BeforeAndAfterRunner.runProtected(BeforeAndAfterRunner.java:34)
	at org.junit.internal.runners.TestMethodRunner.runMethod(TestMethodRunner.java:75)
	at org.junit.internal.runners.TestMethodRunner.run(TestMethodRunner.java:45)
	at org.junit.internal.runners.TestClassMethodsRunner.invokeTestMethod(TestClassMethodsRunner.java:71)
	at org.junit.internal.runners.TestClassMethodsRunner.run(TestClassMethodsRunner.java:35)
	at org.junit.internal.runners.TestClassRunner$1.runUnprotected(TestClassRunner.java:42)
	at org.junit.internal.runners.BeforeAndAfterRunner.runProtected(BeforeAndAfterRunner.java:34)
	at org.junit.internal.runners.TestClassRunner.run(TestClassRunner.java:52)
	at org.junit.internal.runners.CompositeRunner.run(CompositeRunner.java:29)
	at org.junit.runner.JUnitCore.run(JUnitCore.java:121)
	at org.junit.runner.JUnitCore.run(JUnitCore.java:100)
	at org.junit.runner.JUnitCore.run(JUnitCore.java:91)
	at org.scalatest.junit.JUnitSuite$class.run(JUnitSuite.scala:261)
	at kafka.message.CompressionUtilTest.run(CompressionUtilsTest.scala:25)
	at org.scalatest.tools.ScalaTestFramework$ScalaTestRunner.run(ScalaTestFramework.scala:40)
	at sbt.TestRunner.run(TestFramework.scala:53)
	at sbt.TestRunner.runTest$1(TestFramework.scala:67)
	at sbt.TestRunner.run(TestFramework.scala:76)
	at sbt.TestFramework$$anonfun$10$$anonfun$apply$11.runTest$2(TestFramework.scala:194)
	at sbt.TestFramework$$anonfun$10$$anonfun$apply$11$$anonfun$apply$12.apply(TestFramework.scala:205)
	at sbt.TestFramework$$anonfun$10$$anonfun$apply$11$$anonfun$apply$12.apply(TestFramework.scala:205)
	at sbt.NamedTestTask.run(TestFramework.scala:92)
	at sbt.ScalaProject$$anonfun$sbt$ScalaProject$$toTask$1.apply(ScalaProject.scala:193)
	at sbt.ScalaProject$$anonfun$sbt$ScalaProject$$toTask$1.apply(ScalaProject.scala:193)
	at sbt.TaskManager$Task.invoke(TaskManager.scala:62)
	at sbt.impl.RunTask.doRun$1(RunTask.scala:77)
	at sbt.impl.RunTask.runTask(RunTask.scala:85)
	at sbt.impl.RunTask.sbt$impl$RunTask$$runIfNotRoot(RunTask.scala:60)
	at sbt.impl.RunTask$$anonfun$runTasksExceptRoot$2.apply(RunTask.scala:48)
	at sbt.impl.RunTask$$anonfun$runTasksExceptRoot$2.apply(RunTask.scala:48)
	at sbt.Distributor$Run$Worker$$anonfun$2.apply(ParallelRunner.scala:131)
	at sbt.Distributor$Run$Worker$$anonfun$2.apply(ParallelRunner.scala:131)
	at sbt.Control$.trapUnit(Control.scala:19)
	at sbt.Distributor$Run$Worker.run(ParallelRunner.scala:131)
[info] == core-kafka / kafka.message.CompressionUtilTest ==
[info] 
[info] == core-kafka / kafka.consumer.ZookeeperConsumerConnectorTest ==
[info] Test Starting: testBasic(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] Test Passed: testBasic(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] Test Starting: testCompression(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] Test Passed: testCompression(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] Test Starting: testCompressionSetConsumption(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] Test Passed: testCompressionSetConsumption(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] Test Starting: testConsumerDecoder(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] Test Passed: testConsumerDecoder(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] Test Starting: testLeaderSelectionForPartition(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] Test Passed: testLeaderSelectionForPartition(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] == core-kafka / kafka.consumer.ZookeeperConsumerConnectorTest ==
[info] 
[info] == core-kafka / kafka.integration.AutoOffsetResetTest ==
[info] Test Starting: testResetToEarliestWhenOffsetTooHigh(kafka.integration.AutoOffsetResetTest)
[info] Test Passed: testResetToEarliestWhenOffsetTooHigh(kafka.integration.AutoOffsetResetTest)
[info] Test Starting: testResetToEarliestWhenOffsetTooLow(kafka.integration.AutoOffsetResetTest)
[info] Test Passed: testResetToEarliestWhenOffsetTooLow(kafka.integration.AutoOffsetResetTest)
[info] Test Starting: testResetToLatestWhenOffsetTooHigh(kafka.integration.AutoOffsetResetTest)
[info] Test Passed: testResetToLatestWhenOffsetTooHigh(kafka.integration.AutoOffsetResetTest)
[info] Test Starting: testResetToLatestWhenOffsetTooLow(kafka.integration.AutoOffsetResetTest)
[info] Test Passed: testResetToLatestWhenOffsetTooLow(kafka.integration.AutoOffsetResetTest)
[info] == core-kafka / kafka.integration.AutoOffsetResetTest ==
[info] 
[info] == core-kafka / kafka.zk.ZKEphemeralTest ==
[info] Test Starting: testEphemeralNodeCleanup(kafka.zk.ZKEphemeralTest)
[info] Test Passed: testEphemeralNodeCleanup(kafka.zk.ZKEphemeralTest)
[info] == core-kafka / kafka.zk.ZKEphemeralTest ==
[info] 
[info] == core-kafka / kafka.producer.SyncProducerTest ==
[info] Test Starting: testMessageSizeTooLarge(kafka.producer.SyncProducerTest)
[info] Test Passed: testMessageSizeTooLarge(kafka.producer.SyncProducerTest)
[info] Test Starting: testReachableServer(kafka.producer.SyncProducerTest)
[info] Test Passed: testReachableServer(kafka.producer.SyncProducerTest)
[info] Test Starting: testEmptyProduceRequest(kafka.producer.SyncProducerTest)
[info] Test Passed: testEmptyProduceRequest(kafka.producer.SyncProducerTest)
[info] Test Starting: testProduceCorrectlyReceivesResponse(kafka.producer.SyncProducerTest)
[info] Test Passed: testProduceCorrectlyReceivesResponse(kafka.producer.SyncProducerTest)
[info] Test Starting: testProducerCanTimeout(kafka.producer.SyncProducerTest)
[info] Test Passed: testProducerCanTimeout(kafka.producer.SyncProducerTest)
[info] == core-kafka / kafka.producer.SyncProducerTest ==
[info] 
[info] == core-kafka / kafka.producer.ProducerTest ==
[info] Test Starting: testUpdateBrokerPartitionInfo(kafka.producer.ProducerTest)
[info] Test Passed: testUpdateBrokerPartitionInfo(kafka.producer.ProducerTest)
[info] Test Starting: testSendToNewTopic(kafka.producer.ProducerTest)
[info] Test Passed: testSendToNewTopic(kafka.producer.ProducerTest)
[info] Test Starting: testSendWithDeadBroker(kafka.producer.ProducerTest)
[info] Test Passed: testSendWithDeadBroker(kafka.producer.ProducerTest)
[info] Test Starting: testAsyncSendCanCorrectlyFailWithTimeout(kafka.producer.ProducerTest)
[info] Test Passed: testAsyncSendCanCorrectlyFailWithTimeout(kafka.producer.ProducerTest)
[info] == core-kafka / kafka.producer.ProducerTest ==
[info] 
[info] == core-kafka / kafka.network.RpcDataSerializationTest ==
[info] Test Starting: testSerializationAndDeserialization
[info] Test Passed: testSerializationAndDeserialization
[info] == core-kafka / kafka.network.RpcDataSerializationTest ==
[info] 
[info] == core-kafka / kafka.javaapi.consumer.ZookeeperConsumerConnectorTest ==
[info] Test Starting: testBasic(kafka.javaapi.consumer.ZookeeperConsumerConnectorTest)
[info] Test Passed: testBasic(kafka.javaapi.consumer.ZookeeperConsumerConnectorTest)
[info] == core-kafka / kafka.javaapi.consumer.ZookeeperConsumerConnectorTest ==
[info] 
[info] == core-kafka / kafka.integration.PrimitiveApiTest ==
[info] Test Starting: testProduceAndMultiFetch(kafka.integration.PrimitiveApiTest)
[info] Test Passed: testProduceAndMultiFetch(kafka.integration.PrimitiveApiTest)
[info] Test Starting: testMultiProduce(kafka.integration.PrimitiveApiTest)
[info] Test Passed: testMultiProduce(kafka.integration.PrimitiveApiTest)
[info] Test Starting: testFetchRequestCanProperlySerialize(kafka.integration.PrimitiveApiTest)
[info] Test Passed: testFetchRequestCanProperlySerialize(kafka.integration.PrimitiveApiTest)
[info] Test Starting: testEmptyFetchRequest(kafka.integration.PrimitiveApiTest)
[info] Test Passed: testEmptyFetchRequest(kafka.integration.PrimitiveApiTest)
[info] Test Starting: testDefaultEncoderProducerAndFetch(kafka.integration.PrimitiveApiTest)
[info] Test Passed: testDefaultEncoderProducerAndFetch(kafka.integration.PrimitiveApiTest)
[info] Test Starting: testDefaultEncoderProducerAndFetchWithCompression(kafka.integration.PrimitiveApiTest)
[info] Test Passed: testDefaultEncoderProducerAndFetchWithCompression(kafka.integration.PrimitiveApiTest)
[info] Test Starting: testProduceAndMultiFetchWithCompression(kafka.integration.PrimitiveApiTest)
[info] Test Passed: testProduceAndMultiFetchWithCompression(kafka.integration.PrimitiveApiTest)
[info] Test Starting: testMultiProduceWithCompression(kafka.integration.PrimitiveApiTest)
[info] Test Passed: testMultiProduceWithCompression(kafka.integration.PrimitiveApiTest)
[info] Test Starting: testConsumerEmptyTopic(kafka.integration.PrimitiveApiTest)
[info] Test Passed: testConsumerEmptyTopic(kafka.integration.PrimitiveApiTest)
[info] == core-kafka / kafka.integration.PrimitiveApiTest ==
[info] 
[info] == core-kafka / kafka.consumer.FetcherTest ==
[info] Test Starting: testFetcher(kafka.consumer.FetcherTest)
[info] Test Passed: testFetcher(kafka.consumer.FetcherTest)
[info] == core-kafka / kafka.consumer.FetcherTest ==
[info] 
[info] == core-kafka / unit.kafka.metrics.KafkaTimerTest ==
[info] Test Starting: testKafkaTimer(unit.kafka.metrics.KafkaTimerTest)
[info] Test Passed: testKafkaTimer(unit.kafka.metrics.KafkaTimerTest)
[info] == core-kafka / unit.kafka.metrics.KafkaTimerTest ==
[info] 
[info] == core-kafka / kafka.log.SegmentListTest ==
[info] Test Starting: testAppend
[info] Test Passed: testAppend
[info] Test Starting: testTrunc
[info] Test Passed: testTrunc
[info] Test Starting: testTruncLast
[info] Test Passed: testTruncLast
[info] == core-kafka / kafka.log.SegmentListTest ==
[info] 
[info] == core-kafka / kafka.log.LogTest ==
[info] Test Starting: testTimeBasedLogRoll
[info] Test Passed: testTimeBasedLogRoll
[info] Test Starting: testSizeBasedLogRoll
[info] Test Passed: testSizeBasedLogRoll
[info] Test Starting: testLoadEmptyLog
[info] Test Passed: testLoadEmptyLog
[info] Test Starting: testLoadInvalidLogsFails
[info] Test Passed: testLoadInvalidLogsFails
[info] Test Starting: testAppendAndRead
[info] Test Passed: testAppendAndRead
[info] Test Starting: testReadOutOfRange
[info] Test Passed: testReadOutOfRange
[info] Test Starting: testLogRolls
[info] Test Passed: testLogRolls
[info] Test Starting: testFindSegment
[info] Test Passed: testFindSegment
[info] Test Starting: testEdgeLogRolls
[info] Test Passed: testEdgeLogRolls
[info] Test Starting: testMessageSizeCheck
[info] Test Passed: testMessageSizeCheck
[info] Test Starting: testTruncateTo
[info] Test Passed: testTruncateTo
[info] == core-kafka / kafka.log.LogTest ==
[info] 
[info] == core-kafka / kafka.consumer.TopicFilterTest ==
[info] Test Starting: testWhitelists
[info] Test Passed: testWhitelists
[info] Test Starting: testBlacklists
[info] Test Passed: testBlacklists
[info] == core-kafka / kafka.consumer.TopicFilterTest ==
[info] 
[info] == core-kafka / kafka.network.SocketServerTest ==
[info] Test Starting: simpleRequest
[info] Test Passed: simpleRequest
[info] Test Starting: tooBigRequestIsRejected
[info] Test Passed: tooBigRequestIsRejected
[info] == core-kafka / kafka.network.SocketServerTest ==
[info] 
[info] == core-kafka / test-finish ==
[error] Failed: : Total 136, Failed 1, Errors 0, Passed 135, Skipped 0
[info] == core-kafka / test-finish ==
[info] 
[info] == core-kafka / Test cleanup 1 ==
[info] Deleting directory /var/tmp/sbt_37b84a09
[info] == core-kafka / Test cleanup 1 ==
[info] 
[info] == core-kafka / test-cleanup ==
[info] == core-kafka / test-cleanup ==
[info] 
[info] == perf / test-compile ==
[info]   Source analysis: 0 new/modified, 0 indirectly invalidated, 0 removed.
[info] Compiling test sources...
[info] Nothing to compile.
[info]   Post-analysis: 0 classes.
[info] == perf / test-compile ==
[info] 
[info] == hadoop consumer / test-compile ==
[info]   Source analysis: 0 new/modified, 0 indirectly invalidated, 0 removed.
[info] Compiling test sources...
[info] Nothing to compile.
[info]   Post-analysis: 0 classes.
[info] == hadoop consumer / test-compile ==
[info] 
[info] == hadoop producer / copy-test-resources ==
[info] == hadoop producer / copy-test-resources ==
[info] 
[info] == hadoop producer / copy-resources ==
[info] == hadoop producer / copy-resources ==
[info] 
[info] == java-examples / copy-resources ==
[info] == java-examples / copy-resources ==
[info] 
[info] == perf / copy-test-resources ==
[info] == perf / copy-test-resources ==
[error] Error running kafka.message.CompressionUtilTest: Test FAILED
[error] Error running test: One or more subtasks failed
[info] 
[info] Total time: 204 s, completed Sep 25, 2012 7:24:48 PM
[info] 
[info] Total session time: 205 s, completed Sep 25, 2012 7:24:48 PM
[error] Error during build.
Build step 'Execute shell' marked build as failure