You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@kafka.apache.org by Apache Jenkins Server <je...@builds.apache.org> on 2012/11/02 06:29:06 UTC

Build failed in Jenkins: Kafka-0.8 #98

See <https://builds.apache.org/job/Kafka-0.8/98/changes>

Changes:

[junrao] KafkaRequestHandler needs to handle exceptions; patched by Yang Ye; reviewed by Jun Rao; KAFKA-491

------------------------------------------
[...truncated 1303 lines...]
	at org.junit.Assert.assertEquals(Assert.java:104)
	at kafka.server.LogRecoveryTest.testHWCheckpointWithFailuresSingleLogSegment(LogRecoveryTest.scala:149)
	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
	at java.lang.reflect.Method.invoke(Method.java:597)
	at junit.framework.TestCase.runTest(TestCase.java:164)
	at junit.framework.TestCase.runBare(TestCase.java:130)
	at junit.framework.TestResult$1.protect(TestResult.java:110)
	at junit.framework.TestResult.runProtected(TestResult.java:128)
	at junit.framework.TestResult.run(TestResult.java:113)
	at junit.framework.TestCase.run(TestCase.java:120)
	at junit.framework.TestSuite.runTest(TestSuite.java:228)
	at junit.framework.TestSuite.run(TestSuite.java:223)
	at junit.framework.TestSuite.runTest(TestSuite.java:228)
	at junit.framework.TestSuite.run(TestSuite.java:223)
	at org.scalatest.junit.JUnit3Suite.run(JUnit3Suite.scala:309)
	at org.scalatest.tools.ScalaTestFramework$ScalaTestRunner.run(ScalaTestFramework.scala:40)
	at sbt.TestRunner.run(TestFramework.scala:53)
	at sbt.TestRunner.runTest$1(TestFramework.scala:67)
	at sbt.TestRunner.run(TestFramework.scala:76)
	at sbt.TestFramework$$anonfun$10$$anonfun$apply$11.runTest$2(TestFramework.scala:194)
	at sbt.TestFramework$$anonfun$10$$anonfun$apply$11$$anonfun$apply$12.apply(TestFramework.scala:205)
	at sbt.TestFramework$$anonfun$10$$anonfun$apply$11$$anonfun$apply$12.apply(TestFramework.scala:205)
	at sbt.NamedTestTask.run(TestFramework.scala:92)
	at sbt.ScalaProject$$anonfun$sbt$ScalaProject$$toTask$1.apply(ScalaProject.scala:193)
	at sbt.ScalaProject$$anonfun$sbt$ScalaProject$$toTask$1.apply(ScalaProject.scala:193)
	at sbt.TaskManager$Task.invoke(TaskManager.scala:62)
	at sbt.impl.RunTask.doRun$1(RunTask.scala:77)
	at sbt.impl.RunTask.runTask(RunTask.scala:85)
	at sbt.impl.RunTask.sbt$impl$RunTask$$runIfNotRoot(RunTask.scala:60)
	at sbt.impl.RunTask$$anonfun$runTasksExceptRoot$2.apply(RunTask.scala:48)
	at sbt.impl.RunTask$$anonfun$runTasksExceptRoot$2.apply(RunTask.scala:48)
	at sbt.Distributor$Run$Worker$$anonfun$2.apply(ParallelRunner.scala:131)
	at sbt.Distributor$Run$Worker$$anonfun$2.apply(ParallelRunner.scala:131)
	at sbt.Control$.trapUnit(Control.scala:19)
	at sbt.Distributor$Run$Worker.run(ParallelRunner.scala:131)
[info] Test Starting: testHWCheckpointNoFailuresMultipleLogSegments(kafka.server.LogRecoveryTest)
[info] Test Passed: testHWCheckpointNoFailuresMultipleLogSegments(kafka.server.LogRecoveryTest)
[info] Test Starting: testHWCheckpointWithFailuresMultipleLogSegments(kafka.server.LogRecoveryTest)
[info] Test Passed: testHWCheckpointWithFailuresMultipleLogSegments(kafka.server.LogRecoveryTest)
[info] == core-kafka / kafka.server.LogRecoveryTest ==
[info] 
[info] == core-kafka / kafka.log.FileMessageSetTest ==
[info] Test Starting: testWrittenEqualsRead
[info] Test Passed: testWrittenEqualsRead
[info] Test Starting: testIteratorIsConsistent
[info] Test Passed: testIteratorIsConsistent
[info] Test Starting: testSizeInBytes
[info] Test Passed: testSizeInBytes
[info] Test Starting: testWriteTo
[info] Test Passed: testWriteTo
[info] Test Starting: testFileSize
[info] Test Passed: testFileSize
[info] Test Starting: testIterationOverPartialAndTruncation
[info] Test Passed: testIterationOverPartialAndTruncation
[info] Test Starting: testIterationDoesntChangePosition
[info] Test Passed: testIterationDoesntChangePosition
[info] Test Starting: testRead
[info] Test Passed: testRead
[info] Test Starting: testSearch
[info] Test Passed: testSearch
[info] == core-kafka / kafka.log.FileMessageSetTest ==
[info] 
[info] == core-kafka / kafka.log.LogTest ==
[info] Test Starting: testTimeBasedLogRoll
[error] Test Failed: testTimeBasedLogRoll
junit.framework.AssertionFailedError: There should be exactly one segment. expected:<1> but was:<2>
	at junit.framework.Assert.fail(Assert.java:47)
	at junit.framework.Assert.failNotEquals(Assert.java:277)
	at junit.framework.Assert.assertEquals(Assert.java:64)
	at junit.framework.Assert.assertEquals(Assert.java:195)
	at kafka.log.LogTest.testTimeBasedLogRoll(LogTest.scala:73)
	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
	at java.lang.reflect.Method.invoke(Method.java:597)
	at org.junit.internal.runners.TestMethodRunner.executeMethodBody(TestMethodRunner.java:99)
	at org.junit.internal.runners.TestMethodRunner.runUnprotected(TestMethodRunner.java:81)
	at org.junit.internal.runners.BeforeAndAfterRunner.runProtected(BeforeAndAfterRunner.java:34)
	at org.junit.internal.runners.TestMethodRunner.runMethod(TestMethodRunner.java:75)
	at org.junit.internal.runners.TestMethodRunner.run(TestMethodRunner.java:45)
	at org.junit.internal.runners.TestClassMethodsRunner.invokeTestMethod(TestClassMethodsRunner.java:71)
	at org.junit.internal.runners.TestClassMethodsRunner.run(TestClassMethodsRunner.java:35)
	at org.junit.internal.runners.TestClassRunner$1.runUnprotected(TestClassRunner.java:42)
	at org.junit.internal.runners.BeforeAndAfterRunner.runProtected(BeforeAndAfterRunner.java:34)
	at org.junit.internal.runners.TestClassRunner.run(TestClassRunner.java:52)
	at org.junit.internal.runners.CompositeRunner.run(CompositeRunner.java:29)
	at org.junit.runner.JUnitCore.run(JUnitCore.java:121)
	at org.junit.runner.JUnitCore.run(JUnitCore.java:100)
	at org.junit.runner.JUnitCore.run(JUnitCore.java:91)
	at org.scalatest.junit.JUnitSuite$class.run(JUnitSuite.scala:261)
	at kafka.log.LogTest.run(LogTest.scala:32)
	at org.scalatest.tools.ScalaTestFramework$ScalaTestRunner.run(ScalaTestFramework.scala:40)
	at sbt.TestRunner.run(TestFramework.scala:53)
	at sbt.TestRunner.runTest$1(TestFramework.scala:67)
	at sbt.TestRunner.run(TestFramework.scala:76)
	at sbt.TestFramework$$anonfun$10$$anonfun$apply$11.runTest$2(TestFramework.scala:194)
	at sbt.TestFramework$$anonfun$10$$anonfun$apply$11$$anonfun$apply$12.apply(TestFramework.scala:205)
	at sbt.TestFramework$$anonfun$10$$anonfun$apply$11$$anonfun$apply$12.apply(TestFramework.scala:205)
	at sbt.NamedTestTask.run(TestFramework.scala:92)
	at sbt.ScalaProject$$anonfun$sbt$ScalaProject$$toTask$1.apply(ScalaProject.scala:193)
	at sbt.ScalaProject$$anonfun$sbt$ScalaProject$$toTask$1.apply(ScalaProject.scala:193)
	at sbt.TaskManager$Task.invoke(TaskManager.scala:62)
	at sbt.impl.RunTask.doRun$1(RunTask.scala:77)
	at sbt.impl.RunTask.runTask(RunTask.scala:85)
	at sbt.impl.RunTask.sbt$impl$RunTask$$runIfNotRoot(RunTask.scala:60)
	at sbt.impl.RunTask$$anonfun$runTasksExceptRoot$2.apply(RunTask.scala:48)
	at sbt.impl.RunTask$$anonfun$runTasksExceptRoot$2.apply(RunTask.scala:48)
	at sbt.Distributor$Run$Worker$$anonfun$2.apply(ParallelRunner.scala:131)
	at sbt.Distributor$Run$Worker$$anonfun$2.apply(ParallelRunner.scala:131)
	at sbt.Control$.trapUnit(Control.scala:19)
	at sbt.Distributor$Run$Worker.run(ParallelRunner.scala:131)
[info] Test Starting: testSizeBasedLogRoll
[info] Test Passed: testSizeBasedLogRoll
[info] Test Starting: testLoadEmptyLog
[info] Test Passed: testLoadEmptyLog
[info] Test Starting: testAppendAndRead
[info] Test Passed: testAppendAndRead
[info] Test Starting: testReadOutOfRange
[info] Test Passed: testReadOutOfRange
[info] Test Starting: testLogRolls
[info] Test Passed: testLogRolls
[info] Test Starting: testCompressedMessages
[info] Test Passed: testCompressedMessages
[info] Test Starting: testFindSegment
[info] Test Passed: testFindSegment
[info] Test Starting: testEdgeLogRollsStartingAtZero
[info] Test Passed: testEdgeLogRollsStartingAtZero
[info] Test Starting: testEdgeLogRollsStartingAtNonZero
[info] Test Passed: testEdgeLogRollsStartingAtNonZero
[info] Test Starting: testMessageSizeCheck
[info] Test Passed: testMessageSizeCheck
[info] Test Starting: testLogRecoversToCorrectOffset
[info] Test Passed: testLogRecoversToCorrectOffset
[info] Test Starting: testTruncateTo
[info] Test Passed: testTruncateTo
[info] Test Starting: testAppendWithoutOffsetAssignment
[info] Test Passed: testAppendWithoutOffsetAssignment
[info] Test Starting: testReopenThenTruncate
[info] Test Passed: testReopenThenTruncate
[info] == core-kafka / kafka.log.LogTest ==
[info] 
[info] == core-kafka / kafka.producer.AsyncProducerTest ==
[info] Test Starting: testProducerQueueSize(kafka.producer.AsyncProducerTest)
[info] Test Passed: testProducerQueueSize(kafka.producer.AsyncProducerTest)
[info] Test Starting: testProduceAfterClosed(kafka.producer.AsyncProducerTest)
[info] Test Passed: testProduceAfterClosed(kafka.producer.AsyncProducerTest)
[info] Test Starting: testBatchSize(kafka.producer.AsyncProducerTest)
[info] Test Passed: testBatchSize(kafka.producer.AsyncProducerTest)
[info] Test Starting: testQueueTimeExpired(kafka.producer.AsyncProducerTest)
[info] Test Passed: testQueueTimeExpired(kafka.producer.AsyncProducerTest)
[info] Test Starting: testPartitionAndCollateEvents(kafka.producer.AsyncProducerTest)
[info] Test Passed: testPartitionAndCollateEvents(kafka.producer.AsyncProducerTest)
[info] Test Starting: testSerializeEvents(kafka.producer.AsyncProducerTest)
[info] Test Passed: testSerializeEvents(kafka.producer.AsyncProducerTest)
[info] Test Starting: testInvalidPartition(kafka.producer.AsyncProducerTest)
[info] Test Passed: testInvalidPartition(kafka.producer.AsyncProducerTest)
[info] Test Starting: testNoBroker(kafka.producer.AsyncProducerTest)
[info] Test Passed: testNoBroker(kafka.producer.AsyncProducerTest)
[info] Test Starting: testIncompatibleEncoder(kafka.producer.AsyncProducerTest)
[info] Test Passed: testIncompatibleEncoder(kafka.producer.AsyncProducerTest)
[info] Test Starting: testRandomPartitioner(kafka.producer.AsyncProducerTest)
[info] Test Passed: testRandomPartitioner(kafka.producer.AsyncProducerTest)
[info] Test Starting: testBrokerListAndAsync(kafka.producer.AsyncProducerTest)
[info] Test Passed: testBrokerListAndAsync(kafka.producer.AsyncProducerTest)
[info] Test Starting: testFailedSendRetryLogic(kafka.producer.AsyncProducerTest)
[info] Test Passed: testFailedSendRetryLogic(kafka.producer.AsyncProducerTest)
[info] Test Starting: testJavaProducer(kafka.producer.AsyncProducerTest)
[info] Test Passed: testJavaProducer(kafka.producer.AsyncProducerTest)
[info] Test Starting: testInvalidConfiguration(kafka.producer.AsyncProducerTest)
[info] Test Passed: testInvalidConfiguration(kafka.producer.AsyncProducerTest)
[info] == core-kafka / kafka.producer.AsyncProducerTest ==
[info] 
[info] == core-kafka / kafka.log.LogOffsetTest ==
[info] Test Starting: testGetOffsetsForUnknownTopic(kafka.log.LogOffsetTest)
[info] Test Passed: testGetOffsetsForUnknownTopic(kafka.log.LogOffsetTest)
[info] Test Starting: testGetOffsetsBeforeLatestTime(kafka.log.LogOffsetTest)
[info] Test Passed: testGetOffsetsBeforeLatestTime(kafka.log.LogOffsetTest)
[info] Test Starting: testEmptyLogsGetOffsets(kafka.log.LogOffsetTest)
[info] Test Passed: testEmptyLogsGetOffsets(kafka.log.LogOffsetTest)
[info] Test Starting: testGetOffsetsBeforeNow(kafka.log.LogOffsetTest)
[info] Test Passed: testGetOffsetsBeforeNow(kafka.log.LogOffsetTest)
[info] Test Starting: testGetOffsetsBeforeEarliestTime(kafka.log.LogOffsetTest)
[info] Test Passed: testGetOffsetsBeforeEarliestTime(kafka.log.LogOffsetTest)
[info] == core-kafka / kafka.log.LogOffsetTest ==
[info] 
[info] == core-kafka / kafka.server.ReplicaFetchTest ==
[info] Test Starting: testReplicaFetcherThread(kafka.server.ReplicaFetchTest)
[info] Test Passed: testReplicaFetcherThread(kafka.server.ReplicaFetchTest)
[info] == core-kafka / kafka.server.ReplicaFetchTest ==
[info] 
[info] == core-kafka / kafka.server.IsrExpirationTest ==
[info] Test Starting: testIsrExpirationForStuckFollowers(kafka.server.IsrExpirationTest)
[info] Test Passed: testIsrExpirationForStuckFollowers(kafka.server.IsrExpirationTest)
[info] Test Starting: testIsrExpirationForSlowFollowers(kafka.server.IsrExpirationTest)
[info] Test Passed: testIsrExpirationForSlowFollowers(kafka.server.IsrExpirationTest)
[info] == core-kafka / kafka.server.IsrExpirationTest ==
[info] 
[info] == core-kafka / kafka.message.MessageCompressionTest ==
[info] Test Starting: testSimpleCompressDecompress
[info] Test Passed: testSimpleCompressDecompress
[info] Test Starting: testComplexCompressDecompress
[info] Test Passed: testComplexCompressDecompress
[info] == core-kafka / kafka.message.MessageCompressionTest ==
[info] 
[info] == core-kafka / kafka.producer.SyncProducerTest ==
[info] Test Starting: testReachableServer(kafka.producer.SyncProducerTest)
[info] Test Passed: testReachableServer(kafka.producer.SyncProducerTest)
[info] Test Starting: testEmptyProduceRequest(kafka.producer.SyncProducerTest)
[info] Test Passed: testEmptyProduceRequest(kafka.producer.SyncProducerTest)
[info] Test Starting: testMessageSizeTooLarge(kafka.producer.SyncProducerTest)
[info] Test Passed: testMessageSizeTooLarge(kafka.producer.SyncProducerTest)
[info] Test Starting: testProduceCorrectlyReceivesResponse(kafka.producer.SyncProducerTest)
[info] Test Passed: testProduceCorrectlyReceivesResponse(kafka.producer.SyncProducerTest)
[info] Test Starting: testProducerCanTimeout(kafka.producer.SyncProducerTest)
[info] Test Passed: testProducerCanTimeout(kafka.producer.SyncProducerTest)
[info] == core-kafka / kafka.producer.SyncProducerTest ==
[info] 
[info] == core-kafka / kafka.integration.LazyInitProducerTest ==
[info] Test Starting: testProduceAndMultiFetch(kafka.integration.LazyInitProducerTest)
[info] Test Passed: testProduceAndMultiFetch(kafka.integration.LazyInitProducerTest)
[info] Test Starting: testMultiProduce(kafka.integration.LazyInitProducerTest)
[info] Test Passed: testMultiProduce(kafka.integration.LazyInitProducerTest)
[info] Test Starting: testProduceAndFetch(kafka.integration.LazyInitProducerTest)
[info] Test Passed: testProduceAndFetch(kafka.integration.LazyInitProducerTest)
[info] Test Starting: testMultiProduceResend(kafka.integration.LazyInitProducerTest)
[info] Test Passed: testMultiProduceResend(kafka.integration.LazyInitProducerTest)
[info] == core-kafka / kafka.integration.LazyInitProducerTest ==
[info] 
[info] == core-kafka / Test cleanup 1 ==
[info] Deleting directory /tmp/sbt_21fd04f3
[info] == core-kafka / Test cleanup 1 ==
[info] 
[info] == core-kafka / test-finish ==
[error] Failed: : Total 161, Failed 2, Errors 0, Passed 159, Skipped 0
[info] == core-kafka / test-finish ==
[info] 
[info] == core-kafka / test-cleanup ==
[info] == core-kafka / test-cleanup ==
[error] Error running kafka.server.LogRecoveryTest: Test FAILED
[error] Error running kafka.log.LogTest: Test FAILED
[error] Error running test: One or more subtasks failed
[info] 
[info] Total time: 226 s, completed Nov 2, 2012 5:29:06 AM
[info] 
[info] Total session time: 227 s, completed Nov 2, 2012 5:29:06 AM
[error] Error during build.
Build step 'Execute shell' marked build as failure

Build failed in Jenkins: Kafka-0.8 #103

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/Kafka-0.8/103/changes>

Changes:

[junrao] Add test cases to test log size retention and more; patched by John Fung; reviewed by Jun Rao; KAFKA-591

------------------------------------------
[...truncated 3014 lines...]
	at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
	at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
	at scala.collection.immutable.Map$Map1.foreach(Map.scala:105)
	at scala.collection.TraversableLike$class.map(TraversableLike.scala:206)
	at scala.collection.immutable.Map$Map1.map(Map.scala:93)
	at kafka.server.KafkaApis.appendToLocalLog(KafkaApis.scala:243)
	at kafka.server.KafkaApis.handleProducerRequest(KafkaApis.scala:183)
	at kafka.server.KafkaApis.handle(KafkaApis.scala:57)
	at kafka.server.KafkaRequestHandler.run(KafkaRequestHandler.scala:41)
	at java.lang.Thread.run(Thread.java:662)
[info] Test Passed: testReachableServer(kafka.producer.SyncProducerTest)
[info] Test Starting: testEmptyProduceRequest(kafka.producer.SyncProducerTest)
[info] Test Passed: testEmptyProduceRequest(kafka.producer.SyncProducerTest)
[info] Test Starting: testMessageSizeTooLarge(kafka.producer.SyncProducerTest)
[2012-11-09 23:18:53,243] ERROR [KafkaApi-0] Error processing ProducerRequest on test:0 (kafka.server.KafkaApis:102)
kafka.common.MessageSizeTooLargeException: Message size is 1000023 bytes which exceeds the maximum configured message size of 1000000.
	at kafka.log.Log$$anonfun$analyzeAndValidateMessageSet$1.apply(Log.scala:327)
	at kafka.log.Log$$anonfun$analyzeAndValidateMessageSet$1.apply(Log.scala:313)
	at scala.collection.Iterator$class.foreach(Iterator.scala:631)
	at kafka.utils.IteratorTemplate.foreach(IteratorTemplate.scala:32)
	at kafka.log.Log.analyzeAndValidateMessageSet(Log.scala:313)
	at kafka.log.Log.append(Log.scala:241)
	at kafka.server.KafkaApis$$anonfun$appendToLocalLog$2.apply(KafkaApis.scala:250)
	at kafka.server.KafkaApis$$anonfun$appendToLocalLog$2.apply(KafkaApis.scala:243)
	at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
	at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
	at scala.collection.immutable.Map$Map1.foreach(Map.scala:105)
	at scala.collection.TraversableLike$class.map(TraversableLike.scala:206)
	at scala.collection.immutable.Map$Map1.map(Map.scala:93)
	at kafka.server.KafkaApis.appendToLocalLog(KafkaApis.scala:243)
	at kafka.server.KafkaApis.handleProducerRequest(KafkaApis.scala:183)
	at kafka.server.KafkaApis.handle(KafkaApis.scala:57)
	at kafka.server.KafkaRequestHandler.run(KafkaRequestHandler.scala:41)
	at java.lang.Thread.run(Thread.java:662)
[info] Test Passed: testMessageSizeTooLarge(kafka.producer.SyncProducerTest)
[info] Test Starting: testProduceCorrectlyReceivesResponse(kafka.producer.SyncProducerTest)
[2012-11-09 23:18:53,502] ERROR [KafkaApi-0] Error processing ProducerRequest on topic1:0 (kafka.server.KafkaApis:102)
kafka.common.UnknownTopicOrPartitionException: Topic topic1 partition 0 doesn't exist on 0
	at kafka.server.ReplicaManager.getLeaderReplicaIfLocal(ReplicaManager.scala:142)
	at kafka.server.KafkaApis$$anonfun$appendToLocalLog$2.apply(KafkaApis.scala:248)
	at kafka.server.KafkaApis$$anonfun$appendToLocalLog$2.apply(KafkaApis.scala:243)
	at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
	at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
	at scala.collection.immutable.Map$Map3.foreach(Map.scala:152)
	at scala.collection.TraversableLike$class.map(TraversableLike.scala:206)
	at scala.collection.immutable.Map$Map3.map(Map.scala:132)
	at kafka.server.KafkaApis.appendToLocalLog(KafkaApis.scala:243)
	at kafka.server.KafkaApis.handleProducerRequest(KafkaApis.scala:183)
	at kafka.server.KafkaApis.handle(KafkaApis.scala:57)
	at kafka.server.KafkaRequestHandler.run(KafkaRequestHandler.scala:41)
	at java.lang.Thread.run(Thread.java:662)
[2012-11-09 23:18:53,503] ERROR [KafkaApi-0] Error processing ProducerRequest on topic3:0 (kafka.server.KafkaApis:102)
kafka.common.UnknownTopicOrPartitionException: Topic topic3 partition 0 doesn't exist on 0
	at kafka.server.ReplicaManager.getLeaderReplicaIfLocal(ReplicaManager.scala:142)
	at kafka.server.KafkaApis$$anonfun$appendToLocalLog$2.apply(KafkaApis.scala:248)
	at kafka.server.KafkaApis$$anonfun$appendToLocalLog$2.apply(KafkaApis.scala:243)
	at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
	at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
	at scala.collection.immutable.Map$Map3.foreach(Map.scala:152)
	at scala.collection.TraversableLike$class.map(TraversableLike.scala:206)
	at scala.collection.immutable.Map$Map3.map(Map.scala:132)
	at kafka.server.KafkaApis.appendToLocalLog(KafkaApis.scala:243)
	at kafka.server.KafkaApis.handleProducerRequest(KafkaApis.scala:183)
	at kafka.server.KafkaApis.handle(KafkaApis.scala:57)
	at kafka.server.KafkaRequestHandler.run(KafkaRequestHandler.scala:41)
	at java.lang.Thread.run(Thread.java:662)
[2012-11-09 23:18:53,504] ERROR [KafkaApi-0] Error processing ProducerRequest on topic2:0 (kafka.server.KafkaApis:102)
kafka.common.UnknownTopicOrPartitionException: Topic topic2 partition 0 doesn't exist on 0
	at kafka.server.ReplicaManager.getLeaderReplicaIfLocal(ReplicaManager.scala:142)
	at kafka.server.KafkaApis$$anonfun$appendToLocalLog$2.apply(KafkaApis.scala:248)
	at kafka.server.KafkaApis$$anonfun$appendToLocalLog$2.apply(KafkaApis.scala:243)
	at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
	at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
	at scala.collection.immutable.Map$Map3.foreach(Map.scala:152)
	at scala.collection.TraversableLike$class.map(TraversableLike.scala:206)
	at scala.collection.immutable.Map$Map3.map(Map.scala:132)
	at kafka.server.KafkaApis.appendToLocalLog(KafkaApis.scala:243)
	at kafka.server.KafkaApis.handleProducerRequest(KafkaApis.scala:183)
	at kafka.server.KafkaApis.handle(KafkaApis.scala:57)
	at kafka.server.KafkaRequestHandler.run(KafkaRequestHandler.scala:41)
	at java.lang.Thread.run(Thread.java:662)
[2012-11-09 23:18:54,657] ERROR [KafkaApi-0] Error processing ProducerRequest on topic2:0 (kafka.server.KafkaApis:102)
kafka.common.UnknownTopicOrPartitionException: Topic topic2 partition 0 doesn't exist on 0
	at kafka.server.ReplicaManager.getLeaderReplicaIfLocal(ReplicaManager.scala:142)
	at kafka.server.KafkaApis$$anonfun$appendToLocalLog$2.apply(KafkaApis.scala:248)
	at kafka.server.KafkaApis$$anonfun$appendToLocalLog$2.apply(KafkaApis.scala:243)
	at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
	at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
	at scala.collection.immutable.Map$Map3.foreach(Map.scala:152)
	at scala.collection.TraversableLike$class.map(TraversableLike.scala:206)
	at scala.collection.immutable.Map$Map3.map(Map.scala:132)
	at kafka.server.KafkaApis.appendToLocalLog(KafkaApis.scala:243)
	at kafka.server.KafkaApis.handleProducerRequest(KafkaApis.scala:183)
	at kafka.server.KafkaApis.handle(KafkaApis.scala:57)
	at kafka.server.KafkaRequestHandler.run(KafkaRequestHandler.scala:41)
	at java.lang.Thread.run(Thread.java:662)
[info] Test Passed: testProduceCorrectlyReceivesResponse(kafka.producer.SyncProducerTest)
[info] Test Starting: testProducerCanTimeout(kafka.producer.SyncProducerTest)
[info] Test Passed: testProducerCanTimeout(kafka.producer.SyncProducerTest)
[info] == core-kafka / kafka.producer.SyncProducerTest ==
[info] 
[info] == core-kafka / kafka.log4j.KafkaLog4jAppenderTest ==
[info] Test Starting: testKafkaLog4jConfigs(kafka.log4j.KafkaLog4jAppenderTest)
log4j:WARN No appenders could be found for logger (org.I0Itec.zkclient.ZkEventThread).
log4j:WARN Please initialize the log4j system properly.
[info] Test Passed: testKafkaLog4jConfigs(kafka.log4j.KafkaLog4jAppenderTest)
[info] Test Starting: testLog4jAppends(kafka.log4j.KafkaLog4jAppenderTest)
[info] Test Passed: testLog4jAppends(kafka.log4j.KafkaLog4jAppenderTest)
[info] == core-kafka / kafka.log4j.KafkaLog4jAppenderTest ==
[info] 
[info] == core-kafka / kafka.integration.AutoOffsetResetTest ==
[info] Test Starting: testResetToEarliestWhenOffsetTooHigh(kafka.integration.AutoOffsetResetTest)
[info] Test Passed: testResetToEarliestWhenOffsetTooHigh(kafka.integration.AutoOffsetResetTest)
[info] Test Starting: testResetToEarliestWhenOffsetTooLow(kafka.integration.AutoOffsetResetTest)
[info] Test Passed: testResetToEarliestWhenOffsetTooLow(kafka.integration.AutoOffsetResetTest)
[info] Test Starting: testResetToLatestWhenOffsetTooHigh(kafka.integration.AutoOffsetResetTest)
[info] Test Passed: testResetToLatestWhenOffsetTooHigh(kafka.integration.AutoOffsetResetTest)
[info] Test Starting: testResetToLatestWhenOffsetTooLow(kafka.integration.AutoOffsetResetTest)
[info] Test Passed: testResetToLatestWhenOffsetTooLow(kafka.integration.AutoOffsetResetTest)
[info] == core-kafka / kafka.integration.AutoOffsetResetTest ==
[info] 
[info] == core-kafka / kafka.log.OffsetIndexTest ==
[info] Test Starting: truncate
[info] Test Passed: truncate
[info] Test Starting: randomLookupTest
[info] Test Passed: randomLookupTest
[info] Test Starting: lookupExtremeCases
[info] Test Passed: lookupExtremeCases
[info] Test Starting: appendTooMany
[info] Test Passed: appendTooMany
[info] Test Starting: appendOutOfOrder
[info] Test Passed: appendOutOfOrder
[info] Test Starting: testReopen
[info] Test Passed: testReopen
[info] == core-kafka / kafka.log.OffsetIndexTest ==
[info] 
[info] == core-kafka / kafka.integration.LazyInitProducerTest ==
[info] Test Starting: testProduceAndMultiFetch(kafka.integration.LazyInitProducerTest)
[info] Test Passed: testProduceAndMultiFetch(kafka.integration.LazyInitProducerTest)
[info] Test Starting: testMultiProduce(kafka.integration.LazyInitProducerTest)
[info] Test Passed: testMultiProduce(kafka.integration.LazyInitProducerTest)
[info] Test Starting: testProduceAndFetch(kafka.integration.LazyInitProducerTest)
[info] Test Passed: testProduceAndFetch(kafka.integration.LazyInitProducerTest)
[info] Test Starting: testMultiProduceResend(kafka.integration.LazyInitProducerTest)
[info] Test Passed: testMultiProduceResend(kafka.integration.LazyInitProducerTest)
[info] == core-kafka / kafka.integration.LazyInitProducerTest ==
[info] 
[info] == core-kafka / kafka.message.ByteBufferMessageSetTest ==
[info] Test Starting: testWrittenEqualsRead
[info] Test Passed: testWrittenEqualsRead
[info] Test Starting: testIteratorIsConsistent
[info] Test Passed: testIteratorIsConsistent
[info] Test Starting: testSizeInBytes
[info] Test Passed: testSizeInBytes
[info] Test Starting: testWriteTo
[info] Test Passed: testWriteTo
[info] Test Starting: testValidBytes
[info] Test Passed: testValidBytes
[info] Test Starting: testValidBytesWithCompression
[info] Test Passed: testValidBytesWithCompression
[info] Test Starting: testEquals
[info] Test Passed: testEquals
[info] Test Starting: testIterator
[info] Test Passed: testIterator
[info] Test Starting: testOffsetAssignment
[info] Test Passed: testOffsetAssignment
[info] == core-kafka / kafka.message.ByteBufferMessageSetTest ==
[info] 
[info] == core-kafka / kafka.log.LogOffsetTest ==
[info] Test Starting: testGetOffsetsForUnknownTopic(kafka.log.LogOffsetTest)
[info] Test Passed: testGetOffsetsForUnknownTopic(kafka.log.LogOffsetTest)
[info] Test Starting: testGetOffsetsBeforeLatestTime(kafka.log.LogOffsetTest)
[info] Test Passed: testGetOffsetsBeforeLatestTime(kafka.log.LogOffsetTest)
[info] Test Starting: testEmptyLogsGetOffsets(kafka.log.LogOffsetTest)
[info] Test Passed: testEmptyLogsGetOffsets(kafka.log.LogOffsetTest)
[info] Test Starting: testGetOffsetsBeforeNow(kafka.log.LogOffsetTest)
[info] Test Passed: testGetOffsetsBeforeNow(kafka.log.LogOffsetTest)
[info] Test Starting: testGetOffsetsBeforeEarliestTime(kafka.log.LogOffsetTest)
[info] Test Passed: testGetOffsetsBeforeEarliestTime(kafka.log.LogOffsetTest)
[info] == core-kafka / kafka.log.LogOffsetTest ==
[info] 
[info] == core-kafka / kafka.consumer.TopicFilterTest ==
[info] Test Starting: testWhitelists
[info] Test Passed: testWhitelists
[info] Test Starting: testBlacklists
[info] Test Passed: testBlacklists
[info] == core-kafka / kafka.consumer.TopicFilterTest ==
[info] 
[info] == core-kafka / kafka.javaapi.message.ByteBufferMessageSetTest ==
[info] Test Starting: testWrittenEqualsRead
[info] Test Passed: testWrittenEqualsRead
[info] Test Starting: testIteratorIsConsistent
[info] Test Passed: testIteratorIsConsistent
[info] Test Starting: testSizeInBytes
[info] Test Passed: testSizeInBytes
[info] Test Starting: testEquals
[info] Test Passed: testEquals
[info] Test Starting: testIteratorIsConsistentWithCompression
[info] Test Passed: testIteratorIsConsistentWithCompression
[info] Test Starting: testSizeInBytesWithCompression
[info] Test Passed: testSizeInBytesWithCompression
[info] Test Starting: testEqualsWithCompression
[info] Test Passed: testEqualsWithCompression
[info] == core-kafka / kafka.javaapi.message.ByteBufferMessageSetTest ==
[info] 
[info] == core-kafka / kafka.integration.FetcherTest ==
[info] Test Starting: testFetcher(kafka.integration.FetcherTest)
[info] Test Passed: testFetcher(kafka.integration.FetcherTest)
[info] == core-kafka / kafka.integration.FetcherTest ==
[info] 
[info] == core-kafka / kafka.javaapi.consumer.ZookeeperConsumerConnectorTest ==
[info] Test Starting: testBasic(kafka.javaapi.consumer.ZookeeperConsumerConnectorTest)
[info] Test Passed: testBasic(kafka.javaapi.consumer.ZookeeperConsumerConnectorTest)
[info] == core-kafka / kafka.javaapi.consumer.ZookeeperConsumerConnectorTest ==
[info] 
[info] == core-kafka / kafka.metrics.KafkaTimerTest ==
[info] Test Starting: testKafkaTimer(kafka.metrics.KafkaTimerTest)
[info] Test Passed: testKafkaTimer(kafka.metrics.KafkaTimerTest)
[info] == core-kafka / kafka.metrics.KafkaTimerTest ==
[info] 
[info] == core-kafka / kafka.server.RequestPurgatoryTest ==
[info] Test Starting: testRequestSatisfaction(kafka.server.RequestPurgatoryTest)
[info] Test Passed: testRequestSatisfaction(kafka.server.RequestPurgatoryTest)
[info] Test Starting: testRequestExpiry(kafka.server.RequestPurgatoryTest)
[info] Test Passed: testRequestExpiry(kafka.server.RequestPurgatoryTest)
[info] == core-kafka / kafka.server.RequestPurgatoryTest ==
[info] 
[info] == core-kafka / Test cleanup 1 ==
[info] Deleting directory /tmp/sbt_8262927a
[info] == core-kafka / Test cleanup 1 ==
[info] 
[info] == core-kafka / test-finish ==
[error] Failed: : Total 165, Failed 2, Errors 0, Passed 163, Skipped 0
[info] == core-kafka / test-finish ==
[info] 
[info] == core-kafka / test-cleanup ==
[info] == core-kafka / test-cleanup ==
[info] 
[info] == hadoop consumer / copy-test-resources ==
[info] == hadoop consumer / copy-test-resources ==
[error] Error running kafka.log.LogTest: Test FAILED
[error] Error running kafka.server.LogRecoveryTest: Test FAILED
[error] Error running test: One or more subtasks failed
[info] 
[info] Total time: 173 s, completed Nov 9, 2012 11:19:19 PM
[info] 
[info] Total session time: 174 s, completed Nov 9, 2012 11:19:19 PM
[error] Error during build.
Build step 'Execute shell' marked build as failure

Build failed in Jenkins: Kafka-0.8 #102

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/Kafka-0.8/102/changes>

Changes:

[junrao] Fix commit() in zk consumer for compressed messages; patched by Swapnil Ghike; reviewed by Jun Rao; KAFKA-546

------------------------------------------
[...truncated 3920 lines...]
	at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
	at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
	at scala.collection.immutable.Map$Map3.foreach(Map.scala:152)
	at scala.collection.TraversableLike$class.map(TraversableLike.scala:206)
	at scala.collection.immutable.Map$Map3.map(Map.scala:132)
	at kafka.server.KafkaApis.appendToLocalLog(KafkaApis.scala:243)
	at kafka.server.KafkaApis.handleProducerRequest(KafkaApis.scala:183)
	at kafka.server.KafkaApis.handle(KafkaApis.scala:57)
	at kafka.server.KafkaRequestHandler.run(KafkaRequestHandler.scala:41)
	at java.lang.Thread.run(Thread.java:662)
[2012-11-08 07:10:42,302] ERROR [KafkaApi-0] Error processing ProducerRequest on topic2:0 (kafka.server.KafkaApis:102)
kafka.common.UnknownTopicOrPartitionException: Topic topic2 partition 0 doesn't exist on 0
	at kafka.server.ReplicaManager.getLeaderReplicaIfLocal(ReplicaManager.scala:142)
	at kafka.server.KafkaApis$$anonfun$appendToLocalLog$2.apply(KafkaApis.scala:248)
	at kafka.server.KafkaApis$$anonfun$appendToLocalLog$2.apply(KafkaApis.scala:243)
	at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
	at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
	at scala.collection.immutable.Map$Map3.foreach(Map.scala:152)
	at scala.collection.TraversableLike$class.map(TraversableLike.scala:206)
	at scala.collection.immutable.Map$Map3.map(Map.scala:132)
	at kafka.server.KafkaApis.appendToLocalLog(KafkaApis.scala:243)
	at kafka.server.KafkaApis.handleProducerRequest(KafkaApis.scala:183)
	at kafka.server.KafkaApis.handle(KafkaApis.scala:57)
	at kafka.server.KafkaRequestHandler.run(KafkaRequestHandler.scala:41)
	at java.lang.Thread.run(Thread.java:662)
[2012-11-08 07:10:43,620] ERROR [KafkaApi-0] Error processing ProducerRequest on topic2:0 (kafka.server.KafkaApis:102)
kafka.common.UnknownTopicOrPartitionException: Topic topic2 partition 0 doesn't exist on 0
	at kafka.server.ReplicaManager.getLeaderReplicaIfLocal(ReplicaManager.scala:142)
	at kafka.server.KafkaApis$$anonfun$appendToLocalLog$2.apply(KafkaApis.scala:248)
	at kafka.server.KafkaApis$$anonfun$appendToLocalLog$2.apply(KafkaApis.scala:243)
	at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
	at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
	at scala.collection.immutable.Map$Map3.foreach(Map.scala:152)
	at scala.collection.TraversableLike$class.map(TraversableLike.scala:206)
	at scala.collection.immutable.Map$Map3.map(Map.scala:132)
	at kafka.server.KafkaApis.appendToLocalLog(KafkaApis.scala:243)
	at kafka.server.KafkaApis.handleProducerRequest(KafkaApis.scala:183)
	at kafka.server.KafkaApis.handle(KafkaApis.scala:57)
	at kafka.server.KafkaRequestHandler.run(KafkaRequestHandler.scala:41)
	at java.lang.Thread.run(Thread.java:662)
[info] Test Passed: testProduceCorrectlyReceivesResponse(kafka.producer.SyncProducerTest)
[info] Test Starting: testProducerCanTimeout(kafka.producer.SyncProducerTest)
[info] Test Passed: testProducerCanTimeout(kafka.producer.SyncProducerTest)
[info] == core-kafka / kafka.producer.SyncProducerTest ==
[info] 
[info] == core-kafka / kafka.message.MessageCompressionTest ==
[info] Test Starting: testSimpleCompressDecompress
[info] Test Passed: testSimpleCompressDecompress
[info] Test Starting: testComplexCompressDecompress
[info] Test Passed: testComplexCompressDecompress
[info] == core-kafka / kafka.message.MessageCompressionTest ==
[info] 
[info] == core-kafka / kafka.log.LogOffsetTest ==
[info] Test Starting: testGetOffsetsForUnknownTopic(kafka.log.LogOffsetTest)
[info] Test Passed: testGetOffsetsForUnknownTopic(kafka.log.LogOffsetTest)
[info] Test Starting: testGetOffsetsBeforeLatestTime(kafka.log.LogOffsetTest)
[info] Test Passed: testGetOffsetsBeforeLatestTime(kafka.log.LogOffsetTest)
[info] Test Starting: testEmptyLogsGetOffsets(kafka.log.LogOffsetTest)
[info] Test Passed: testEmptyLogsGetOffsets(kafka.log.LogOffsetTest)
[info] Test Starting: testGetOffsetsBeforeNow(kafka.log.LogOffsetTest)
[info] Test Passed: testGetOffsetsBeforeNow(kafka.log.LogOffsetTest)
[info] Test Starting: testGetOffsetsBeforeEarliestTime(kafka.log.LogOffsetTest)
[2012-11-08 07:10:51,384] WARN fsync-ing the write ahead log in SyncThread:0 took 1193ms which will adversely effect operation latency. See the ZooKeeper troubleshooting guide (org.apache.zookeeper.server.persistence.FileTxnLog:321)
[info] Test Passed: testGetOffsetsBeforeEarliestTime(kafka.log.LogOffsetTest)
[info] == core-kafka / kafka.log.LogOffsetTest ==
[info] 
[info] == core-kafka / kafka.integration.TopicMetadataTest ==
[info] Test Starting: testTopicMetadataRequest(kafka.integration.TopicMetadataTest)
[info] Test Passed: testTopicMetadataRequest(kafka.integration.TopicMetadataTest)
[info] Test Starting: testBasicTopicMetadata(kafka.integration.TopicMetadataTest)
[info] Test Passed: testBasicTopicMetadata(kafka.integration.TopicMetadataTest)
[info] Test Starting: testAutoCreateTopic(kafka.integration.TopicMetadataTest)
[info] Test Passed: testAutoCreateTopic(kafka.integration.TopicMetadataTest)
[info] == core-kafka / kafka.integration.TopicMetadataTest ==
[info] 
[info] == core-kafka / kafka.server.LeaderElectionTest ==
[info] Test Starting: testLeaderElectionAndEpoch(kafka.server.LeaderElectionTest)
[2012-11-08 07:10:55,097] ERROR Closing socket for /127.0.0.1 because of error (kafka.network.Processor:102)
java.io.IOException: Connection reset by peer
	at sun.nio.ch.FileDispatcher.write0(Native Method)
	at sun.nio.ch.SocketDispatcher.write(SocketDispatcher.java:29)
	at sun.nio.ch.IOUtil.writeFromNativeBuffer(IOUtil.java:69)
	at sun.nio.ch.IOUtil.write(IOUtil.java:40)
	at sun.nio.ch.SocketChannelImpl.write(SocketChannelImpl.java:334)
	at kafka.api.TopicDataSend.writeTo(FetchResponse.scala:131)
	at kafka.network.MultiSend.writeTo(Transmission.scala:94)
	at kafka.network.Send$class.writeCompletely(Transmission.scala:75)
	at kafka.network.MultiSend.writeCompletely(Transmission.scala:87)
	at kafka.api.FetchResponseSend.writeTo(FetchResponse.scala:232)
	at kafka.network.Processor.write(SocketServer.scala:318)
	at kafka.network.Processor.run(SocketServer.scala:211)
	at java.lang.Thread.run(Thread.java:662)
[2012-11-08 07:10:56,851] ERROR [BrokerChangeListener on Controller 0]: Error while handling broker changes (kafka.controller.ReplicaStateMachine$BrokerChangeListener:102)
org.I0Itec.zkclient.exception.ZkInterruptedException: java.lang.InterruptedException
	at org.I0Itec.zkclient.ZkClient.retryUntilConnected(ZkClient.java:687)
	at org.I0Itec.zkclient.ZkClient.readData(ZkClient.java:766)
	at org.I0Itec.zkclient.ZkClient.readData(ZkClient.java:761)
	at kafka.utils.ZkUtils$.readDataMaybeNull(ZkUtils.scala:374)
	at kafka.utils.ZkUtils$$anonfun$getPartitionAssignmentForTopics$1.apply(ZkUtils.scala:481)
	at kafka.utils.ZkUtils$$anonfun$getPartitionAssignmentForTopics$1.apply(ZkUtils.scala:480)
	at scala.collection.LinearSeqOptimized$class.foreach(LinearSeqOptimized.scala:61)
	at scala.collection.immutable.List.foreach(List.scala:45)
	at kafka.utils.ZkUtils$.getPartitionAssignmentForTopics(ZkUtils.scala:480)
	at kafka.utils.ZkUtils$.getPartitionsAssignedToBroker(ZkUtils.scala:521)
	at kafka.utils.ZkUtils$$anonfun$getAllReplicasOnBroker$1.apply(ZkUtils.scala:586)
	at kafka.utils.ZkUtils$$anonfun$getAllReplicasOnBroker$1.apply(ZkUtils.scala:583)
	at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
	at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
	at scala.collection.LinearSeqOptimized$class.foreach(LinearSeqOptimized.scala:61)
	at scala.collection.immutable.List.foreach(List.scala:45)
	at scala.collection.TraversableLike$class.map(TraversableLike.scala:206)
	at scala.collection.immutable.List.map(List.scala:45)
	at kafka.utils.ZkUtils$.getAllReplicasOnBroker(ZkUtils.scala:583)
	at kafka.controller.KafkaController.onBrokerStartup(KafkaController.scala:237)
	at kafka.controller.ReplicaStateMachine$BrokerChangeListener$$anonfun$handleChildChange$1.liftedTree1$1(ReplicaStateMachine.scala:250)
	at kafka.controller.ReplicaStateMachine$BrokerChangeListener$$anonfun$handleChildChange$1.apply$mcV$sp(ReplicaStateMachine.scala:239)
	at kafka.controller.ReplicaStateMachine$BrokerChangeListener$$anonfun$handleChildChange$1.apply(ReplicaStateMachine.scala:235)
	at kafka.controller.ReplicaStateMachine$BrokerChangeListener$$anonfun$handleChildChange$1.apply(ReplicaStateMachine.scala:235)
	at kafka.metrics.KafkaTimer.time(KafkaTimer.scala:33)
	at kafka.controller.ReplicaStateMachine$BrokerChangeListener.handleChildChange(ReplicaStateMachine.scala:235)
	at org.I0Itec.zkclient.ZkClient$7.run(ZkClient.java:568)
	at org.I0Itec.zkclient.ZkEventThread.run(ZkEventThread.java:71)
Caused by: java.lang.InterruptedException
	at java.lang.Object.wait(Native Method)
	at java.lang.Object.wait(Object.java:485)
	at org.apache.zookeeper.ClientCnxn.submitRequest(ClientCnxn.java:1344)
	at org.apache.zookeeper.ZooKeeper.getData(ZooKeeper.java:925)
	at org.apache.zookeeper.ZooKeeper.getData(ZooKeeper.java:956)
	at org.I0Itec.zkclient.ZkConnection.readData(ZkConnection.java:103)
	at org.I0Itec.zkclient.ZkClient$9.call(ZkClient.java:770)
	at org.I0Itec.zkclient.ZkClient$9.call(ZkClient.java:766)
	at org.I0Itec.zkclient.ZkClient.retryUntilConnected(ZkClient.java:675)
	... 27 more
[info] Test Passed: testLeaderElectionAndEpoch(kafka.server.LeaderElectionTest)
[info] == core-kafka / kafka.server.LeaderElectionTest ==
[info] 
[info] == core-kafka / kafka.log4j.KafkaLog4jAppenderTest ==
[info] Test Starting: testKafkaLog4jConfigs(kafka.log4j.KafkaLog4jAppenderTest)
log4j:WARN No appenders could be found for logger (org.I0Itec.zkclient.ZkEventThread).
log4j:WARN Please initialize the log4j system properly.
[info] Test Passed: testKafkaLog4jConfigs(kafka.log4j.KafkaLog4jAppenderTest)
[info] Test Starting: testLog4jAppends(kafka.log4j.KafkaLog4jAppenderTest)
[info] Test Passed: testLog4jAppends(kafka.log4j.KafkaLog4jAppenderTest)
[info] == core-kafka / kafka.log4j.KafkaLog4jAppenderTest ==
[info] 
[info] == core-kafka / kafka.server.SimpleFetchTest ==
[info] Test Starting: testNonReplicaSeesHwWhenFetching(kafka.server.SimpleFetchTest)
[info] Test Passed: testNonReplicaSeesHwWhenFetching(kafka.server.SimpleFetchTest)
[info] Test Starting: testReplicaSeesLeoWhenFetching(kafka.server.SimpleFetchTest)
[info] Test Passed: testReplicaSeesLeoWhenFetching(kafka.server.SimpleFetchTest)
[info] == core-kafka / kafka.server.SimpleFetchTest ==
[info] 
[info] == core-kafka / kafka.integration.AutoOffsetResetTest ==
[info] Test Starting: testResetToEarliestWhenOffsetTooHigh(kafka.integration.AutoOffsetResetTest)
[info] Test Passed: testResetToEarliestWhenOffsetTooHigh(kafka.integration.AutoOffsetResetTest)
[info] Test Starting: testResetToEarliestWhenOffsetTooLow(kafka.integration.AutoOffsetResetTest)
[info] Test Passed: testResetToEarliestWhenOffsetTooLow(kafka.integration.AutoOffsetResetTest)
[info] Test Starting: testResetToLatestWhenOffsetTooHigh(kafka.integration.AutoOffsetResetTest)
[info] Test Passed: testResetToLatestWhenOffsetTooHigh(kafka.integration.AutoOffsetResetTest)
[info] Test Starting: testResetToLatestWhenOffsetTooLow(kafka.integration.AutoOffsetResetTest)
[info] Test Passed: testResetToLatestWhenOffsetTooLow(kafka.integration.AutoOffsetResetTest)
[info] == core-kafka / kafka.integration.AutoOffsetResetTest ==
[info] 
[info] == core-kafka / kafka.server.IsrExpirationTest ==
[info] Test Starting: testIsrExpirationForStuckFollowers(kafka.server.IsrExpirationTest)
[info] Test Passed: testIsrExpirationForStuckFollowers(kafka.server.IsrExpirationTest)
[info] Test Starting: testIsrExpirationForSlowFollowers(kafka.server.IsrExpirationTest)
[info] Test Passed: testIsrExpirationForSlowFollowers(kafka.server.IsrExpirationTest)
[info] == core-kafka / kafka.server.IsrExpirationTest ==
[info] 
[info] == core-kafka / kafka.server.ServerShutdownTest ==
[info] Test Starting: testCleanShutdown(kafka.server.ServerShutdownTest)
[info] Test Passed: testCleanShutdown(kafka.server.ServerShutdownTest)
[info] == core-kafka / kafka.server.ServerShutdownTest ==
[info] 
[info] == core-kafka / kafka.server.LogRecoveryTest ==
[info] Test Starting: testHWCheckpointNoFailuresSingleLogSegment(kafka.server.LogRecoveryTest)
[info] Test Passed: testHWCheckpointNoFailuresSingleLogSegment(kafka.server.LogRecoveryTest)
[info] Test Starting: testHWCheckpointWithFailuresSingleLogSegment(kafka.server.LogRecoveryTest)
[error] Test Failed: testHWCheckpointWithFailuresSingleLogSegment(kafka.server.LogRecoveryTest)
java.lang.AssertionError: expected:<2> but was:<1>
	at org.junit.Assert.fail(Assert.java:69)
	at org.junit.Assert.failNotEquals(Assert.java:314)
	at org.junit.Assert.assertEquals(Assert.java:94)
	at org.junit.Assert.assertEquals(Assert.java:104)
	at kafka.server.LogRecoveryTest.testHWCheckpointWithFailuresSingleLogSegment(LogRecoveryTest.scala:149)
	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
	at java.lang.reflect.Method.invoke(Method.java:597)
	at junit.framework.TestCase.runTest(TestCase.java:164)
	at junit.framework.TestCase.runBare(TestCase.java:130)
	at junit.framework.TestResult$1.protect(TestResult.java:110)
	at junit.framework.TestResult.runProtected(TestResult.java:128)
	at junit.framework.TestResult.run(TestResult.java:113)
	at junit.framework.TestCase.run(TestCase.java:120)
	at junit.framework.TestSuite.runTest(TestSuite.java:228)
	at junit.framework.TestSuite.run(TestSuite.java:223)
	at junit.framework.TestSuite.runTest(TestSuite.java:228)
	at junit.framework.TestSuite.run(TestSuite.java:223)
	at org.scalatest.junit.JUnit3Suite.run(JUnit3Suite.scala:309)
	at org.scalatest.tools.ScalaTestFramework$ScalaTestRunner.run(ScalaTestFramework.scala:40)
	at sbt.TestRunner.run(TestFramework.scala:53)
	at sbt.TestRunner.runTest$1(TestFramework.scala:67)
	at sbt.TestRunner.run(TestFramework.scala:76)
	at sbt.TestFramework$$anonfun$10$$anonfun$apply$11.runTest$2(TestFramework.scala:194)
	at sbt.TestFramework$$anonfun$10$$anonfun$apply$11$$anonfun$apply$12.apply(TestFramework.scala:205)
	at sbt.TestFramework$$anonfun$10$$anonfun$apply$11$$anonfun$apply$12.apply(TestFramework.scala:205)
	at sbt.NamedTestTask.run(TestFramework.scala:92)
	at sbt.ScalaProject$$anonfun$sbt$ScalaProject$$toTask$1.apply(ScalaProject.scala:193)
	at sbt.ScalaProject$$anonfun$sbt$ScalaProject$$toTask$1.apply(ScalaProject.scala:193)
	at sbt.TaskManager$Task.invoke(TaskManager.scala:62)
	at sbt.impl.RunTask.doRun$1(RunTask.scala:77)
	at sbt.impl.RunTask.runTask(RunTask.scala:85)
	at sbt.impl.RunTask.sbt$impl$RunTask$$runIfNotRoot(RunTask.scala:60)
	at sbt.impl.RunTask$$anonfun$runTasksExceptRoot$2.apply(RunTask.scala:48)
	at sbt.impl.RunTask$$anonfun$runTasksExceptRoot$2.apply(RunTask.scala:48)
	at sbt.Distributor$Run$Worker$$anonfun$2.apply(ParallelRunner.scala:131)
	at sbt.Distributor$Run$Worker$$anonfun$2.apply(ParallelRunner.scala:131)
	at sbt.Control$.trapUnit(Control.scala:19)
	at sbt.Distributor$Run$Worker.run(ParallelRunner.scala:131)
[info] Test Starting: testHWCheckpointNoFailuresMultipleLogSegments(kafka.server.LogRecoveryTest)
[info] Test Passed: testHWCheckpointNoFailuresMultipleLogSegments(kafka.server.LogRecoveryTest)
[info] Test Starting: testHWCheckpointWithFailuresMultipleLogSegments(kafka.server.LogRecoveryTest)
[info] Test Passed: testHWCheckpointWithFailuresMultipleLogSegments(kafka.server.LogRecoveryTest)
[info] == core-kafka / kafka.server.LogRecoveryTest ==
[info] 
[info] == core-kafka / test-finish ==
[error] Failed: : Total 165, Failed 2, Errors 0, Passed 163, Skipped 0
[info] == core-kafka / test-finish ==
[info] 
[info] == core-kafka / Test cleanup 1 ==
[info] Deleting directory /tmp/sbt_268a9bb5
[info] == core-kafka / Test cleanup 1 ==
[info] 
[info] == core-kafka / test-cleanup ==
[info] == core-kafka / test-cleanup ==
[info] 
[info] == perf / copy-resources ==
[info] == perf / copy-resources ==
[error] Error running kafka.integration.LazyInitProducerTest: Test FAILED
[error] Error running kafka.server.LogRecoveryTest: Test FAILED
[error] Error running test: One or more subtasks failed
[info] 
[info] Total time: 257 s, completed Nov 8, 2012 7:11:33 AM
[info] 
[info] Total session time: 258 s, completed Nov 8, 2012 7:11:33 AM
[error] Error during build.
Build step 'Execute shell' marked build as failure

Build failed in Jenkins: Kafka-0.8 #101

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/Kafka-0.8/101/changes>

Changes:

[nehanarkhede] KAFKA-593 Empty log index file created when it shouldn't be empty; Patched by Yang Ye; reviewed by Jun and Jay

------------------------------------------
[...truncated 885 lines...]
[info] Test Starting: testBasic(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] Test Passed: testBasic(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] Test Starting: testCompression(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] Test Passed: testCompression(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] Test Starting: testCompressionSetConsumption(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] Test Passed: testCompressionSetConsumption(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] Test Starting: testConsumerDecoder(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] Test Passed: testConsumerDecoder(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] Test Starting: testLeaderSelectionForPartition(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] Test Passed: testLeaderSelectionForPartition(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] == core-kafka / kafka.consumer.ZookeeperConsumerConnectorTest ==
[info] 
[info] == core-kafka / kafka.server.ReplicaFetchTest ==
[info] Test Starting: testReplicaFetcherThread(kafka.server.ReplicaFetchTest)
[info] Test Passed: testReplicaFetcherThread(kafka.server.ReplicaFetchTest)
[info] == core-kafka / kafka.server.ReplicaFetchTest ==
[info] 
[info] == core-kafka / kafka.server.ServerShutdownTest ==
[info] Test Starting: testCleanShutdown(kafka.server.ServerShutdownTest)
[info] Test Passed: testCleanShutdown(kafka.server.ServerShutdownTest)
[info] == core-kafka / kafka.server.ServerShutdownTest ==
[info] 
[info] == core-kafka / kafka.integration.FetcherTest ==
[info] Test Starting: testFetcher(kafka.integration.FetcherTest)
[info] Test Passed: testFetcher(kafka.integration.FetcherTest)
[info] == core-kafka / kafka.integration.FetcherTest ==
[info] 
[info] == core-kafka / kafka.integration.TopicMetadataTest ==
[info] Test Starting: testTopicMetadataRequest(kafka.integration.TopicMetadataTest)
[info] Test Passed: testTopicMetadataRequest(kafka.integration.TopicMetadataTest)
[info] Test Starting: testBasicTopicMetadata(kafka.integration.TopicMetadataTest)
[info] Test Passed: testBasicTopicMetadata(kafka.integration.TopicMetadataTest)
[info] Test Starting: testAutoCreateTopic(kafka.integration.TopicMetadataTest)
[info] Test Passed: testAutoCreateTopic(kafka.integration.TopicMetadataTest)
[info] == core-kafka / kafka.integration.TopicMetadataTest ==
[info] 
[info] == core-kafka / kafka.producer.SyncProducerTest ==
[info] Test Starting: testReachableServer(kafka.producer.SyncProducerTest)
[info] Test Passed: testReachableServer(kafka.producer.SyncProducerTest)
[info] Test Starting: testEmptyProduceRequest(kafka.producer.SyncProducerTest)
[info] Test Passed: testEmptyProduceRequest(kafka.producer.SyncProducerTest)
[info] Test Starting: testMessageSizeTooLarge(kafka.producer.SyncProducerTest)
[info] Test Passed: testMessageSizeTooLarge(kafka.producer.SyncProducerTest)
[info] Test Starting: testProduceCorrectlyReceivesResponse(kafka.producer.SyncProducerTest)
[info] Test Passed: testProduceCorrectlyReceivesResponse(kafka.producer.SyncProducerTest)
[info] Test Starting: testProducerCanTimeout(kafka.producer.SyncProducerTest)
[info] Test Passed: testProducerCanTimeout(kafka.producer.SyncProducerTest)
[info] == core-kafka / kafka.producer.SyncProducerTest ==
[info] 
[info] == core-kafka / kafka.message.ByteBufferMessageSetTest ==
[info] Test Starting: testWrittenEqualsRead
[info] Test Passed: testWrittenEqualsRead
[info] Test Starting: testIteratorIsConsistent
[info] Test Passed: testIteratorIsConsistent
[info] Test Starting: testSizeInBytes
[info] Test Passed: testSizeInBytes
[info] Test Starting: testEquals
[info] Test Passed: testEquals
[info] Test Starting: testWriteTo
[info] Test Passed: testWriteTo
[info] Test Starting: testValidBytes
[info] Test Passed: testValidBytes
[info] Test Starting: testValidBytesWithCompression
[info] Test Passed: testValidBytesWithCompression
[info] Test Starting: testIterator
[info] Test Passed: testIterator
[info] Test Starting: testOffsetAssignment
[info] Test Passed: testOffsetAssignment
[info] == core-kafka / kafka.message.ByteBufferMessageSetTest ==
[info] 
[info] == core-kafka / kafka.log.FileMessageSetTest ==
[info] Test Starting: testWrittenEqualsRead
[info] Test Passed: testWrittenEqualsRead
[info] Test Starting: testIteratorIsConsistent
[info] Test Passed: testIteratorIsConsistent
[info] Test Starting: testSizeInBytes
[info] Test Passed: testSizeInBytes
[info] Test Starting: testWriteTo
[info] Test Passed: testWriteTo
[info] Test Starting: testFileSize
[info] Test Passed: testFileSize
[info] Test Starting: testIterationOverPartialAndTruncation
[info] Test Passed: testIterationOverPartialAndTruncation
[info] Test Starting: testIterationDoesntChangePosition
[info] Test Passed: testIterationDoesntChangePosition
[info] Test Starting: testRead
[info] Test Passed: testRead
[info] Test Starting: testSearch
[info] Test Passed: testSearch
[info] == core-kafka / kafka.log.FileMessageSetTest ==
[info] 
[info] == core-kafka / kafka.integration.AutoOffsetResetTest ==
[info] Test Starting: testResetToEarliestWhenOffsetTooHigh(kafka.integration.AutoOffsetResetTest)
[info] Test Passed: testResetToEarliestWhenOffsetTooHigh(kafka.integration.AutoOffsetResetTest)
[info] Test Starting: testResetToEarliestWhenOffsetTooLow(kafka.integration.AutoOffsetResetTest)
[info] Test Passed: testResetToEarliestWhenOffsetTooLow(kafka.integration.AutoOffsetResetTest)
[info] Test Starting: testResetToLatestWhenOffsetTooHigh(kafka.integration.AutoOffsetResetTest)
[info] Test Passed: testResetToLatestWhenOffsetTooHigh(kafka.integration.AutoOffsetResetTest)
[info] Test Starting: testResetToLatestWhenOffsetTooLow(kafka.integration.AutoOffsetResetTest)
[info] Test Passed: testResetToLatestWhenOffsetTooLow(kafka.integration.AutoOffsetResetTest)
[info] == core-kafka / kafka.integration.AutoOffsetResetTest ==
[info] 
[info] == core-kafka / kafka.server.LeaderElectionTest ==
[info] Test Starting: testLeaderElectionAndEpoch(kafka.server.LeaderElectionTest)
[info] Test Passed: testLeaderElectionAndEpoch(kafka.server.LeaderElectionTest)
[info] == core-kafka / kafka.server.LeaderElectionTest ==
[info] 
[info] == core-kafka / kafka.zk.ZKEphemeralTest ==
[info] Test Starting: testEphemeralNodeCleanup(kafka.zk.ZKEphemeralTest)
[info] Test Passed: testEphemeralNodeCleanup(kafka.zk.ZKEphemeralTest)
[info] == core-kafka / kafka.zk.ZKEphemeralTest ==
[info] 
[info] == core-kafka / kafka.integration.LazyInitProducerTest ==
[info] Test Starting: testProduceAndMultiFetch(kafka.integration.LazyInitProducerTest)
[info] Test Passed: testProduceAndMultiFetch(kafka.integration.LazyInitProducerTest)
[info] Test Starting: testMultiProduce(kafka.integration.LazyInitProducerTest)
[info] Test Passed: testMultiProduce(kafka.integration.LazyInitProducerTest)
[info] Test Starting: testProduceAndFetch(kafka.integration.LazyInitProducerTest)
[info] Test Passed: testProduceAndFetch(kafka.integration.LazyInitProducerTest)
[info] Test Starting: testMultiProduceResend(kafka.integration.LazyInitProducerTest)
[info] Test Passed: testMultiProduceResend(kafka.integration.LazyInitProducerTest)
[info] == core-kafka / kafka.integration.LazyInitProducerTest ==
[info] 
[info] == core-kafka / kafka.admin.AdminTest ==
[info] Test Starting: testReplicaAssignment(kafka.admin.AdminTest)
[info] Test Passed: testReplicaAssignment(kafka.admin.AdminTest)
[info] Test Starting: testManualReplicaAssignment(kafka.admin.AdminTest)
[info] Test Passed: testManualReplicaAssignment(kafka.admin.AdminTest)
[info] Test Starting: testTopicCreationInZK(kafka.admin.AdminTest)
[info] Test Passed: testTopicCreationInZK(kafka.admin.AdminTest)
[info] Test Starting: testGetTopicMetadata(kafka.admin.AdminTest)
[info] Test Passed: testGetTopicMetadata(kafka.admin.AdminTest)
[info] Test Starting: testPartitionReassignmentWithLeaderInNewReplicas(kafka.admin.AdminTest)
[info] Test Passed: testPartitionReassignmentWithLeaderInNewReplicas(kafka.admin.AdminTest)
[info] Test Starting: testPartitionReassignmentWithLeaderNotInNewReplicas(kafka.admin.AdminTest)
[info] Test Passed: testPartitionReassignmentWithLeaderNotInNewReplicas(kafka.admin.AdminTest)
[info] Test Starting: testPartitionReassignmentNonOverlappingReplicas(kafka.admin.AdminTest)
[info] Test Passed: testPartitionReassignmentNonOverlappingReplicas(kafka.admin.AdminTest)
[info] Test Starting: testReassigningNonExistingPartition(kafka.admin.AdminTest)
[info] Test Passed: testReassigningNonExistingPartition(kafka.admin.AdminTest)
[info] Test Starting: testResumePartitionReassignmentThatWasCompleted(kafka.admin.AdminTest)
[info] Test Passed: testResumePartitionReassignmentThatWasCompleted(kafka.admin.AdminTest)
[info] Test Starting: testPreferredReplicaJsonData(kafka.admin.AdminTest)
[info] Test Passed: testPreferredReplicaJsonData(kafka.admin.AdminTest)
[info] Test Starting: testBasicPreferredReplicaElection(kafka.admin.AdminTest)
[info] Test Passed: testBasicPreferredReplicaElection(kafka.admin.AdminTest)
[info] Test Starting: testShutdownBroker(kafka.admin.AdminTest)
[info] Test Passed: testShutdownBroker(kafka.admin.AdminTest)
[info] == core-kafka / kafka.admin.AdminTest ==
[info] 
[info] == core-kafka / kafka.log.LogTest ==
[info] Test Starting: testTimeBasedLogRoll
[error] Test Failed: testTimeBasedLogRoll
junit.framework.AssertionFailedError: There should be exactly one segment. expected:<1> but was:<2>
	at junit.framework.Assert.fail(Assert.java:47)
	at junit.framework.Assert.failNotEquals(Assert.java:277)
	at junit.framework.Assert.assertEquals(Assert.java:64)
	at junit.framework.Assert.assertEquals(Assert.java:195)
	at kafka.log.LogTest.testTimeBasedLogRoll(LogTest.scala:73)
	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
	at java.lang.reflect.Method.invoke(Method.java:597)
	at org.junit.internal.runners.TestMethodRunner.executeMethodBody(TestMethodRunner.java:99)
	at org.junit.internal.runners.TestMethodRunner.runUnprotected(TestMethodRunner.java:81)
	at org.junit.internal.runners.BeforeAndAfterRunner.runProtected(BeforeAndAfterRunner.java:34)
	at org.junit.internal.runners.TestMethodRunner.runMethod(TestMethodRunner.java:75)
	at org.junit.internal.runners.TestMethodRunner.run(TestMethodRunner.java:45)
	at org.junit.internal.runners.TestClassMethodsRunner.invokeTestMethod(TestClassMethodsRunner.java:71)
	at org.junit.internal.runners.TestClassMethodsRunner.run(TestClassMethodsRunner.java:35)
	at org.junit.internal.runners.TestClassRunner$1.runUnprotected(TestClassRunner.java:42)
	at org.junit.internal.runners.BeforeAndAfterRunner.runProtected(BeforeAndAfterRunner.java:34)
	at org.junit.internal.runners.TestClassRunner.run(TestClassRunner.java:52)
	at org.junit.internal.runners.CompositeRunner.run(CompositeRunner.java:29)
	at org.junit.runner.JUnitCore.run(JUnitCore.java:121)
	at org.junit.runner.JUnitCore.run(JUnitCore.java:100)
	at org.junit.runner.JUnitCore.run(JUnitCore.java:91)
	at org.scalatest.junit.JUnitSuite$class.run(JUnitSuite.scala:261)
	at kafka.log.LogTest.run(LogTest.scala:32)
	at org.scalatest.tools.ScalaTestFramework$ScalaTestRunner.run(ScalaTestFramework.scala:40)
	at sbt.TestRunner.run(TestFramework.scala:53)
	at sbt.TestRunner.runTest$1(TestFramework.scala:67)
	at sbt.TestRunner.run(TestFramework.scala:76)
	at sbt.TestFramework$$anonfun$10$$anonfun$apply$11.runTest$2(TestFramework.scala:194)
	at sbt.TestFramework$$anonfun$10$$anonfun$apply$11$$anonfun$apply$12.apply(TestFramework.scala:205)
	at sbt.TestFramework$$anonfun$10$$anonfun$apply$11$$anonfun$apply$12.apply(TestFramework.scala:205)
	at sbt.NamedTestTask.run(TestFramework.scala:92)
	at sbt.ScalaProject$$anonfun$sbt$ScalaProject$$toTask$1.apply(ScalaProject.scala:193)
	at sbt.ScalaProject$$anonfun$sbt$ScalaProject$$toTask$1.apply(ScalaProject.scala:193)
	at sbt.TaskManager$Task.invoke(TaskManager.scala:62)
	at sbt.impl.RunTask.doRun$1(RunTask.scala:77)
	at sbt.impl.RunTask.runTask(RunTask.scala:85)
	at sbt.impl.RunTask.sbt$impl$RunTask$$runIfNotRoot(RunTask.scala:60)
	at sbt.impl.RunTask$$anonfun$runTasksExceptRoot$2.apply(RunTask.scala:48)
	at sbt.impl.RunTask$$anonfun$runTasksExceptRoot$2.apply(RunTask.scala:48)
	at sbt.Distributor$Run$Worker$$anonfun$2.apply(ParallelRunner.scala:131)
	at sbt.Distributor$Run$Worker$$anonfun$2.apply(ParallelRunner.scala:131)
	at sbt.Control$.trapUnit(Control.scala:19)
	at sbt.Distributor$Run$Worker.run(ParallelRunner.scala:131)
[info] Test Starting: testSizeBasedLogRoll
[info] Test Passed: testSizeBasedLogRoll
[info] Test Starting: testLoadEmptyLog
[info] Test Passed: testLoadEmptyLog
[info] Test Starting: testAppendAndRead
[info] Test Passed: testAppendAndRead
[info] Test Starting: testReadOutOfRange
[info] Test Passed: testReadOutOfRange
[info] Test Starting: testLogRolls
[info] Test Passed: testLogRolls
[info] Test Starting: testCompressedMessages
[info] Test Passed: testCompressedMessages
[info] Test Starting: testFindSegment
[info] Test Passed: testFindSegment
[info] Test Starting: testEdgeLogRollsStartingAtZero
[info] Test Passed: testEdgeLogRollsStartingAtZero
[info] Test Starting: testEdgeLogRollsStartingAtNonZero
[info] Test Passed: testEdgeLogRollsStartingAtNonZero
[info] Test Starting: testMessageSizeCheck
[info] Test Passed: testMessageSizeCheck
[info] Test Starting: testLogRecoversToCorrectOffset
[info] Test Passed: testLogRecoversToCorrectOffset
[info] Test Starting: testTruncateTo
[info] Test Passed: testTruncateTo
[info] Test Starting: testIndexResizingAtTruncation
[info] Test Passed: testIndexResizingAtTruncation
[info] Test Starting: testAppendWithoutOffsetAssignment
[info] Test Passed: testAppendWithoutOffsetAssignment
[info] Test Starting: testReopenThenTruncate
[info] Test Passed: testReopenThenTruncate
[info] == core-kafka / kafka.log.LogTest ==
[info] 
[info] == core-kafka / test-finish ==
[error] Failed: : Total 164, Failed 1, Errors 0, Passed 163, Skipped 0
[info] == core-kafka / test-finish ==
[info] 
[info] == core-kafka / Test cleanup 1 ==
[info] Deleting directory /tmp/sbt_6aafdbf3
[info] == core-kafka / Test cleanup 1 ==
[info] 
[info] == core-kafka / test-cleanup ==
[info] == core-kafka / test-cleanup ==
[error] Error running kafka.log.LogTest: Test FAILED
[error] Error running test: One or more subtasks failed
[info] 
[info] Total time: 178 s, completed Nov 6, 2012 4:08:36 AM
[info] 
[info] Total session time: 178 s, completed Nov 6, 2012 4:08:36 AM
[error] Error during build.
Build step 'Execute shell' marked build as failure

Build failed in Jenkins: Kafka-0.8 #100

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/Kafka-0.8/100/changes>

Changes:

[junrao] extend DumpLogSegments to verify consistency btw data and index; patched by Yang Ye; reviewed by Jun Rao; KAFKA-577

[junrao] LogSegment.firstAppendTime not reset after truncate to; patched by Swapnil Ghike; reviewed by Jun Rao, Neha Narkhede; KAFKA-596

------------------------------------------
[...truncated 1613 lines...]
[2012-11-06 00:58:30,404] WARN Session 0x13ad339c2c40001 for server null, unexpected error, closing socket connection and attempting reconnect (org.apache.zookeeper.ClientCnxn:1188)
java.net.ConnectException: Connection refused
	at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
	at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:567)
	at org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1146)
[2012-11-06 00:58:30,540] WARN Session 0x0 for server null, unexpected error, closing socket connection and attempting reconnect (org.apache.zookeeper.ClientCnxn:1188)
java.net.ConnectException: Connection refused
	at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
	at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:567)
	at org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1146)
[2012-11-06 00:58:30,748] WARN Session 0x13ad339c2c40004 for server null, unexpected error, closing socket connection and attempting reconnect (org.apache.zookeeper.ClientCnxn:1188)
java.net.ConnectException: Connection refused
	at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
	at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:567)
	at org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1146)
[2012-11-06 00:58:30,972] WARN Session 0x13ad339c2c40003 for server null, unexpected error, closing socket connection and attempting reconnect (org.apache.zookeeper.ClientCnxn:1188)
java.net.ConnectException: Connection refused
	at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
	at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:567)
	at org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1146)
[info] Test Passed: testLogRecoversToCorrectOffset
[info] Test Starting: testTruncateTo
[2012-11-06 00:58:31,210] ERROR [Kafka Log on Broker 0], Target offset 11 cannot be greater than the last message offset 10 in the log /tmp/kafka-342189/00000000000000000000.log (kafka.log.Log:96)
[info] Test Passed: testTruncateTo
[info] Test Starting: testAppendWithoutOffsetAssignment
[info] Test Passed: testAppendWithoutOffsetAssignment
[info] Test Starting: testReopenThenTruncate
[2012-11-06 00:58:31,318] WARN Session 0x0 for server null, unexpected error, closing socket connection and attempting reconnect (org.apache.zookeeper.ClientCnxn:1188)
java.net.ConnectException: Connection refused
	at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
	at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:567)
	at org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1146)
[info] Test Passed: testReopenThenTruncate
[info] == core-kafka / kafka.log.LogTest ==
[info] 
[info] == core-kafka / kafka.log4j.KafkaLog4jAppenderTest ==
[info] Test Starting: testKafkaLog4jConfigs(kafka.log4j.KafkaLog4jAppenderTest)
log4j:WARN No appenders could be found for logger (org.I0Itec.zkclient.ZkEventThread).
log4j:WARN Please initialize the log4j system properly.
[info] Test Passed: testKafkaLog4jConfigs(kafka.log4j.KafkaLog4jAppenderTest)
[info] Test Starting: testLog4jAppends(kafka.log4j.KafkaLog4jAppenderTest)
[info] Test Passed: testLog4jAppends(kafka.log4j.KafkaLog4jAppenderTest)
[info] == core-kafka / kafka.log4j.KafkaLog4jAppenderTest ==
[info] 
[info] == core-kafka / kafka.message.MessageTest ==
[info] Test Starting: testFieldValues
[info] Test Passed: testFieldValues
[info] Test Starting: testChecksum
[info] Test Passed: testChecksum
[info] Test Starting: testEquality
[info] Test Passed: testEquality
[info] Test Starting: testIsHashable
[info] Test Passed: testIsHashable
[info] == core-kafka / kafka.message.MessageTest ==
[info] 
[info] == core-kafka / kafka.javaapi.message.ByteBufferMessageSetTest ==
[info] Test Starting: testWrittenEqualsRead
[info] Test Passed: testWrittenEqualsRead
[info] Test Starting: testIteratorIsConsistent
[info] Test Passed: testIteratorIsConsistent
[info] Test Starting: testSizeInBytes
[info] Test Passed: testSizeInBytes
[info] Test Starting: testEquals
[info] Test Passed: testEquals
[info] Test Starting: testIteratorIsConsistentWithCompression
[info] Test Passed: testIteratorIsConsistentWithCompression
[info] Test Starting: testSizeInBytesWithCompression
[info] Test Passed: testSizeInBytesWithCompression
[info] Test Starting: testEqualsWithCompression
[info] Test Passed: testEqualsWithCompression
[info] == core-kafka / kafka.javaapi.message.ByteBufferMessageSetTest ==
[info] 
[info] == core-kafka / kafka.api.RequestResponseSerializationTest ==
[info] Test Starting: testSerializationAndDeserialization
[info] Test Passed: testSerializationAndDeserialization
[info] == core-kafka / kafka.api.RequestResponseSerializationTest ==
[info] 
[info] == core-kafka / kafka.log.OffsetIndexTest ==
[info] Test Starting: truncate
[info] Test Passed: truncate
[info] Test Starting: randomLookupTest
[info] Test Passed: randomLookupTest
[info] Test Starting: lookupExtremeCases
[info] Test Passed: lookupExtremeCases
[info] Test Starting: appendTooMany
[info] Test Passed: appendTooMany
[info] Test Starting: appendOutOfOrder
[info] Test Passed: appendOutOfOrder
[info] Test Starting: testReopen
[info] Test Passed: testReopen
[info] == core-kafka / kafka.log.OffsetIndexTest ==
[info] 
[info] == core-kafka / kafka.network.SocketServerTest ==
[info] Test Starting: simpleRequest
[info] Test Passed: simpleRequest
[info] Test Starting: tooBigRequestIsRejected
[info] Test Passed: tooBigRequestIsRejected
[info] == core-kafka / kafka.network.SocketServerTest ==
[info] 
[info] == core-kafka / kafka.producer.ProducerTest ==
[info] Test Starting: testUpdateBrokerPartitionInfo(kafka.producer.ProducerTest)
[info] Test Passed: testUpdateBrokerPartitionInfo(kafka.producer.ProducerTest)
[info] Test Starting: testSendToNewTopic(kafka.producer.ProducerTest)
[info] Test Passed: testSendToNewTopic(kafka.producer.ProducerTest)
[info] Test Starting: testSendWithDeadBroker(kafka.producer.ProducerTest)
[info] Test Passed: testSendWithDeadBroker(kafka.producer.ProducerTest)
[info] Test Starting: testAsyncSendCanCorrectlyFailWithTimeout(kafka.producer.ProducerTest)
[info] Test Passed: testAsyncSendCanCorrectlyFailWithTimeout(kafka.producer.ProducerTest)
[info] == core-kafka / kafka.producer.ProducerTest ==
[info] 
[info] == core-kafka / kafka.javaapi.consumer.ZookeeperConsumerConnectorTest ==
[info] Test Starting: testBasic(kafka.javaapi.consumer.ZookeeperConsumerConnectorTest)
[info] Test Passed: testBasic(kafka.javaapi.consumer.ZookeeperConsumerConnectorTest)
[info] == core-kafka / kafka.javaapi.consumer.ZookeeperConsumerConnectorTest ==
[info] 
[info] == core-kafka / kafka.producer.SyncProducerTest ==
[info] Test Starting: testReachableServer(kafka.producer.SyncProducerTest)
[info] Test Passed: testReachableServer(kafka.producer.SyncProducerTest)
[info] Test Starting: testEmptyProduceRequest(kafka.producer.SyncProducerTest)
[info] Test Passed: testEmptyProduceRequest(kafka.producer.SyncProducerTest)
[info] Test Starting: testMessageSizeTooLarge(kafka.producer.SyncProducerTest)
[info] Test Passed: testMessageSizeTooLarge(kafka.producer.SyncProducerTest)
[info] Test Starting: testProduceCorrectlyReceivesResponse(kafka.producer.SyncProducerTest)
[info] Test Passed: testProduceCorrectlyReceivesResponse(kafka.producer.SyncProducerTest)
[info] Test Starting: testProducerCanTimeout(kafka.producer.SyncProducerTest)
[info] Test Passed: testProducerCanTimeout(kafka.producer.SyncProducerTest)
[info] == core-kafka / kafka.producer.SyncProducerTest ==
[info] 
[info] == core-kafka / kafka.server.ReplicaFetchTest ==
[info] Test Starting: testReplicaFetcherThread(kafka.server.ReplicaFetchTest)
[info] Test Passed: testReplicaFetcherThread(kafka.server.ReplicaFetchTest)
[info] == core-kafka / kafka.server.ReplicaFetchTest ==
[info] 
[info] == core-kafka / kafka.consumer.ZookeeperConsumerConnectorTest ==
[info] Test Starting: testBasic(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] Test Passed: testBasic(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] Test Starting: testCompression(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] Test Passed: testCompression(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] Test Starting: testCompressionSetConsumption(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] Test Passed: testCompressionSetConsumption(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] Test Starting: testConsumerDecoder(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] Test Passed: testConsumerDecoder(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] Test Starting: testLeaderSelectionForPartition(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] Test Passed: testLeaderSelectionForPartition(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] == core-kafka / kafka.consumer.ZookeeperConsumerConnectorTest ==
[info] 
[info] == core-kafka / kafka.server.RequestPurgatoryTest ==
[info] Test Starting: testRequestSatisfaction(kafka.server.RequestPurgatoryTest)
[info] Test Passed: testRequestSatisfaction(kafka.server.RequestPurgatoryTest)
[info] Test Starting: testRequestExpiry(kafka.server.RequestPurgatoryTest)
[info] Test Passed: testRequestExpiry(kafka.server.RequestPurgatoryTest)
[info] == core-kafka / kafka.server.RequestPurgatoryTest ==
[info] 
[info] == core-kafka / kafka.log.FileMessageSetTest ==
[info] Test Starting: testWrittenEqualsRead
[info] Test Passed: testWrittenEqualsRead
[info] Test Starting: testIteratorIsConsistent
[info] Test Passed: testIteratorIsConsistent
[info] Test Starting: testSizeInBytes
[info] Test Passed: testSizeInBytes
[info] Test Starting: testWriteTo
[info] Test Passed: testWriteTo
[info] Test Starting: testFileSize
[info] Test Passed: testFileSize
[info] Test Starting: testIterationOverPartialAndTruncation
[info] Test Passed: testIterationOverPartialAndTruncation
[info] Test Starting: testIterationDoesntChangePosition
[info] Test Passed: testIterationDoesntChangePosition
[info] Test Starting: testRead
[info] Test Passed: testRead
[info] Test Starting: testSearch
[info] Test Passed: testSearch
[info] == core-kafka / kafka.log.FileMessageSetTest ==
[info] 
[info] == core-kafka / kafka.log.LogOffsetTest ==
[info] Test Starting: testGetOffsetsForUnknownTopic(kafka.log.LogOffsetTest)
[info] Test Passed: testGetOffsetsForUnknownTopic(kafka.log.LogOffsetTest)
[info] Test Starting: testGetOffsetsBeforeLatestTime(kafka.log.LogOffsetTest)
[info] Test Passed: testGetOffsetsBeforeLatestTime(kafka.log.LogOffsetTest)
[info] Test Starting: testEmptyLogsGetOffsets(kafka.log.LogOffsetTest)
[info] Test Passed: testEmptyLogsGetOffsets(kafka.log.LogOffsetTest)
[info] Test Starting: testGetOffsetsBeforeNow(kafka.log.LogOffsetTest)
[info] Test Passed: testGetOffsetsBeforeNow(kafka.log.LogOffsetTest)
[info] Test Starting: testGetOffsetsBeforeEarliestTime(kafka.log.LogOffsetTest)
[info] Test Passed: testGetOffsetsBeforeEarliestTime(kafka.log.LogOffsetTest)
[info] == core-kafka / kafka.log.LogOffsetTest ==
[info] 
[info] == core-kafka / kafka.utils.UtilsTest ==
[info] Test Starting: testSwallow
[info] Test Passed: testSwallow
[info] Test Starting: testCircularIterator
[info] Test Passed: testCircularIterator
[info] == core-kafka / kafka.utils.UtilsTest ==
[info] 
[info] == core-kafka / kafka.integration.TopicMetadataTest ==
[info] Test Starting: testTopicMetadataRequest(kafka.integration.TopicMetadataTest)
[info] Test Passed: testTopicMetadataRequest(kafka.integration.TopicMetadataTest)
[info] Test Starting: testBasicTopicMetadata(kafka.integration.TopicMetadataTest)
[info] Test Passed: testBasicTopicMetadata(kafka.integration.TopicMetadataTest)
[info] Test Starting: testAutoCreateTopic(kafka.integration.TopicMetadataTest)
[info] Test Passed: testAutoCreateTopic(kafka.integration.TopicMetadataTest)
[info] == core-kafka / kafka.integration.TopicMetadataTest ==
[info] 
[info] == core-kafka / kafka.admin.AdminTest ==
[info] Test Starting: testReplicaAssignment(kafka.admin.AdminTest)
[info] Test Passed: testReplicaAssignment(kafka.admin.AdminTest)
[info] Test Starting: testManualReplicaAssignment(kafka.admin.AdminTest)
[info] Test Passed: testManualReplicaAssignment(kafka.admin.AdminTest)
[info] Test Starting: testTopicCreationInZK(kafka.admin.AdminTest)
[info] Test Passed: testTopicCreationInZK(kafka.admin.AdminTest)
[info] Test Starting: testGetTopicMetadata(kafka.admin.AdminTest)
[info] Test Passed: testGetTopicMetadata(kafka.admin.AdminTest)
[info] Test Starting: testPartitionReassignmentWithLeaderInNewReplicas(kafka.admin.AdminTest)
[info] Test Passed: testPartitionReassignmentWithLeaderInNewReplicas(kafka.admin.AdminTest)
[info] Test Starting: testPartitionReassignmentWithLeaderNotInNewReplicas(kafka.admin.AdminTest)
[info] Test Passed: testPartitionReassignmentWithLeaderNotInNewReplicas(kafka.admin.AdminTest)
[info] Test Starting: testPartitionReassignmentNonOverlappingReplicas(kafka.admin.AdminTest)
[info] Test Passed: testPartitionReassignmentNonOverlappingReplicas(kafka.admin.AdminTest)
[info] Test Starting: testReassigningNonExistingPartition(kafka.admin.AdminTest)
[info] Test Passed: testReassigningNonExistingPartition(kafka.admin.AdminTest)
[info] Test Starting: testResumePartitionReassignmentThatWasCompleted(kafka.admin.AdminTest)
[info] Test Passed: testResumePartitionReassignmentThatWasCompleted(kafka.admin.AdminTest)
[info] Test Starting: testPreferredReplicaJsonData(kafka.admin.AdminTest)
[info] Test Passed: testPreferredReplicaJsonData(kafka.admin.AdminTest)
[info] Test Starting: testBasicPreferredReplicaElection(kafka.admin.AdminTest)
[info] Test Passed: testBasicPreferredReplicaElection(kafka.admin.AdminTest)
[info] Test Starting: testShutdownBroker(kafka.admin.AdminTest)
[info] Test Passed: testShutdownBroker(kafka.admin.AdminTest)
[info] == core-kafka / kafka.admin.AdminTest ==
[info] 
[info] == core-kafka / test-finish ==
[error] Failed: : Total 163, Failed 2, Errors 0, Passed 161, Skipped 0
[info] == core-kafka / test-finish ==
[info] 
[info] == core-kafka / Test cleanup 1 ==
[info] Deleting directory /tmp/sbt_bff857f0
[info] == core-kafka / Test cleanup 1 ==
[info] 
[info] == core-kafka / test-cleanup ==
[info] == core-kafka / test-cleanup ==
[error] Error running kafka.integration.PrimitiveApiTest: Test FAILED
[error] Error running kafka.integration.LazyInitProducerTest: Test FAILED
[error] Error running test: One or more subtasks failed
[info] 
[info] Total time: 244 s, completed Nov 6, 2012 12:59:53 AM
[info] 
[info] Total session time: 244 s, completed Nov 6, 2012 12:59:53 AM
[error] Error during build.
Build step 'Execute shell' marked build as failure

Build failed in Jenkins: Kafka-0.8 #99

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/Kafka-0.8/99/changes>

Changes:

[jkreps] KAFKA-188 Support multiple data directories. 
Patch reviewed by Neha and Jun.

------------------------------------------
[...truncated 2248 lines...]
[info] Test Passed: testKafkaLog4jConfigs(kafka.log4j.KafkaLog4jAppenderTest)
[info] Test Starting: testLog4jAppends(kafka.log4j.KafkaLog4jAppenderTest)
[info] Test Passed: testLog4jAppends(kafka.log4j.KafkaLog4jAppenderTest)
[info] == core-kafka / kafka.log4j.KafkaLog4jAppenderTest ==
[info] 
[info] == core-kafka / kafka.network.SocketServerTest ==
[info] Test Starting: simpleRequest
[info] Test Passed: simpleRequest
[info] Test Starting: tooBigRequestIsRejected
[info] Test Passed: tooBigRequestIsRejected
[info] == core-kafka / kafka.network.SocketServerTest ==
[info] 
[info] == core-kafka / kafka.javaapi.message.ByteBufferMessageSetTest ==
[info] Test Starting: testWrittenEqualsRead
[info] Test Passed: testWrittenEqualsRead
[info] Test Starting: testIteratorIsConsistent
[info] Test Passed: testIteratorIsConsistent
[info] Test Starting: testIteratorIsConsistentWithCompression
[info] Test Passed: testIteratorIsConsistentWithCompression
[info] Test Starting: testSizeInBytes
[info] Test Passed: testSizeInBytes
[info] Test Starting: testSizeInBytesWithCompression
[info] Test Passed: testSizeInBytesWithCompression
[info] Test Starting: testEquals
[info] Test Passed: testEquals
[info] Test Starting: testEqualsWithCompression
[info] Test Passed: testEqualsWithCompression
[info] == core-kafka / kafka.javaapi.message.ByteBufferMessageSetTest ==
[info] 
[info] == core-kafka / kafka.server.SimpleFetchTest ==
[info] Test Starting: testNonReplicaSeesHwWhenFetching(kafka.server.SimpleFetchTest)
[info] Test Passed: testNonReplicaSeesHwWhenFetching(kafka.server.SimpleFetchTest)
[info] Test Starting: testReplicaSeesLeoWhenFetching(kafka.server.SimpleFetchTest)
[info] Test Passed: testReplicaSeesLeoWhenFetching(kafka.server.SimpleFetchTest)
[info] == core-kafka / kafka.server.SimpleFetchTest ==
[info] 
[info] == core-kafka / kafka.integration.PrimitiveApiTest ==
[info] Test Starting: testProduceAndMultiFetch(kafka.integration.PrimitiveApiTest)
[info] Test Passed: testProduceAndMultiFetch(kafka.integration.PrimitiveApiTest)
[info] Test Starting: testMultiProduce(kafka.integration.PrimitiveApiTest)
[info] Test Passed: testMultiProduce(kafka.integration.PrimitiveApiTest)
[info] Test Starting: testFetchRequestCanProperlySerialize(kafka.integration.PrimitiveApiTest)
[info] Test Passed: testFetchRequestCanProperlySerialize(kafka.integration.PrimitiveApiTest)
[info] Test Starting: testEmptyFetchRequest(kafka.integration.PrimitiveApiTest)
[info] Test Passed: testEmptyFetchRequest(kafka.integration.PrimitiveApiTest)
[info] Test Starting: testDefaultEncoderProducerAndFetch(kafka.integration.PrimitiveApiTest)
[info] Test Passed: testDefaultEncoderProducerAndFetch(kafka.integration.PrimitiveApiTest)
[info] Test Starting: testDefaultEncoderProducerAndFetchWithCompression(kafka.integration.PrimitiveApiTest)
[info] Test Passed: testDefaultEncoderProducerAndFetchWithCompression(kafka.integration.PrimitiveApiTest)
[info] Test Starting: testProduceAndMultiFetchWithCompression(kafka.integration.PrimitiveApiTest)
[info] Test Passed: testProduceAndMultiFetchWithCompression(kafka.integration.PrimitiveApiTest)
[info] Test Starting: testMultiProduceWithCompression(kafka.integration.PrimitiveApiTest)
[error] Test Failed: testMultiProduceWithCompression(kafka.integration.PrimitiveApiTest)
junit.framework.AssertionFailedError: Iterators have uneven length-- second has more: 4 > 2
	at junit.framework.Assert.fail(Assert.java:47)
	at junit.framework.Assert.assertTrue(Assert.java:20)
	at junit.framework.Assert.assertFalse(Assert.java:34)
	at kafka.utils.TestUtils$.checkEquals(TestUtils.scala:220)
	at kafka.integration.PrimitiveApiTest$$anonfun$testMultiProduceWithCompression$2.apply(PrimitiveApiTest.scala:311)
	at kafka.integration.PrimitiveApiTest$$anonfun$testMultiProduceWithCompression$2.apply(PrimitiveApiTest.scala:309)
	at scala.collection.LinearSeqOptimized$class.foreach(LinearSeqOptimized.scala:61)
	at scala.collection.immutable.List.foreach(List.scala:45)
	at kafka.integration.PrimitiveApiTest.testMultiProduceWithCompression(PrimitiveApiTest.scala:309)
	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
	at java.lang.reflect.Method.invoke(Method.java:597)
	at junit.framework.TestCase.runTest(TestCase.java:164)
	at junit.framework.TestCase.runBare(TestCase.java:130)
	at junit.framework.TestResult$1.protect(TestResult.java:110)
	at junit.framework.TestResult.runProtected(TestResult.java:128)
	at junit.framework.TestResult.run(TestResult.java:113)
	at junit.framework.TestCase.run(TestCase.java:120)
	at junit.framework.TestSuite.runTest(TestSuite.java:228)
	at junit.framework.TestSuite.run(TestSuite.java:223)
	at junit.framework.TestSuite.runTest(TestSuite.java:228)
	at junit.framework.TestSuite.run(TestSuite.java:223)
	at org.scalatest.junit.JUnit3Suite.run(JUnit3Suite.scala:309)
	at org.scalatest.tools.ScalaTestFramework$ScalaTestRunner.run(ScalaTestFramework.scala:40)
	at sbt.TestRunner.run(TestFramework.scala:53)
	at sbt.TestRunner.runTest$1(TestFramework.scala:67)
	at sbt.TestRunner.run(TestFramework.scala:76)
	at sbt.TestFramework$$anonfun$10$$anonfun$apply$11.runTest$2(TestFramework.scala:194)
	at sbt.TestFramework$$anonfun$10$$anonfun$apply$11$$anonfun$apply$12.apply(TestFramework.scala:205)
	at sbt.TestFramework$$anonfun$10$$anonfun$apply$11$$anonfun$apply$12.apply(TestFramework.scala:205)
	at sbt.NamedTestTask.run(TestFramework.scala:92)
	at sbt.ScalaProject$$anonfun$sbt$ScalaProject$$toTask$1.apply(ScalaProject.scala:193)
	at sbt.ScalaProject$$anonfun$sbt$ScalaProject$$toTask$1.apply(ScalaProject.scala:193)
	at sbt.TaskManager$Task.invoke(TaskManager.scala:62)
	at sbt.impl.RunTask.doRun$1(RunTask.scala:77)
	at sbt.impl.RunTask.runTask(RunTask.scala:85)
	at sbt.impl.RunTask.sbt$impl$RunTask$$runIfNotRoot(RunTask.scala:60)
	at sbt.impl.RunTask$$anonfun$runTasksExceptRoot$2.apply(RunTask.scala:48)
	at sbt.impl.RunTask$$anonfun$runTasksExceptRoot$2.apply(RunTask.scala:48)
	at sbt.Distributor$Run$Worker$$anonfun$2.apply(ParallelRunner.scala:131)
	at sbt.Distributor$Run$Worker$$anonfun$2.apply(ParallelRunner.scala:131)
	at sbt.Control$.trapUnit(Control.scala:19)
	at sbt.Distributor$Run$Worker.run(ParallelRunner.scala:131)
[info] Test Starting: testConsumerEmptyTopic(kafka.integration.PrimitiveApiTest)
[info] Test Passed: testConsumerEmptyTopic(kafka.integration.PrimitiveApiTest)
[info] == core-kafka / kafka.integration.PrimitiveApiTest ==
[info] 
[info] == core-kafka / kafka.consumer.ZookeeperConsumerConnectorTest ==
[info] Test Starting: testBasic(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] Test Passed: testBasic(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] Test Starting: testCompression(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] Test Passed: testCompression(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] Test Starting: testCompressionSetConsumption(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] Test Passed: testCompressionSetConsumption(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] Test Starting: testConsumerDecoder(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] Test Passed: testConsumerDecoder(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] Test Starting: testLeaderSelectionForPartition(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] Test Passed: testLeaderSelectionForPartition(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] == core-kafka / kafka.consumer.ZookeeperConsumerConnectorTest ==
[info] 
[info] == core-kafka / kafka.utils.UtilsTest ==
[info] Test Starting: testSwallow
[info] Test Passed: testSwallow
[info] Test Starting: testCircularIterator
[info] Test Passed: testCircularIterator
[info] == core-kafka / kafka.utils.UtilsTest ==
[info] 
[info] == core-kafka / kafka.message.MessageTest ==
[info] Test Starting: testFieldValues
[info] Test Passed: testFieldValues
[info] Test Starting: testChecksum
[info] Test Passed: testChecksum
[info] Test Starting: testEquality
[info] Test Passed: testEquality
[info] Test Starting: testIsHashable
[info] Test Passed: testIsHashable
[info] == core-kafka / kafka.message.MessageTest ==
[info] 
[info] == core-kafka / kafka.server.LogRecoveryTest ==
[info] Test Starting: testHWCheckpointNoFailuresSingleLogSegment(kafka.server.LogRecoveryTest)
[info] Test Passed: testHWCheckpointNoFailuresSingleLogSegment(kafka.server.LogRecoveryTest)
[info] Test Starting: testHWCheckpointWithFailuresSingleLogSegment(kafka.server.LogRecoveryTest)
[info] Test Passed: testHWCheckpointWithFailuresSingleLogSegment(kafka.server.LogRecoveryTest)
[info] Test Starting: testHWCheckpointNoFailuresMultipleLogSegments(kafka.server.LogRecoveryTest)
[info] Test Passed: testHWCheckpointNoFailuresMultipleLogSegments(kafka.server.LogRecoveryTest)
[info] Test Starting: testHWCheckpointWithFailuresMultipleLogSegments(kafka.server.LogRecoveryTest)
[info] Test Passed: testHWCheckpointWithFailuresMultipleLogSegments(kafka.server.LogRecoveryTest)
[info] == core-kafka / kafka.server.LogRecoveryTest ==
[info] 
[info] == core-kafka / kafka.log.FileMessageSetTest ==
[info] Test Starting: testWrittenEqualsRead
[info] Test Passed: testWrittenEqualsRead
[info] Test Starting: testIteratorIsConsistent
[info] Test Passed: testIteratorIsConsistent
[info] Test Starting: testSizeInBytes
[info] Test Passed: testSizeInBytes
[info] Test Starting: testWriteTo
[info] Test Passed: testWriteTo
[info] Test Starting: testFileSize
[info] Test Passed: testFileSize
[info] Test Starting: testIterationOverPartialAndTruncation
[info] Test Passed: testIterationOverPartialAndTruncation
[info] Test Starting: testIterationDoesntChangePosition
[info] Test Passed: testIterationDoesntChangePosition
[info] Test Starting: testRead
[info] Test Passed: testRead
[info] Test Starting: testSearch
[info] Test Passed: testSearch
[info] == core-kafka / kafka.log.FileMessageSetTest ==
[info] 
[info] == core-kafka / kafka.server.HighwatermarkPersistenceTest ==
[info] Test Starting: testHighWatermarkPersistenceSinglePartition(kafka.server.HighwatermarkPersistenceTest)
[info] Test Passed: testHighWatermarkPersistenceSinglePartition(kafka.server.HighwatermarkPersistenceTest)
[info] Test Starting: testHighWatermarkPersistenceMultiplePartitions(kafka.server.HighwatermarkPersistenceTest)
[info] Test Passed: testHighWatermarkPersistenceMultiplePartitions(kafka.server.HighwatermarkPersistenceTest)
[info] == core-kafka / kafka.server.HighwatermarkPersistenceTest ==
[info] 
[info] == core-kafka / kafka.server.RequestPurgatoryTest ==
[info] Test Starting: testRequestSatisfaction(kafka.server.RequestPurgatoryTest)
[info] Test Passed: testRequestSatisfaction(kafka.server.RequestPurgatoryTest)
[info] Test Starting: testRequestExpiry(kafka.server.RequestPurgatoryTest)
[info] Test Passed: testRequestExpiry(kafka.server.RequestPurgatoryTest)
[info] == core-kafka / kafka.server.RequestPurgatoryTest ==
[info] 
[info] == core-kafka / kafka.producer.SyncProducerTest ==
[info] Test Starting: testReachableServer(kafka.producer.SyncProducerTest)
[info] Test Passed: testReachableServer(kafka.producer.SyncProducerTest)
[info] Test Starting: testEmptyProduceRequest(kafka.producer.SyncProducerTest)
[info] Test Passed: testEmptyProduceRequest(kafka.producer.SyncProducerTest)
[info] Test Starting: testMessageSizeTooLarge(kafka.producer.SyncProducerTest)
[info] Test Passed: testMessageSizeTooLarge(kafka.producer.SyncProducerTest)
[info] Test Starting: testProduceCorrectlyReceivesResponse(kafka.producer.SyncProducerTest)
[info] Test Passed: testProduceCorrectlyReceivesResponse(kafka.producer.SyncProducerTest)
[info] Test Starting: testProducerCanTimeout(kafka.producer.SyncProducerTest)
[info] Test Passed: testProducerCanTimeout(kafka.producer.SyncProducerTest)
[info] == core-kafka / kafka.producer.SyncProducerTest ==
[info] 
[info] == core-kafka / kafka.integration.AutoOffsetResetTest ==
[info] Test Starting: testResetToEarliestWhenOffsetTooHigh(kafka.integration.AutoOffsetResetTest)
[info] Test Passed: testResetToEarliestWhenOffsetTooHigh(kafka.integration.AutoOffsetResetTest)
[info] Test Starting: testResetToEarliestWhenOffsetTooLow(kafka.integration.AutoOffsetResetTest)
[info] Test Passed: testResetToEarliestWhenOffsetTooLow(kafka.integration.AutoOffsetResetTest)
[info] Test Starting: testResetToLatestWhenOffsetTooHigh(kafka.integration.AutoOffsetResetTest)
[info] Test Passed: testResetToLatestWhenOffsetTooHigh(kafka.integration.AutoOffsetResetTest)
[info] Test Starting: testResetToLatestWhenOffsetTooLow(kafka.integration.AutoOffsetResetTest)
[info] Test Passed: testResetToLatestWhenOffsetTooLow(kafka.integration.AutoOffsetResetTest)
[info] == core-kafka / kafka.integration.AutoOffsetResetTest ==
[info] 
[info] == core-kafka / kafka.message.ByteBufferMessageSetTest ==
[info] Test Starting: testWrittenEqualsRead
[info] Test Passed: testWrittenEqualsRead
[info] Test Starting: testIteratorIsConsistent
[info] Test Passed: testIteratorIsConsistent
[info] Test Starting: testSizeInBytes
[info] Test Passed: testSizeInBytes
[info] Test Starting: testEquals
[info] Test Passed: testEquals
[info] Test Starting: testWriteTo
[info] Test Passed: testWriteTo
[info] Test Starting: testValidBytes
[info] Test Passed: testValidBytes
[info] Test Starting: testValidBytesWithCompression
[info] Test Passed: testValidBytesWithCompression
[info] Test Starting: testIterator
[info] Test Passed: testIterator
[info] Test Starting: testOffsetAssignment
[info] Test Passed: testOffsetAssignment
[info] == core-kafka / kafka.message.ByteBufferMessageSetTest ==
[info] 
[info] == core-kafka / test-finish ==
[error] Failed: : Total 163, Failed 2, Errors 0, Passed 161, Skipped 0
[info] == core-kafka / test-finish ==
[info] 
[info] == core-kafka / Test cleanup 1 ==
[info] Deleting directory /tmp/sbt_b9aec80b
[info] == core-kafka / Test cleanup 1 ==
[info] 
[info] == core-kafka / test-cleanup ==
[info] == core-kafka / test-cleanup ==
[info] 
[info] == java-examples / copy-resources ==
[info] == java-examples / copy-resources ==
[info] 
[info] == perf / copy-resources ==
[info] == perf / copy-resources ==
[error] Error running kafka.log.LogTest: Test FAILED
[error] Error running kafka.integration.PrimitiveApiTest: Test FAILED
[error] Error running test: One or more subtasks failed
[info] 
[info] Total time: 231 s, completed Nov 2, 2012 7:44:55 PM
[info] 
[info] Total session time: 232 s, completed Nov 2, 2012 7:44:55 PM
[error] Error during build.
Build step 'Execute shell' marked build as failure