You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@kafka.apache.org by Apache Jenkins Server <je...@builds.apache.org> on 2012/11/06 05:08:37 UTC
Build failed in Jenkins: Kafka-0.8 #101
See <https://builds.apache.org/job/Kafka-0.8/101/changes>
Changes:
[nehanarkhede] KAFKA-593 Empty log index file created when it shouldn't be empty; Patched by Yang Ye; reviewed by Jun and Jay
------------------------------------------
[...truncated 885 lines...]
[0m[[0minfo[0m] [0mTest Starting: testBasic(kafka.consumer.ZookeeperConsumerConnectorTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testBasic(kafka.consumer.ZookeeperConsumerConnectorTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testCompression(kafka.consumer.ZookeeperConsumerConnectorTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testCompression(kafka.consumer.ZookeeperConsumerConnectorTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testCompressionSetConsumption(kafka.consumer.ZookeeperConsumerConnectorTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testCompressionSetConsumption(kafka.consumer.ZookeeperConsumerConnectorTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testConsumerDecoder(kafka.consumer.ZookeeperConsumerConnectorTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testConsumerDecoder(kafka.consumer.ZookeeperConsumerConnectorTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testLeaderSelectionForPartition(kafka.consumer.ZookeeperConsumerConnectorTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testLeaderSelectionForPartition(kafka.consumer.ZookeeperConsumerConnectorTest)[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.consumer.ZookeeperConsumerConnectorTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.server.ReplicaFetchTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testReplicaFetcherThread(kafka.server.ReplicaFetchTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testReplicaFetcherThread(kafka.server.ReplicaFetchTest)[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.server.ReplicaFetchTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.server.ServerShutdownTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testCleanShutdown(kafka.server.ServerShutdownTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testCleanShutdown(kafka.server.ServerShutdownTest)[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.server.ServerShutdownTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.integration.FetcherTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testFetcher(kafka.integration.FetcherTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testFetcher(kafka.integration.FetcherTest)[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.integration.FetcherTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.integration.TopicMetadataTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testTopicMetadataRequest(kafka.integration.TopicMetadataTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testTopicMetadataRequest(kafka.integration.TopicMetadataTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testBasicTopicMetadata(kafka.integration.TopicMetadataTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testBasicTopicMetadata(kafka.integration.TopicMetadataTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testAutoCreateTopic(kafka.integration.TopicMetadataTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testAutoCreateTopic(kafka.integration.TopicMetadataTest)[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.integration.TopicMetadataTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.producer.SyncProducerTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testReachableServer(kafka.producer.SyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testReachableServer(kafka.producer.SyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testEmptyProduceRequest(kafka.producer.SyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testEmptyProduceRequest(kafka.producer.SyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testMessageSizeTooLarge(kafka.producer.SyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testMessageSizeTooLarge(kafka.producer.SyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testProduceCorrectlyReceivesResponse(kafka.producer.SyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testProduceCorrectlyReceivesResponse(kafka.producer.SyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testProducerCanTimeout(kafka.producer.SyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testProducerCanTimeout(kafka.producer.SyncProducerTest)[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.producer.SyncProducerTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.message.ByteBufferMessageSetTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testWrittenEqualsRead[0m
[0m[[0minfo[0m] [0mTest Passed: testWrittenEqualsRead[0m
[0m[[0minfo[0m] [0mTest Starting: testIteratorIsConsistent[0m
[0m[[0minfo[0m] [0mTest Passed: testIteratorIsConsistent[0m
[0m[[0minfo[0m] [0mTest Starting: testSizeInBytes[0m
[0m[[0minfo[0m] [0mTest Passed: testSizeInBytes[0m
[0m[[0minfo[0m] [0mTest Starting: testEquals[0m
[0m[[0minfo[0m] [0mTest Passed: testEquals[0m
[0m[[0minfo[0m] [0mTest Starting: testWriteTo[0m
[0m[[0minfo[0m] [0mTest Passed: testWriteTo[0m
[0m[[0minfo[0m] [0mTest Starting: testValidBytes[0m
[0m[[0minfo[0m] [0mTest Passed: testValidBytes[0m
[0m[[0minfo[0m] [0mTest Starting: testValidBytesWithCompression[0m
[0m[[0minfo[0m] [0mTest Passed: testValidBytesWithCompression[0m
[0m[[0minfo[0m] [0mTest Starting: testIterator[0m
[0m[[0minfo[0m] [0mTest Passed: testIterator[0m
[0m[[0minfo[0m] [0mTest Starting: testOffsetAssignment[0m
[0m[[0minfo[0m] [0mTest Passed: testOffsetAssignment[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.message.ByteBufferMessageSetTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.log.FileMessageSetTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testWrittenEqualsRead[0m
[0m[[0minfo[0m] [0mTest Passed: testWrittenEqualsRead[0m
[0m[[0minfo[0m] [0mTest Starting: testIteratorIsConsistent[0m
[0m[[0minfo[0m] [0mTest Passed: testIteratorIsConsistent[0m
[0m[[0minfo[0m] [0mTest Starting: testSizeInBytes[0m
[0m[[0minfo[0m] [0mTest Passed: testSizeInBytes[0m
[0m[[0minfo[0m] [0mTest Starting: testWriteTo[0m
[0m[[0minfo[0m] [0mTest Passed: testWriteTo[0m
[0m[[0minfo[0m] [0mTest Starting: testFileSize[0m
[0m[[0minfo[0m] [0mTest Passed: testFileSize[0m
[0m[[0minfo[0m] [0mTest Starting: testIterationOverPartialAndTruncation[0m
[0m[[0minfo[0m] [0mTest Passed: testIterationOverPartialAndTruncation[0m
[0m[[0minfo[0m] [0mTest Starting: testIterationDoesntChangePosition[0m
[0m[[0minfo[0m] [0mTest Passed: testIterationDoesntChangePosition[0m
[0m[[0minfo[0m] [0mTest Starting: testRead[0m
[0m[[0minfo[0m] [0mTest Passed: testRead[0m
[0m[[0minfo[0m] [0mTest Starting: testSearch[0m
[0m[[0minfo[0m] [0mTest Passed: testSearch[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.log.FileMessageSetTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.integration.AutoOffsetResetTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testResetToEarliestWhenOffsetTooHigh(kafka.integration.AutoOffsetResetTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testResetToEarliestWhenOffsetTooHigh(kafka.integration.AutoOffsetResetTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testResetToEarliestWhenOffsetTooLow(kafka.integration.AutoOffsetResetTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testResetToEarliestWhenOffsetTooLow(kafka.integration.AutoOffsetResetTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testResetToLatestWhenOffsetTooHigh(kafka.integration.AutoOffsetResetTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testResetToLatestWhenOffsetTooHigh(kafka.integration.AutoOffsetResetTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testResetToLatestWhenOffsetTooLow(kafka.integration.AutoOffsetResetTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testResetToLatestWhenOffsetTooLow(kafka.integration.AutoOffsetResetTest)[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.integration.AutoOffsetResetTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.server.LeaderElectionTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testLeaderElectionAndEpoch(kafka.server.LeaderElectionTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testLeaderElectionAndEpoch(kafka.server.LeaderElectionTest)[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.server.LeaderElectionTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.zk.ZKEphemeralTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testEphemeralNodeCleanup(kafka.zk.ZKEphemeralTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testEphemeralNodeCleanup(kafka.zk.ZKEphemeralTest)[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.zk.ZKEphemeralTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.integration.LazyInitProducerTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testProduceAndMultiFetch(kafka.integration.LazyInitProducerTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testProduceAndMultiFetch(kafka.integration.LazyInitProducerTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testMultiProduce(kafka.integration.LazyInitProducerTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testMultiProduce(kafka.integration.LazyInitProducerTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testProduceAndFetch(kafka.integration.LazyInitProducerTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testProduceAndFetch(kafka.integration.LazyInitProducerTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testMultiProduceResend(kafka.integration.LazyInitProducerTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testMultiProduceResend(kafka.integration.LazyInitProducerTest)[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.integration.LazyInitProducerTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.admin.AdminTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testReplicaAssignment(kafka.admin.AdminTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testReplicaAssignment(kafka.admin.AdminTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testManualReplicaAssignment(kafka.admin.AdminTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testManualReplicaAssignment(kafka.admin.AdminTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testTopicCreationInZK(kafka.admin.AdminTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testTopicCreationInZK(kafka.admin.AdminTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testGetTopicMetadata(kafka.admin.AdminTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testGetTopicMetadata(kafka.admin.AdminTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testPartitionReassignmentWithLeaderInNewReplicas(kafka.admin.AdminTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testPartitionReassignmentWithLeaderInNewReplicas(kafka.admin.AdminTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testPartitionReassignmentWithLeaderNotInNewReplicas(kafka.admin.AdminTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testPartitionReassignmentWithLeaderNotInNewReplicas(kafka.admin.AdminTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testPartitionReassignmentNonOverlappingReplicas(kafka.admin.AdminTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testPartitionReassignmentNonOverlappingReplicas(kafka.admin.AdminTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testReassigningNonExistingPartition(kafka.admin.AdminTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testReassigningNonExistingPartition(kafka.admin.AdminTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testResumePartitionReassignmentThatWasCompleted(kafka.admin.AdminTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testResumePartitionReassignmentThatWasCompleted(kafka.admin.AdminTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testPreferredReplicaJsonData(kafka.admin.AdminTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testPreferredReplicaJsonData(kafka.admin.AdminTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testBasicPreferredReplicaElection(kafka.admin.AdminTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testBasicPreferredReplicaElection(kafka.admin.AdminTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testShutdownBroker(kafka.admin.AdminTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testShutdownBroker(kafka.admin.AdminTest)[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.admin.AdminTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.log.LogTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testTimeBasedLogRoll[0m
[0m[[31merror[0m] [0mTest Failed: testTimeBasedLogRoll[0m
junit.framework.AssertionFailedError: There should be exactly one segment. expected:<1> but was:<2>
at junit.framework.Assert.fail(Assert.java:47)
at junit.framework.Assert.failNotEquals(Assert.java:277)
at junit.framework.Assert.assertEquals(Assert.java:64)
at junit.framework.Assert.assertEquals(Assert.java:195)
at kafka.log.LogTest.testTimeBasedLogRoll(LogTest.scala:73)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at org.junit.internal.runners.TestMethodRunner.executeMethodBody(TestMethodRunner.java:99)
at org.junit.internal.runners.TestMethodRunner.runUnprotected(TestMethodRunner.java:81)
at org.junit.internal.runners.BeforeAndAfterRunner.runProtected(BeforeAndAfterRunner.java:34)
at org.junit.internal.runners.TestMethodRunner.runMethod(TestMethodRunner.java:75)
at org.junit.internal.runners.TestMethodRunner.run(TestMethodRunner.java:45)
at org.junit.internal.runners.TestClassMethodsRunner.invokeTestMethod(TestClassMethodsRunner.java:71)
at org.junit.internal.runners.TestClassMethodsRunner.run(TestClassMethodsRunner.java:35)
at org.junit.internal.runners.TestClassRunner$1.runUnprotected(TestClassRunner.java:42)
at org.junit.internal.runners.BeforeAndAfterRunner.runProtected(BeforeAndAfterRunner.java:34)
at org.junit.internal.runners.TestClassRunner.run(TestClassRunner.java:52)
at org.junit.internal.runners.CompositeRunner.run(CompositeRunner.java:29)
at org.junit.runner.JUnitCore.run(JUnitCore.java:121)
at org.junit.runner.JUnitCore.run(JUnitCore.java:100)
at org.junit.runner.JUnitCore.run(JUnitCore.java:91)
at org.scalatest.junit.JUnitSuite$class.run(JUnitSuite.scala:261)
at kafka.log.LogTest.run(LogTest.scala:32)
at org.scalatest.tools.ScalaTestFramework$ScalaTestRunner.run(ScalaTestFramework.scala:40)
at sbt.TestRunner.run(TestFramework.scala:53)
at sbt.TestRunner.runTest$1(TestFramework.scala:67)
at sbt.TestRunner.run(TestFramework.scala:76)
at sbt.TestFramework$$anonfun$10$$anonfun$apply$11.runTest$2(TestFramework.scala:194)
at sbt.TestFramework$$anonfun$10$$anonfun$apply$11$$anonfun$apply$12.apply(TestFramework.scala:205)
at sbt.TestFramework$$anonfun$10$$anonfun$apply$11$$anonfun$apply$12.apply(TestFramework.scala:205)
at sbt.NamedTestTask.run(TestFramework.scala:92)
at sbt.ScalaProject$$anonfun$sbt$ScalaProject$$toTask$1.apply(ScalaProject.scala:193)
at sbt.ScalaProject$$anonfun$sbt$ScalaProject$$toTask$1.apply(ScalaProject.scala:193)
at sbt.TaskManager$Task.invoke(TaskManager.scala:62)
at sbt.impl.RunTask.doRun$1(RunTask.scala:77)
at sbt.impl.RunTask.runTask(RunTask.scala:85)
at sbt.impl.RunTask.sbt$impl$RunTask$$runIfNotRoot(RunTask.scala:60)
at sbt.impl.RunTask$$anonfun$runTasksExceptRoot$2.apply(RunTask.scala:48)
at sbt.impl.RunTask$$anonfun$runTasksExceptRoot$2.apply(RunTask.scala:48)
at sbt.Distributor$Run$Worker$$anonfun$2.apply(ParallelRunner.scala:131)
at sbt.Distributor$Run$Worker$$anonfun$2.apply(ParallelRunner.scala:131)
at sbt.Control$.trapUnit(Control.scala:19)
at sbt.Distributor$Run$Worker.run(ParallelRunner.scala:131)
[0m[[0minfo[0m] [0mTest Starting: testSizeBasedLogRoll[0m
[0m[[0minfo[0m] [0mTest Passed: testSizeBasedLogRoll[0m
[0m[[0minfo[0m] [0mTest Starting: testLoadEmptyLog[0m
[0m[[0minfo[0m] [0mTest Passed: testLoadEmptyLog[0m
[0m[[0minfo[0m] [0mTest Starting: testAppendAndRead[0m
[0m[[0minfo[0m] [0mTest Passed: testAppendAndRead[0m
[0m[[0minfo[0m] [0mTest Starting: testReadOutOfRange[0m
[0m[[0minfo[0m] [0mTest Passed: testReadOutOfRange[0m
[0m[[0minfo[0m] [0mTest Starting: testLogRolls[0m
[0m[[0minfo[0m] [0mTest Passed: testLogRolls[0m
[0m[[0minfo[0m] [0mTest Starting: testCompressedMessages[0m
[0m[[0minfo[0m] [0mTest Passed: testCompressedMessages[0m
[0m[[0minfo[0m] [0mTest Starting: testFindSegment[0m
[0m[[0minfo[0m] [0mTest Passed: testFindSegment[0m
[0m[[0minfo[0m] [0mTest Starting: testEdgeLogRollsStartingAtZero[0m
[0m[[0minfo[0m] [0mTest Passed: testEdgeLogRollsStartingAtZero[0m
[0m[[0minfo[0m] [0mTest Starting: testEdgeLogRollsStartingAtNonZero[0m
[0m[[0minfo[0m] [0mTest Passed: testEdgeLogRollsStartingAtNonZero[0m
[0m[[0minfo[0m] [0mTest Starting: testMessageSizeCheck[0m
[0m[[0minfo[0m] [0mTest Passed: testMessageSizeCheck[0m
[0m[[0minfo[0m] [0mTest Starting: testLogRecoversToCorrectOffset[0m
[0m[[0minfo[0m] [0mTest Passed: testLogRecoversToCorrectOffset[0m
[0m[[0minfo[0m] [0mTest Starting: testTruncateTo[0m
[0m[[0minfo[0m] [0mTest Passed: testTruncateTo[0m
[0m[[0minfo[0m] [0mTest Starting: testIndexResizingAtTruncation[0m
[0m[[0minfo[0m] [0mTest Passed: testIndexResizingAtTruncation[0m
[0m[[0minfo[0m] [0mTest Starting: testAppendWithoutOffsetAssignment[0m
[0m[[0minfo[0m] [0mTest Passed: testAppendWithoutOffsetAssignment[0m
[0m[[0minfo[0m] [0mTest Starting: testReopenThenTruncate[0m
[0m[[0minfo[0m] [0mTest Passed: testReopenThenTruncate[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.log.LogTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / test-finish ==[0m
[0m[[31merror[0m] [0mFailed: : Total 164, Failed 1, Errors 0, Passed 163, Skipped 0[0m
[0m[[0minfo[0m] [34m== core-kafka / test-finish ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / Test cleanup 1 ==[0m
[0m[[0minfo[0m] [0mDeleting directory /tmp/sbt_6aafdbf3[0m
[0m[[0minfo[0m] [34m== core-kafka / Test cleanup 1 ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / test-cleanup ==[0m
[0m[[0minfo[0m] [34m== core-kafka / test-cleanup ==[0m
[0m[[31merror[0m] [0mError running kafka.log.LogTest: Test FAILED[0m
[0m[[31merror[0m] [0mError running test: One or more subtasks failed[0m
[0m[[0minfo[0m] [0m[0m
[0m[[0minfo[0m] [0mTotal time: 178 s, completed Nov 6, 2012 4:08:36 AM[0m
[0m[[0minfo[0m] [0m[0m
[0m[[0minfo[0m] [0mTotal session time: 178 s, completed Nov 6, 2012 4:08:36 AM[0m
[0m[[31merror[0m] [0mError during build.[0m
Build step 'Execute shell' marked build as failure
Build failed in Jenkins: Kafka-0.8 #103
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/Kafka-0.8/103/changes>
Changes:
[junrao] Add test cases to test log size retention and more; patched by John Fung; reviewed by Jun Rao; KAFKA-591
------------------------------------------
[...truncated 3014 lines...]
at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
at scala.collection.immutable.Map$Map1.foreach(Map.scala:105)
at scala.collection.TraversableLike$class.map(TraversableLike.scala:206)
at scala.collection.immutable.Map$Map1.map(Map.scala:93)
at kafka.server.KafkaApis.appendToLocalLog(KafkaApis.scala:243)
at kafka.server.KafkaApis.handleProducerRequest(KafkaApis.scala:183)
at kafka.server.KafkaApis.handle(KafkaApis.scala:57)
at kafka.server.KafkaRequestHandler.run(KafkaRequestHandler.scala:41)
at java.lang.Thread.run(Thread.java:662)
[0m[[0minfo[0m] [0mTest Passed: testReachableServer(kafka.producer.SyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testEmptyProduceRequest(kafka.producer.SyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testEmptyProduceRequest(kafka.producer.SyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testMessageSizeTooLarge(kafka.producer.SyncProducerTest)[0m
[2012-11-09 23:18:53,243] ERROR [KafkaApi-0] Error processing ProducerRequest on test:0 (kafka.server.KafkaApis:102)
kafka.common.MessageSizeTooLargeException: Message size is 1000023 bytes which exceeds the maximum configured message size of 1000000.
at kafka.log.Log$$anonfun$analyzeAndValidateMessageSet$1.apply(Log.scala:327)
at kafka.log.Log$$anonfun$analyzeAndValidateMessageSet$1.apply(Log.scala:313)
at scala.collection.Iterator$class.foreach(Iterator.scala:631)
at kafka.utils.IteratorTemplate.foreach(IteratorTemplate.scala:32)
at kafka.log.Log.analyzeAndValidateMessageSet(Log.scala:313)
at kafka.log.Log.append(Log.scala:241)
at kafka.server.KafkaApis$$anonfun$appendToLocalLog$2.apply(KafkaApis.scala:250)
at kafka.server.KafkaApis$$anonfun$appendToLocalLog$2.apply(KafkaApis.scala:243)
at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
at scala.collection.immutable.Map$Map1.foreach(Map.scala:105)
at scala.collection.TraversableLike$class.map(TraversableLike.scala:206)
at scala.collection.immutable.Map$Map1.map(Map.scala:93)
at kafka.server.KafkaApis.appendToLocalLog(KafkaApis.scala:243)
at kafka.server.KafkaApis.handleProducerRequest(KafkaApis.scala:183)
at kafka.server.KafkaApis.handle(KafkaApis.scala:57)
at kafka.server.KafkaRequestHandler.run(KafkaRequestHandler.scala:41)
at java.lang.Thread.run(Thread.java:662)
[0m[[0minfo[0m] [0mTest Passed: testMessageSizeTooLarge(kafka.producer.SyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testProduceCorrectlyReceivesResponse(kafka.producer.SyncProducerTest)[0m
[2012-11-09 23:18:53,502] ERROR [KafkaApi-0] Error processing ProducerRequest on topic1:0 (kafka.server.KafkaApis:102)
kafka.common.UnknownTopicOrPartitionException: Topic topic1 partition 0 doesn't exist on 0
at kafka.server.ReplicaManager.getLeaderReplicaIfLocal(ReplicaManager.scala:142)
at kafka.server.KafkaApis$$anonfun$appendToLocalLog$2.apply(KafkaApis.scala:248)
at kafka.server.KafkaApis$$anonfun$appendToLocalLog$2.apply(KafkaApis.scala:243)
at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
at scala.collection.immutable.Map$Map3.foreach(Map.scala:152)
at scala.collection.TraversableLike$class.map(TraversableLike.scala:206)
at scala.collection.immutable.Map$Map3.map(Map.scala:132)
at kafka.server.KafkaApis.appendToLocalLog(KafkaApis.scala:243)
at kafka.server.KafkaApis.handleProducerRequest(KafkaApis.scala:183)
at kafka.server.KafkaApis.handle(KafkaApis.scala:57)
at kafka.server.KafkaRequestHandler.run(KafkaRequestHandler.scala:41)
at java.lang.Thread.run(Thread.java:662)
[2012-11-09 23:18:53,503] ERROR [KafkaApi-0] Error processing ProducerRequest on topic3:0 (kafka.server.KafkaApis:102)
kafka.common.UnknownTopicOrPartitionException: Topic topic3 partition 0 doesn't exist on 0
at kafka.server.ReplicaManager.getLeaderReplicaIfLocal(ReplicaManager.scala:142)
at kafka.server.KafkaApis$$anonfun$appendToLocalLog$2.apply(KafkaApis.scala:248)
at kafka.server.KafkaApis$$anonfun$appendToLocalLog$2.apply(KafkaApis.scala:243)
at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
at scala.collection.immutable.Map$Map3.foreach(Map.scala:152)
at scala.collection.TraversableLike$class.map(TraversableLike.scala:206)
at scala.collection.immutable.Map$Map3.map(Map.scala:132)
at kafka.server.KafkaApis.appendToLocalLog(KafkaApis.scala:243)
at kafka.server.KafkaApis.handleProducerRequest(KafkaApis.scala:183)
at kafka.server.KafkaApis.handle(KafkaApis.scala:57)
at kafka.server.KafkaRequestHandler.run(KafkaRequestHandler.scala:41)
at java.lang.Thread.run(Thread.java:662)
[2012-11-09 23:18:53,504] ERROR [KafkaApi-0] Error processing ProducerRequest on topic2:0 (kafka.server.KafkaApis:102)
kafka.common.UnknownTopicOrPartitionException: Topic topic2 partition 0 doesn't exist on 0
at kafka.server.ReplicaManager.getLeaderReplicaIfLocal(ReplicaManager.scala:142)
at kafka.server.KafkaApis$$anonfun$appendToLocalLog$2.apply(KafkaApis.scala:248)
at kafka.server.KafkaApis$$anonfun$appendToLocalLog$2.apply(KafkaApis.scala:243)
at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
at scala.collection.immutable.Map$Map3.foreach(Map.scala:152)
at scala.collection.TraversableLike$class.map(TraversableLike.scala:206)
at scala.collection.immutable.Map$Map3.map(Map.scala:132)
at kafka.server.KafkaApis.appendToLocalLog(KafkaApis.scala:243)
at kafka.server.KafkaApis.handleProducerRequest(KafkaApis.scala:183)
at kafka.server.KafkaApis.handle(KafkaApis.scala:57)
at kafka.server.KafkaRequestHandler.run(KafkaRequestHandler.scala:41)
at java.lang.Thread.run(Thread.java:662)
[2012-11-09 23:18:54,657] ERROR [KafkaApi-0] Error processing ProducerRequest on topic2:0 (kafka.server.KafkaApis:102)
kafka.common.UnknownTopicOrPartitionException: Topic topic2 partition 0 doesn't exist on 0
at kafka.server.ReplicaManager.getLeaderReplicaIfLocal(ReplicaManager.scala:142)
at kafka.server.KafkaApis$$anonfun$appendToLocalLog$2.apply(KafkaApis.scala:248)
at kafka.server.KafkaApis$$anonfun$appendToLocalLog$2.apply(KafkaApis.scala:243)
at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
at scala.collection.immutable.Map$Map3.foreach(Map.scala:152)
at scala.collection.TraversableLike$class.map(TraversableLike.scala:206)
at scala.collection.immutable.Map$Map3.map(Map.scala:132)
at kafka.server.KafkaApis.appendToLocalLog(KafkaApis.scala:243)
at kafka.server.KafkaApis.handleProducerRequest(KafkaApis.scala:183)
at kafka.server.KafkaApis.handle(KafkaApis.scala:57)
at kafka.server.KafkaRequestHandler.run(KafkaRequestHandler.scala:41)
at java.lang.Thread.run(Thread.java:662)
[0m[[0minfo[0m] [0mTest Passed: testProduceCorrectlyReceivesResponse(kafka.producer.SyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testProducerCanTimeout(kafka.producer.SyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testProducerCanTimeout(kafka.producer.SyncProducerTest)[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.producer.SyncProducerTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.log4j.KafkaLog4jAppenderTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testKafkaLog4jConfigs(kafka.log4j.KafkaLog4jAppenderTest)[0m
log4j:WARN No appenders could be found for logger (org.I0Itec.zkclient.ZkEventThread).
log4j:WARN Please initialize the log4j system properly.
[0m[[0minfo[0m] [0mTest Passed: testKafkaLog4jConfigs(kafka.log4j.KafkaLog4jAppenderTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testLog4jAppends(kafka.log4j.KafkaLog4jAppenderTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testLog4jAppends(kafka.log4j.KafkaLog4jAppenderTest)[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.log4j.KafkaLog4jAppenderTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.integration.AutoOffsetResetTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testResetToEarliestWhenOffsetTooHigh(kafka.integration.AutoOffsetResetTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testResetToEarliestWhenOffsetTooHigh(kafka.integration.AutoOffsetResetTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testResetToEarliestWhenOffsetTooLow(kafka.integration.AutoOffsetResetTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testResetToEarliestWhenOffsetTooLow(kafka.integration.AutoOffsetResetTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testResetToLatestWhenOffsetTooHigh(kafka.integration.AutoOffsetResetTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testResetToLatestWhenOffsetTooHigh(kafka.integration.AutoOffsetResetTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testResetToLatestWhenOffsetTooLow(kafka.integration.AutoOffsetResetTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testResetToLatestWhenOffsetTooLow(kafka.integration.AutoOffsetResetTest)[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.integration.AutoOffsetResetTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.log.OffsetIndexTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: truncate[0m
[0m[[0minfo[0m] [0mTest Passed: truncate[0m
[0m[[0minfo[0m] [0mTest Starting: randomLookupTest[0m
[0m[[0minfo[0m] [0mTest Passed: randomLookupTest[0m
[0m[[0minfo[0m] [0mTest Starting: lookupExtremeCases[0m
[0m[[0minfo[0m] [0mTest Passed: lookupExtremeCases[0m
[0m[[0minfo[0m] [0mTest Starting: appendTooMany[0m
[0m[[0minfo[0m] [0mTest Passed: appendTooMany[0m
[0m[[0minfo[0m] [0mTest Starting: appendOutOfOrder[0m
[0m[[0minfo[0m] [0mTest Passed: appendOutOfOrder[0m
[0m[[0minfo[0m] [0mTest Starting: testReopen[0m
[0m[[0minfo[0m] [0mTest Passed: testReopen[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.log.OffsetIndexTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.integration.LazyInitProducerTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testProduceAndMultiFetch(kafka.integration.LazyInitProducerTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testProduceAndMultiFetch(kafka.integration.LazyInitProducerTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testMultiProduce(kafka.integration.LazyInitProducerTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testMultiProduce(kafka.integration.LazyInitProducerTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testProduceAndFetch(kafka.integration.LazyInitProducerTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testProduceAndFetch(kafka.integration.LazyInitProducerTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testMultiProduceResend(kafka.integration.LazyInitProducerTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testMultiProduceResend(kafka.integration.LazyInitProducerTest)[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.integration.LazyInitProducerTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.message.ByteBufferMessageSetTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testWrittenEqualsRead[0m
[0m[[0minfo[0m] [0mTest Passed: testWrittenEqualsRead[0m
[0m[[0minfo[0m] [0mTest Starting: testIteratorIsConsistent[0m
[0m[[0minfo[0m] [0mTest Passed: testIteratorIsConsistent[0m
[0m[[0minfo[0m] [0mTest Starting: testSizeInBytes[0m
[0m[[0minfo[0m] [0mTest Passed: testSizeInBytes[0m
[0m[[0minfo[0m] [0mTest Starting: testWriteTo[0m
[0m[[0minfo[0m] [0mTest Passed: testWriteTo[0m
[0m[[0minfo[0m] [0mTest Starting: testValidBytes[0m
[0m[[0minfo[0m] [0mTest Passed: testValidBytes[0m
[0m[[0minfo[0m] [0mTest Starting: testValidBytesWithCompression[0m
[0m[[0minfo[0m] [0mTest Passed: testValidBytesWithCompression[0m
[0m[[0minfo[0m] [0mTest Starting: testEquals[0m
[0m[[0minfo[0m] [0mTest Passed: testEquals[0m
[0m[[0minfo[0m] [0mTest Starting: testIterator[0m
[0m[[0minfo[0m] [0mTest Passed: testIterator[0m
[0m[[0minfo[0m] [0mTest Starting: testOffsetAssignment[0m
[0m[[0minfo[0m] [0mTest Passed: testOffsetAssignment[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.message.ByteBufferMessageSetTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.log.LogOffsetTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testGetOffsetsForUnknownTopic(kafka.log.LogOffsetTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testGetOffsetsForUnknownTopic(kafka.log.LogOffsetTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testGetOffsetsBeforeLatestTime(kafka.log.LogOffsetTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testGetOffsetsBeforeLatestTime(kafka.log.LogOffsetTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testEmptyLogsGetOffsets(kafka.log.LogOffsetTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testEmptyLogsGetOffsets(kafka.log.LogOffsetTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testGetOffsetsBeforeNow(kafka.log.LogOffsetTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testGetOffsetsBeforeNow(kafka.log.LogOffsetTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testGetOffsetsBeforeEarliestTime(kafka.log.LogOffsetTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testGetOffsetsBeforeEarliestTime(kafka.log.LogOffsetTest)[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.log.LogOffsetTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.consumer.TopicFilterTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testWhitelists[0m
[0m[[0minfo[0m] [0mTest Passed: testWhitelists[0m
[0m[[0minfo[0m] [0mTest Starting: testBlacklists[0m
[0m[[0minfo[0m] [0mTest Passed: testBlacklists[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.consumer.TopicFilterTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.javaapi.message.ByteBufferMessageSetTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testWrittenEqualsRead[0m
[0m[[0minfo[0m] [0mTest Passed: testWrittenEqualsRead[0m
[0m[[0minfo[0m] [0mTest Starting: testIteratorIsConsistent[0m
[0m[[0minfo[0m] [0mTest Passed: testIteratorIsConsistent[0m
[0m[[0minfo[0m] [0mTest Starting: testSizeInBytes[0m
[0m[[0minfo[0m] [0mTest Passed: testSizeInBytes[0m
[0m[[0minfo[0m] [0mTest Starting: testEquals[0m
[0m[[0minfo[0m] [0mTest Passed: testEquals[0m
[0m[[0minfo[0m] [0mTest Starting: testIteratorIsConsistentWithCompression[0m
[0m[[0minfo[0m] [0mTest Passed: testIteratorIsConsistentWithCompression[0m
[0m[[0minfo[0m] [0mTest Starting: testSizeInBytesWithCompression[0m
[0m[[0minfo[0m] [0mTest Passed: testSizeInBytesWithCompression[0m
[0m[[0minfo[0m] [0mTest Starting: testEqualsWithCompression[0m
[0m[[0minfo[0m] [0mTest Passed: testEqualsWithCompression[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.javaapi.message.ByteBufferMessageSetTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.integration.FetcherTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testFetcher(kafka.integration.FetcherTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testFetcher(kafka.integration.FetcherTest)[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.integration.FetcherTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.javaapi.consumer.ZookeeperConsumerConnectorTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testBasic(kafka.javaapi.consumer.ZookeeperConsumerConnectorTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testBasic(kafka.javaapi.consumer.ZookeeperConsumerConnectorTest)[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.javaapi.consumer.ZookeeperConsumerConnectorTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.metrics.KafkaTimerTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testKafkaTimer(kafka.metrics.KafkaTimerTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testKafkaTimer(kafka.metrics.KafkaTimerTest)[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.metrics.KafkaTimerTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.server.RequestPurgatoryTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testRequestSatisfaction(kafka.server.RequestPurgatoryTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testRequestSatisfaction(kafka.server.RequestPurgatoryTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testRequestExpiry(kafka.server.RequestPurgatoryTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testRequestExpiry(kafka.server.RequestPurgatoryTest)[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.server.RequestPurgatoryTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / Test cleanup 1 ==[0m
[0m[[0minfo[0m] [0mDeleting directory /tmp/sbt_8262927a[0m
[0m[[0minfo[0m] [34m== core-kafka / Test cleanup 1 ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / test-finish ==[0m
[0m[[31merror[0m] [0mFailed: : Total 165, Failed 2, Errors 0, Passed 163, Skipped 0[0m
[0m[[0minfo[0m] [34m== core-kafka / test-finish ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / test-cleanup ==[0m
[0m[[0minfo[0m] [34m== core-kafka / test-cleanup ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== hadoop consumer / copy-test-resources ==[0m
[0m[[0minfo[0m] [34m== hadoop consumer / copy-test-resources ==[0m
[0m[[31merror[0m] [0mError running kafka.log.LogTest: Test FAILED[0m
[0m[[31merror[0m] [0mError running kafka.server.LogRecoveryTest: Test FAILED[0m
[0m[[31merror[0m] [0mError running test: One or more subtasks failed[0m
[0m[[0minfo[0m] [0m[0m
[0m[[0minfo[0m] [0mTotal time: 173 s, completed Nov 9, 2012 11:19:19 PM[0m
[0m[[0minfo[0m] [0m[0m
[0m[[0minfo[0m] [0mTotal session time: 174 s, completed Nov 9, 2012 11:19:19 PM[0m
[0m[[31merror[0m] [0mError during build.[0m
Build step 'Execute shell' marked build as failure
Build failed in Jenkins: Kafka-0.8 #102
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/Kafka-0.8/102/changes>
Changes:
[junrao] Fix commit() in zk consumer for compressed messages; patched by Swapnil Ghike; reviewed by Jun Rao; KAFKA-546
------------------------------------------
[...truncated 3920 lines...]
at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
at scala.collection.immutable.Map$Map3.foreach(Map.scala:152)
at scala.collection.TraversableLike$class.map(TraversableLike.scala:206)
at scala.collection.immutable.Map$Map3.map(Map.scala:132)
at kafka.server.KafkaApis.appendToLocalLog(KafkaApis.scala:243)
at kafka.server.KafkaApis.handleProducerRequest(KafkaApis.scala:183)
at kafka.server.KafkaApis.handle(KafkaApis.scala:57)
at kafka.server.KafkaRequestHandler.run(KafkaRequestHandler.scala:41)
at java.lang.Thread.run(Thread.java:662)
[2012-11-08 07:10:42,302] ERROR [KafkaApi-0] Error processing ProducerRequest on topic2:0 (kafka.server.KafkaApis:102)
kafka.common.UnknownTopicOrPartitionException: Topic topic2 partition 0 doesn't exist on 0
at kafka.server.ReplicaManager.getLeaderReplicaIfLocal(ReplicaManager.scala:142)
at kafka.server.KafkaApis$$anonfun$appendToLocalLog$2.apply(KafkaApis.scala:248)
at kafka.server.KafkaApis$$anonfun$appendToLocalLog$2.apply(KafkaApis.scala:243)
at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
at scala.collection.immutable.Map$Map3.foreach(Map.scala:152)
at scala.collection.TraversableLike$class.map(TraversableLike.scala:206)
at scala.collection.immutable.Map$Map3.map(Map.scala:132)
at kafka.server.KafkaApis.appendToLocalLog(KafkaApis.scala:243)
at kafka.server.KafkaApis.handleProducerRequest(KafkaApis.scala:183)
at kafka.server.KafkaApis.handle(KafkaApis.scala:57)
at kafka.server.KafkaRequestHandler.run(KafkaRequestHandler.scala:41)
at java.lang.Thread.run(Thread.java:662)
[2012-11-08 07:10:43,620] ERROR [KafkaApi-0] Error processing ProducerRequest on topic2:0 (kafka.server.KafkaApis:102)
kafka.common.UnknownTopicOrPartitionException: Topic topic2 partition 0 doesn't exist on 0
at kafka.server.ReplicaManager.getLeaderReplicaIfLocal(ReplicaManager.scala:142)
at kafka.server.KafkaApis$$anonfun$appendToLocalLog$2.apply(KafkaApis.scala:248)
at kafka.server.KafkaApis$$anonfun$appendToLocalLog$2.apply(KafkaApis.scala:243)
at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
at scala.collection.immutable.Map$Map3.foreach(Map.scala:152)
at scala.collection.TraversableLike$class.map(TraversableLike.scala:206)
at scala.collection.immutable.Map$Map3.map(Map.scala:132)
at kafka.server.KafkaApis.appendToLocalLog(KafkaApis.scala:243)
at kafka.server.KafkaApis.handleProducerRequest(KafkaApis.scala:183)
at kafka.server.KafkaApis.handle(KafkaApis.scala:57)
at kafka.server.KafkaRequestHandler.run(KafkaRequestHandler.scala:41)
at java.lang.Thread.run(Thread.java:662)
[0m[[0minfo[0m] [0mTest Passed: testProduceCorrectlyReceivesResponse(kafka.producer.SyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testProducerCanTimeout(kafka.producer.SyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testProducerCanTimeout(kafka.producer.SyncProducerTest)[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.producer.SyncProducerTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.message.MessageCompressionTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testSimpleCompressDecompress[0m
[0m[[0minfo[0m] [0mTest Passed: testSimpleCompressDecompress[0m
[0m[[0minfo[0m] [0mTest Starting: testComplexCompressDecompress[0m
[0m[[0minfo[0m] [0mTest Passed: testComplexCompressDecompress[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.message.MessageCompressionTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.log.LogOffsetTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testGetOffsetsForUnknownTopic(kafka.log.LogOffsetTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testGetOffsetsForUnknownTopic(kafka.log.LogOffsetTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testGetOffsetsBeforeLatestTime(kafka.log.LogOffsetTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testGetOffsetsBeforeLatestTime(kafka.log.LogOffsetTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testEmptyLogsGetOffsets(kafka.log.LogOffsetTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testEmptyLogsGetOffsets(kafka.log.LogOffsetTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testGetOffsetsBeforeNow(kafka.log.LogOffsetTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testGetOffsetsBeforeNow(kafka.log.LogOffsetTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testGetOffsetsBeforeEarliestTime(kafka.log.LogOffsetTest)[0m
[2012-11-08 07:10:51,384] WARN fsync-ing the write ahead log in SyncThread:0 took 1193ms which will adversely effect operation latency. See the ZooKeeper troubleshooting guide (org.apache.zookeeper.server.persistence.FileTxnLog:321)
[0m[[0minfo[0m] [0mTest Passed: testGetOffsetsBeforeEarliestTime(kafka.log.LogOffsetTest)[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.log.LogOffsetTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.integration.TopicMetadataTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testTopicMetadataRequest(kafka.integration.TopicMetadataTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testTopicMetadataRequest(kafka.integration.TopicMetadataTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testBasicTopicMetadata(kafka.integration.TopicMetadataTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testBasicTopicMetadata(kafka.integration.TopicMetadataTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testAutoCreateTopic(kafka.integration.TopicMetadataTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testAutoCreateTopic(kafka.integration.TopicMetadataTest)[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.integration.TopicMetadataTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.server.LeaderElectionTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testLeaderElectionAndEpoch(kafka.server.LeaderElectionTest)[0m
[2012-11-08 07:10:55,097] ERROR Closing socket for /127.0.0.1 because of error (kafka.network.Processor:102)
java.io.IOException: Connection reset by peer
at sun.nio.ch.FileDispatcher.write0(Native Method)
at sun.nio.ch.SocketDispatcher.write(SocketDispatcher.java:29)
at sun.nio.ch.IOUtil.writeFromNativeBuffer(IOUtil.java:69)
at sun.nio.ch.IOUtil.write(IOUtil.java:40)
at sun.nio.ch.SocketChannelImpl.write(SocketChannelImpl.java:334)
at kafka.api.TopicDataSend.writeTo(FetchResponse.scala:131)
at kafka.network.MultiSend.writeTo(Transmission.scala:94)
at kafka.network.Send$class.writeCompletely(Transmission.scala:75)
at kafka.network.MultiSend.writeCompletely(Transmission.scala:87)
at kafka.api.FetchResponseSend.writeTo(FetchResponse.scala:232)
at kafka.network.Processor.write(SocketServer.scala:318)
at kafka.network.Processor.run(SocketServer.scala:211)
at java.lang.Thread.run(Thread.java:662)
[2012-11-08 07:10:56,851] ERROR [BrokerChangeListener on Controller 0]: Error while handling broker changes (kafka.controller.ReplicaStateMachine$BrokerChangeListener:102)
org.I0Itec.zkclient.exception.ZkInterruptedException: java.lang.InterruptedException
at org.I0Itec.zkclient.ZkClient.retryUntilConnected(ZkClient.java:687)
at org.I0Itec.zkclient.ZkClient.readData(ZkClient.java:766)
at org.I0Itec.zkclient.ZkClient.readData(ZkClient.java:761)
at kafka.utils.ZkUtils$.readDataMaybeNull(ZkUtils.scala:374)
at kafka.utils.ZkUtils$$anonfun$getPartitionAssignmentForTopics$1.apply(ZkUtils.scala:481)
at kafka.utils.ZkUtils$$anonfun$getPartitionAssignmentForTopics$1.apply(ZkUtils.scala:480)
at scala.collection.LinearSeqOptimized$class.foreach(LinearSeqOptimized.scala:61)
at scala.collection.immutable.List.foreach(List.scala:45)
at kafka.utils.ZkUtils$.getPartitionAssignmentForTopics(ZkUtils.scala:480)
at kafka.utils.ZkUtils$.getPartitionsAssignedToBroker(ZkUtils.scala:521)
at kafka.utils.ZkUtils$$anonfun$getAllReplicasOnBroker$1.apply(ZkUtils.scala:586)
at kafka.utils.ZkUtils$$anonfun$getAllReplicasOnBroker$1.apply(ZkUtils.scala:583)
at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
at scala.collection.LinearSeqOptimized$class.foreach(LinearSeqOptimized.scala:61)
at scala.collection.immutable.List.foreach(List.scala:45)
at scala.collection.TraversableLike$class.map(TraversableLike.scala:206)
at scala.collection.immutable.List.map(List.scala:45)
at kafka.utils.ZkUtils$.getAllReplicasOnBroker(ZkUtils.scala:583)
at kafka.controller.KafkaController.onBrokerStartup(KafkaController.scala:237)
at kafka.controller.ReplicaStateMachine$BrokerChangeListener$$anonfun$handleChildChange$1.liftedTree1$1(ReplicaStateMachine.scala:250)
at kafka.controller.ReplicaStateMachine$BrokerChangeListener$$anonfun$handleChildChange$1.apply$mcV$sp(ReplicaStateMachine.scala:239)
at kafka.controller.ReplicaStateMachine$BrokerChangeListener$$anonfun$handleChildChange$1.apply(ReplicaStateMachine.scala:235)
at kafka.controller.ReplicaStateMachine$BrokerChangeListener$$anonfun$handleChildChange$1.apply(ReplicaStateMachine.scala:235)
at kafka.metrics.KafkaTimer.time(KafkaTimer.scala:33)
at kafka.controller.ReplicaStateMachine$BrokerChangeListener.handleChildChange(ReplicaStateMachine.scala:235)
at org.I0Itec.zkclient.ZkClient$7.run(ZkClient.java:568)
at org.I0Itec.zkclient.ZkEventThread.run(ZkEventThread.java:71)
Caused by: java.lang.InterruptedException
at java.lang.Object.wait(Native Method)
at java.lang.Object.wait(Object.java:485)
at org.apache.zookeeper.ClientCnxn.submitRequest(ClientCnxn.java:1344)
at org.apache.zookeeper.ZooKeeper.getData(ZooKeeper.java:925)
at org.apache.zookeeper.ZooKeeper.getData(ZooKeeper.java:956)
at org.I0Itec.zkclient.ZkConnection.readData(ZkConnection.java:103)
at org.I0Itec.zkclient.ZkClient$9.call(ZkClient.java:770)
at org.I0Itec.zkclient.ZkClient$9.call(ZkClient.java:766)
at org.I0Itec.zkclient.ZkClient.retryUntilConnected(ZkClient.java:675)
... 27 more
[0m[[0minfo[0m] [0mTest Passed: testLeaderElectionAndEpoch(kafka.server.LeaderElectionTest)[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.server.LeaderElectionTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.log4j.KafkaLog4jAppenderTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testKafkaLog4jConfigs(kafka.log4j.KafkaLog4jAppenderTest)[0m
log4j:WARN No appenders could be found for logger (org.I0Itec.zkclient.ZkEventThread).
log4j:WARN Please initialize the log4j system properly.
[0m[[0minfo[0m] [0mTest Passed: testKafkaLog4jConfigs(kafka.log4j.KafkaLog4jAppenderTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testLog4jAppends(kafka.log4j.KafkaLog4jAppenderTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testLog4jAppends(kafka.log4j.KafkaLog4jAppenderTest)[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.log4j.KafkaLog4jAppenderTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.server.SimpleFetchTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testNonReplicaSeesHwWhenFetching(kafka.server.SimpleFetchTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testNonReplicaSeesHwWhenFetching(kafka.server.SimpleFetchTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testReplicaSeesLeoWhenFetching(kafka.server.SimpleFetchTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testReplicaSeesLeoWhenFetching(kafka.server.SimpleFetchTest)[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.server.SimpleFetchTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.integration.AutoOffsetResetTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testResetToEarliestWhenOffsetTooHigh(kafka.integration.AutoOffsetResetTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testResetToEarliestWhenOffsetTooHigh(kafka.integration.AutoOffsetResetTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testResetToEarliestWhenOffsetTooLow(kafka.integration.AutoOffsetResetTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testResetToEarliestWhenOffsetTooLow(kafka.integration.AutoOffsetResetTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testResetToLatestWhenOffsetTooHigh(kafka.integration.AutoOffsetResetTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testResetToLatestWhenOffsetTooHigh(kafka.integration.AutoOffsetResetTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testResetToLatestWhenOffsetTooLow(kafka.integration.AutoOffsetResetTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testResetToLatestWhenOffsetTooLow(kafka.integration.AutoOffsetResetTest)[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.integration.AutoOffsetResetTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.server.IsrExpirationTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testIsrExpirationForStuckFollowers(kafka.server.IsrExpirationTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testIsrExpirationForStuckFollowers(kafka.server.IsrExpirationTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testIsrExpirationForSlowFollowers(kafka.server.IsrExpirationTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testIsrExpirationForSlowFollowers(kafka.server.IsrExpirationTest)[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.server.IsrExpirationTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.server.ServerShutdownTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testCleanShutdown(kafka.server.ServerShutdownTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testCleanShutdown(kafka.server.ServerShutdownTest)[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.server.ServerShutdownTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.server.LogRecoveryTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testHWCheckpointNoFailuresSingleLogSegment(kafka.server.LogRecoveryTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testHWCheckpointNoFailuresSingleLogSegment(kafka.server.LogRecoveryTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testHWCheckpointWithFailuresSingleLogSegment(kafka.server.LogRecoveryTest)[0m
[0m[[31merror[0m] [0mTest Failed: testHWCheckpointWithFailuresSingleLogSegment(kafka.server.LogRecoveryTest)[0m
java.lang.AssertionError: expected:<2> but was:<1>
at org.junit.Assert.fail(Assert.java:69)
at org.junit.Assert.failNotEquals(Assert.java:314)
at org.junit.Assert.assertEquals(Assert.java:94)
at org.junit.Assert.assertEquals(Assert.java:104)
at kafka.server.LogRecoveryTest.testHWCheckpointWithFailuresSingleLogSegment(LogRecoveryTest.scala:149)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at junit.framework.TestCase.runTest(TestCase.java:164)
at junit.framework.TestCase.runBare(TestCase.java:130)
at junit.framework.TestResult$1.protect(TestResult.java:110)
at junit.framework.TestResult.runProtected(TestResult.java:128)
at junit.framework.TestResult.run(TestResult.java:113)
at junit.framework.TestCase.run(TestCase.java:120)
at junit.framework.TestSuite.runTest(TestSuite.java:228)
at junit.framework.TestSuite.run(TestSuite.java:223)
at junit.framework.TestSuite.runTest(TestSuite.java:228)
at junit.framework.TestSuite.run(TestSuite.java:223)
at org.scalatest.junit.JUnit3Suite.run(JUnit3Suite.scala:309)
at org.scalatest.tools.ScalaTestFramework$ScalaTestRunner.run(ScalaTestFramework.scala:40)
at sbt.TestRunner.run(TestFramework.scala:53)
at sbt.TestRunner.runTest$1(TestFramework.scala:67)
at sbt.TestRunner.run(TestFramework.scala:76)
at sbt.TestFramework$$anonfun$10$$anonfun$apply$11.runTest$2(TestFramework.scala:194)
at sbt.TestFramework$$anonfun$10$$anonfun$apply$11$$anonfun$apply$12.apply(TestFramework.scala:205)
at sbt.TestFramework$$anonfun$10$$anonfun$apply$11$$anonfun$apply$12.apply(TestFramework.scala:205)
at sbt.NamedTestTask.run(TestFramework.scala:92)
at sbt.ScalaProject$$anonfun$sbt$ScalaProject$$toTask$1.apply(ScalaProject.scala:193)
at sbt.ScalaProject$$anonfun$sbt$ScalaProject$$toTask$1.apply(ScalaProject.scala:193)
at sbt.TaskManager$Task.invoke(TaskManager.scala:62)
at sbt.impl.RunTask.doRun$1(RunTask.scala:77)
at sbt.impl.RunTask.runTask(RunTask.scala:85)
at sbt.impl.RunTask.sbt$impl$RunTask$$runIfNotRoot(RunTask.scala:60)
at sbt.impl.RunTask$$anonfun$runTasksExceptRoot$2.apply(RunTask.scala:48)
at sbt.impl.RunTask$$anonfun$runTasksExceptRoot$2.apply(RunTask.scala:48)
at sbt.Distributor$Run$Worker$$anonfun$2.apply(ParallelRunner.scala:131)
at sbt.Distributor$Run$Worker$$anonfun$2.apply(ParallelRunner.scala:131)
at sbt.Control$.trapUnit(Control.scala:19)
at sbt.Distributor$Run$Worker.run(ParallelRunner.scala:131)
[0m[[0minfo[0m] [0mTest Starting: testHWCheckpointNoFailuresMultipleLogSegments(kafka.server.LogRecoveryTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testHWCheckpointNoFailuresMultipleLogSegments(kafka.server.LogRecoveryTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testHWCheckpointWithFailuresMultipleLogSegments(kafka.server.LogRecoveryTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testHWCheckpointWithFailuresMultipleLogSegments(kafka.server.LogRecoveryTest)[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.server.LogRecoveryTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / test-finish ==[0m
[0m[[31merror[0m] [0mFailed: : Total 165, Failed 2, Errors 0, Passed 163, Skipped 0[0m
[0m[[0minfo[0m] [34m== core-kafka / test-finish ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / Test cleanup 1 ==[0m
[0m[[0minfo[0m] [0mDeleting directory /tmp/sbt_268a9bb5[0m
[0m[[0minfo[0m] [34m== core-kafka / Test cleanup 1 ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / test-cleanup ==[0m
[0m[[0minfo[0m] [34m== core-kafka / test-cleanup ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== perf / copy-resources ==[0m
[0m[[0minfo[0m] [34m== perf / copy-resources ==[0m
[0m[[31merror[0m] [0mError running kafka.integration.LazyInitProducerTest: Test FAILED[0m
[0m[[31merror[0m] [0mError running kafka.server.LogRecoveryTest: Test FAILED[0m
[0m[[31merror[0m] [0mError running test: One or more subtasks failed[0m
[0m[[0minfo[0m] [0m[0m
[0m[[0minfo[0m] [0mTotal time: 257 s, completed Nov 8, 2012 7:11:33 AM[0m
[0m[[0minfo[0m] [0m[0m
[0m[[0minfo[0m] [0mTotal session time: 258 s, completed Nov 8, 2012 7:11:33 AM[0m
[0m[[31merror[0m] [0mError during build.[0m
Build step 'Execute shell' marked build as failure