You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@kafka.apache.org by Apache Jenkins Server <je...@builds.apache.org> on 2012/07/26 02:40:46 UTC
Build failed in Jenkins: Kafka-0.8 #10
See <https://builds.apache.org/job/Kafka-0.8/10/changes>
Changes:
[nehanarkhede] KAFKA-405 Improve high watermark maintenance to store high watermarks for all partitions in a single .highwatermark file; patched by Neha Narkhede; reviewed by Jay Kreps and Jun Rao
------------------------------------------
[...truncated 895 lines...]
at org.junit.internal.runners.BeforeAndAfterRunner.runProtected(BeforeAndAfterRunner.java:34)
at org.junit.internal.runners.TestMethodRunner.runMethod(TestMethodRunner.java:75)
at org.junit.internal.runners.TestMethodRunner.run(TestMethodRunner.java:45)
at org.junit.internal.runners.TestClassMethodsRunner.invokeTestMethod(TestClassMethodsRunner.java:71)
at org.junit.internal.runners.TestClassMethodsRunner.run(TestClassMethodsRunner.java:35)
at org.junit.internal.runners.TestClassRunner$1.runUnprotected(TestClassRunner.java:42)
at org.junit.internal.runners.BeforeAndAfterRunner.runProtected(BeforeAndAfterRunner.java:34)
at org.junit.internal.runners.TestClassRunner.run(TestClassRunner.java:52)
at org.junit.internal.runners.CompositeRunner.run(CompositeRunner.java:29)
at org.junit.runner.JUnitCore.run(JUnitCore.java:121)
at org.junit.runner.JUnitCore.run(JUnitCore.java:100)
at org.junit.runner.JUnitCore.run(JUnitCore.java:91)
at org.scalatest.junit.JUnitSuite$class.run(JUnitSuite.scala:261)
at kafka.message.CompressionUtilTest.run(CompressionUtilsTest.scala:25)
at org.scalatest.tools.ScalaTestFramework$ScalaTestRunner.run(ScalaTestFramework.scala:40)
at sbt.TestRunner.run(TestFramework.scala:53)
at sbt.TestRunner.runTest$1(TestFramework.scala:67)
at sbt.TestRunner.run(TestFramework.scala:76)
at sbt.TestFramework$$anonfun$10$$anonfun$apply$11.runTest$2(TestFramework.scala:194)
at sbt.TestFramework$$anonfun$10$$anonfun$apply$11$$anonfun$apply$12.apply(TestFramework.scala:205)
at sbt.TestFramework$$anonfun$10$$anonfun$apply$11$$anonfun$apply$12.apply(TestFramework.scala:205)
at sbt.NamedTestTask.run(TestFramework.scala:92)
at sbt.ScalaProject$$anonfun$sbt$ScalaProject$$toTask$1.apply(ScalaProject.scala:193)
at sbt.ScalaProject$$anonfun$sbt$ScalaProject$$toTask$1.apply(ScalaProject.scala:193)
at sbt.TaskManager$Task.invoke(TaskManager.scala:62)
at sbt.impl.RunTask.doRun$1(RunTask.scala:77)
at sbt.impl.RunTask.runTask(RunTask.scala:85)
at sbt.impl.RunTask.sbt$impl$RunTask$$runIfNotRoot(RunTask.scala:60)
at sbt.impl.RunTask$$anonfun$runTasksExceptRoot$2.apply(RunTask.scala:48)
at sbt.impl.RunTask$$anonfun$runTasksExceptRoot$2.apply(RunTask.scala:48)
at sbt.Distributor$Run$Worker$$anonfun$2.apply(ParallelRunner.scala:131)
at sbt.Distributor$Run$Worker$$anonfun$2.apply(ParallelRunner.scala:131)
at sbt.Control$.trapUnit(Control.scala:19)
at sbt.Distributor$Run$Worker.run(ParallelRunner.scala:131)
Caused by: java.lang.UnsatisfiedLinkError: no snappyjava in java.library.path
at java.lang.ClassLoader.loadLibrary(ClassLoader.java:1734)
at java.lang.Runtime.loadLibrary0(Runtime.java:823)
at java.lang.System.loadLibrary(System.java:1028)
at org.xerial.snappy.SnappyNativeLoader.loadLibrary(SnappyNativeLoader.java:52)
... 53 more
[0m[[31merror[0m] [0mTest Failed: testSnappyCompressDecompressExplicit[0m
org.xerial.snappy.SnappyError: [FAILED_TO_LOAD_NATIVE_LIBRARY] null
at org.xerial.snappy.SnappyLoader.load(SnappyLoader.java:229)
at org.xerial.snappy.Snappy.<clinit>(Snappy.java:44)
at org.xerial.snappy.SnappyOutputStream.<init>(SnappyOutputStream.java:79)
at org.xerial.snappy.SnappyOutputStream.<init>(SnappyOutputStream.java:66)
at kafka.message.SnappyCompression.<init>(CompressionUtils.scala:61)
at kafka.message.CompressionFactory$.apply(CompressionUtils.scala:82)
at kafka.message.CompressionUtils$.compress(CompressionUtils.scala:109)
at kafka.message.CompressionUtilTest.testSnappyCompressDecompressExplicit(CompressionUtilsTest.scala:65)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at org.junit.internal.runners.TestMethodRunner.executeMethodBody(TestMethodRunner.java:99)
at org.junit.internal.runners.TestMethodRunner.runUnprotected(TestMethodRunner.java:81)
at org.junit.internal.runners.BeforeAndAfterRunner.runProtected(BeforeAndAfterRunner.java:34)
at org.junit.internal.runners.TestMethodRunner.runMethod(TestMethodRunner.java:75)
at org.junit.internal.runners.TestMethodRunner.run(TestMethodRunner.java:45)
at org.junit.internal.runners.TestClassMethodsRunner.invokeTestMethod(TestClassMethodsRunner.java:71)
at org.junit.internal.runners.TestClassMethodsRunner.run(TestClassMethodsRunner.java:35)
at org.junit.internal.runners.TestClassRunner$1.runUnprotected(TestClassRunner.java:42)
at org.junit.internal.runners.BeforeAndAfterRunner.runProtected(BeforeAndAfterRunner.java:34)
at org.junit.internal.runners.TestClassRunner.run(TestClassRunner.java:52)
at org.junit.internal.runners.CompositeRunner.run(CompositeRunner.java:29)
at org.junit.runner.JUnitCore.run(JUnitCore.java:121)
at org.junit.runner.JUnitCore.run(JUnitCore.java:100)
at org.junit.runner.JUnitCore.run(JUnitCore.java:91)
at org.scalatest.junit.JUnitSuite$class.run(JUnitSuite.scala:261)
at kafka.message.CompressionUtilTest.run(CompressionUtilsTest.scala:25)
at org.scalatest.tools.ScalaTestFramework$ScalaTestRunner.run(ScalaTestFramework.scala:40)
at sbt.TestRunner.run(TestFramework.scala:53)
at sbt.TestRunner.runTest$1(TestFramework.scala:67)
at sbt.TestRunner.run(TestFramework.scala:76)
at sbt.TestFramework$$anonfun$10$$anonfun$apply$11.runTest$2(TestFramework.scala:194)
at sbt.TestFramework$$anonfun$10$$anonfun$apply$11$$anonfun$apply$12.apply(TestFramework.scala:205)
at sbt.TestFramework$$anonfun$10$$anonfun$apply$11$$anonfun$apply$12.apply(TestFramework.scala:205)
at sbt.NamedTestTask.run(TestFramework.scala:92)
at sbt.ScalaProject$$anonfun$sbt$ScalaProject$$toTask$1.apply(ScalaProject.scala:193)
at sbt.ScalaProject$$anonfun$sbt$ScalaProject$$toTask$1.apply(ScalaProject.scala:193)
at sbt.TaskManager$Task.invoke(TaskManager.scala:62)
at sbt.impl.RunTask.doRun$1(RunTask.scala:77)
at sbt.impl.RunTask.runTask(RunTask.scala:85)
at sbt.impl.RunTask.sbt$impl$RunTask$$runIfNotRoot(RunTask.scala:60)
at sbt.impl.RunTask$$anonfun$runTasksExceptRoot$2.apply(RunTask.scala:48)
at sbt.impl.RunTask$$anonfun$runTasksExceptRoot$2.apply(RunTask.scala:48)
at sbt.Distributor$Run$Worker$$anonfun$2.apply(ParallelRunner.scala:131)
at sbt.Distributor$Run$Worker$$anonfun$2.apply(ParallelRunner.scala:131)
at sbt.Control$.trapUnit(Control.scala:19)
at sbt.Distributor$Run$Worker.run(ParallelRunner.scala:131)
[0m[[0minfo[0m] [34m== core-kafka / kafka.message.CompressionUtilTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.server.ServerShutdownTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testCleanShutdown(kafka.server.ServerShutdownTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testCleanShutdown(kafka.server.ServerShutdownTest)[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.server.ServerShutdownTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.javaapi.message.ByteBufferMessageSetTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testWrittenEqualsRead[0m
[0m[[0minfo[0m] [0mTest Passed: testWrittenEqualsRead[0m
[0m[[0minfo[0m] [0mTest Starting: testIteratorIsConsistent[0m
[0m[[0minfo[0m] [0mTest Passed: testIteratorIsConsistent[0m
[0m[[0minfo[0m] [0mTest Starting: testSizeInBytes[0m
[0m[[0minfo[0m] [0mTest Passed: testSizeInBytes[0m
[0m[[0minfo[0m] [0mTest Starting: testIteratorIsConsistentWithCompression[0m
[0m[[0minfo[0m] [0mTest Passed: testIteratorIsConsistentWithCompression[0m
[0m[[0minfo[0m] [0mTest Starting: testSizeInBytesWithCompression[0m
[0m[[0minfo[0m] [0mTest Passed: testSizeInBytesWithCompression[0m
[0m[[0minfo[0m] [0mTest Starting: testValidBytes[0m
[0m[[0minfo[0m] [0mTest Passed: testValidBytes[0m
[0m[[0minfo[0m] [0mTest Starting: testValidBytesWithCompression[0m
[0m[[0minfo[0m] [0mTest Passed: testValidBytesWithCompression[0m
[0m[[0minfo[0m] [0mTest Starting: testEquals[0m
[0m[[0minfo[0m] [0mTest Passed: testEquals[0m
[0m[[0minfo[0m] [0mTest Starting: testEqualsWithCompression[0m
[0m[[0minfo[0m] [0mTest Passed: testEqualsWithCompression[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.javaapi.message.ByteBufferMessageSetTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.producer.SyncProducerTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testReachableServer(kafka.producer.SyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testReachableServer(kafka.producer.SyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testSingleMessageSizeTooLarge(kafka.producer.SyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testSingleMessageSizeTooLarge(kafka.producer.SyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testCompressedMessageSizeTooLarge(kafka.producer.SyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testCompressedMessageSizeTooLarge(kafka.producer.SyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testProduceCorrectlyReceivesResponse(kafka.producer.SyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testProduceCorrectlyReceivesResponse(kafka.producer.SyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testProducerCanTimeout(kafka.producer.SyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testProducerCanTimeout(kafka.producer.SyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testProduceRequestForUnknownTopic(kafka.producer.SyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testProduceRequestForUnknownTopic(kafka.producer.SyncProducerTest)[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.producer.SyncProducerTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.log.LogCorruptionTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testMessageSizeTooLarge(kafka.log.LogCorruptionTest)[0m
This is good
[0m[[0minfo[0m] [0mTest Passed: testMessageSizeTooLarge(kafka.log.LogCorruptionTest)[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.log.LogCorruptionTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.network.SocketServerTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: simpleRequest[0m
[0m[[0minfo[0m] [0mTest Passed: simpleRequest[0m
[0m[[0minfo[0m] [0mTest Starting: tooBigRequestIsRejected[0m
[0m[[0minfo[0m] [0mTest Passed: tooBigRequestIsRejected[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.network.SocketServerTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.controller.ControllerToBrokerRequestTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testLeaderAndISRRequest(kafka.controller.ControllerToBrokerRequestTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testLeaderAndISRRequest(kafka.controller.ControllerToBrokerRequestTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testLeaderAndISRResponse(kafka.controller.ControllerToBrokerRequestTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testLeaderAndISRResponse(kafka.controller.ControllerToBrokerRequestTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testStopReplicaRequest(kafka.controller.ControllerToBrokerRequestTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testStopReplicaRequest(kafka.controller.ControllerToBrokerRequestTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testStopReplicaResponse(kafka.controller.ControllerToBrokerRequestTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testStopReplicaResponse(kafka.controller.ControllerToBrokerRequestTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testEndToEndLeaderAndISRRequest(kafka.controller.ControllerToBrokerRequestTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testEndToEndLeaderAndISRRequest(kafka.controller.ControllerToBrokerRequestTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testEndToEndStopReplicaRequest(kafka.controller.ControllerToBrokerRequestTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testEndToEndStopReplicaRequest(kafka.controller.ControllerToBrokerRequestTest)[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.controller.ControllerToBrokerRequestTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.consumer.TopicCountTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testPartition[0m
[0m[[0minfo[0m] [0mTest Passed: testPartition[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.consumer.TopicCountTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.controller.ControllerBasicTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testControllerFailOver(kafka.controller.ControllerBasicTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testControllerFailOver(kafka.controller.ControllerBasicTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testControllerCommandSend(kafka.controller.ControllerBasicTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testControllerCommandSend(kafka.controller.ControllerBasicTest)[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.controller.ControllerBasicTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.message.ByteBufferMessageSetTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testWrittenEqualsRead[0m
[0m[[0minfo[0m] [0mTest Passed: testWrittenEqualsRead[0m
[0m[[0minfo[0m] [0mTest Starting: testIteratorIsConsistent[0m
[0m[[0minfo[0m] [0mTest Passed: testIteratorIsConsistent[0m
[0m[[0minfo[0m] [0mTest Starting: testSizeInBytes[0m
[0m[[0minfo[0m] [0mTest Passed: testSizeInBytes[0m
[0m[[0minfo[0m] [0mTest Starting: testWriteTo[0m
[0m[[0minfo[0m] [0mTest Passed: testWriteTo[0m
[0m[[0minfo[0m] [0mTest Starting: testValidBytes[0m
[0m[[0minfo[0m] [0mTest Passed: testValidBytes[0m
[0m[[0minfo[0m] [0mTest Starting: testEquals[0m
[0m[[0minfo[0m] [0mTest Passed: testEquals[0m
[0m[[0minfo[0m] [0mTest Starting: testSmallFetchSize[0m
[0m[[0minfo[0m] [0mTest Passed: testSmallFetchSize[0m
[0m[[0minfo[0m] [0mTest Starting: testIterator[0m
[0m[[0minfo[0m] [0mTest Passed: testIterator[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.message.ByteBufferMessageSetTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.message.MessageTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testFieldValues[0m
[0m[[0minfo[0m] [0mTest Passed: testFieldValues[0m
[0m[[0minfo[0m] [0mTest Starting: testChecksum[0m
[0m[[0minfo[0m] [0mTest Passed: testChecksum[0m
[0m[[0minfo[0m] [0mTest Starting: testEquality[0m
[0m[[0minfo[0m] [0mTest Passed: testEquality[0m
[0m[[0minfo[0m] [0mTest Starting: testIsHashable[0m
[0m[[0minfo[0m] [0mTest Passed: testIsHashable[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.message.MessageTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.integration.PrimitiveApiTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testProduceAndMultiFetch(kafka.integration.PrimitiveApiTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testProduceAndMultiFetch(kafka.integration.PrimitiveApiTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testMultiProduce(kafka.integration.PrimitiveApiTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testMultiProduce(kafka.integration.PrimitiveApiTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testFetchRequestCanProperlySerialize(kafka.integration.PrimitiveApiTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testFetchRequestCanProperlySerialize(kafka.integration.PrimitiveApiTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testFetchRequestEnforcesUniqueTopicsForOffsetDetails(kafka.integration.PrimitiveApiTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testFetchRequestEnforcesUniqueTopicsForOffsetDetails(kafka.integration.PrimitiveApiTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testDefaultEncoderProducerAndFetch(kafka.integration.PrimitiveApiTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testDefaultEncoderProducerAndFetch(kafka.integration.PrimitiveApiTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testDefaultEncoderProducerAndFetchWithCompression(kafka.integration.PrimitiveApiTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testDefaultEncoderProducerAndFetchWithCompression(kafka.integration.PrimitiveApiTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testProduceAndMultiFetchWithCompression(kafka.integration.PrimitiveApiTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testProduceAndMultiFetchWithCompression(kafka.integration.PrimitiveApiTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testMultiProduceWithCompression(kafka.integration.PrimitiveApiTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testMultiProduceWithCompression(kafka.integration.PrimitiveApiTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testConsumerNotExistTopic(kafka.integration.PrimitiveApiTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testConsumerNotExistTopic(kafka.integration.PrimitiveApiTest)[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.integration.PrimitiveApiTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / Test cleanup 1 ==[0m
[0m[[0minfo[0m] [0mDeleting directory /var/tmp/sbt_292d6150[0m
[0m[[0minfo[0m] [34m== core-kafka / Test cleanup 1 ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / test-finish ==[0m
[0m[[31merror[0m] [0mFailed: : Total 145, Failed 1, Errors 0, Passed 144, Skipped 0[0m
[0m[[0minfo[0m] [34m== core-kafka / test-finish ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / test-cleanup ==[0m
[0m[[0minfo[0m] [34m== core-kafka / test-cleanup ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== perf / test-compile ==[0m
[0m[[0minfo[0m] [0m Source analysis: 0 new/modified, 0 indirectly invalidated, 0 removed.[0m
[0m[[0minfo[0m] [0mCompiling test sources...[0m
[0m[[0minfo[0m] [0mNothing to compile.[0m
[0m[[0minfo[0m] [0m Post-analysis: 0 classes.[0m
[0m[[0minfo[0m] [34m== perf / test-compile ==[0m
[0m[[31merror[0m] [0mError running kafka.message.CompressionUtilTest: Test FAILED[0m
[0m[[31merror[0m] [0mError running test: One or more subtasks failed[0m
[0m[[0minfo[0m] [0m[0m
[0m[[0minfo[0m] [0mTotal time: 233 s, completed Jul 26, 2012 12:40:44 AM[0m
[0m[[0minfo[0m] [0m[0m
[0m[[0minfo[0m] [0mTotal session time: 234 s, completed Jul 26, 2012 12:40:44 AM[0m
[0m[[31merror[0m] [0mError during build.[0m
Build step 'Execute shell' marked build as failure
Jenkins build is back to normal : Kafka-0.8 #12
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/Kafka-0.8/12/changes>
Build failed in Jenkins: Kafka-0.8 #11
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/Kafka-0.8/11/changes>
Changes:
[junrao] single_host_multi_brokers system test fails on laptop; patched by John Fung; reviewed by Jun Rao; kafka-413
------------------------------------------
[...truncated 2995 lines...]
at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:27)
at java.lang.reflect.Constructor.newInstance(Constructor.java:513)
at java.lang.Class.newInstance0(Class.java:355)
at java.lang.Class.newInstance(Class.java:308)
at kafka.common.ErrorMapping$.maybeThrowException(ErrorMapping.scala:66)
at kafka.producer.BrokerPartitionInfo$$anonfun$updateInfo$1$$anonfun$apply$5$$anonfun$apply$6.apply(BrokerPartitionInfo.scala:85)
at kafka.producer.BrokerPartitionInfo$$anonfun$updateInfo$1$$anonfun$apply$5$$anonfun$apply$6.apply(BrokerPartitionInfo.scala:85)
at scala.collection.LinearSeqOptimized$class.foreach(LinearSeqOptimized.scala:61)
at scala.collection.immutable.List.foreach(List.scala:45)
at scala.collection.generic.TraversableForwarder$class.foreach(TraversableForwarder.scala:44)
at scala.collection.mutable.ListBuffer.foreach(ListBuffer.scala:42)
at kafka.producer.BrokerPartitionInfo$$anonfun$updateInfo$1$$anonfun$apply$5.apply(BrokerPartitionInfo.scala:85)
at kafka.producer.BrokerPartitionInfo$$anonfun$updateInfo$1$$anonfun$apply$5.apply(BrokerPartitionInfo.scala:84)
at scala.collection.IndexedSeqOptimized$class.foreach(IndexedSeqOptimized.scala:34)
at scala.collection.mutable.WrappedArray.foreach(WrappedArray.scala:32)
at kafka.producer.BrokerPartitionInfo$$anonfun$updateInfo$1.apply(BrokerPartitionInfo.scala:84)
at kafka.producer.BrokerPartitionInfo$$anonfun$updateInfo$1.apply(BrokerPartitionInfo.scala:75)
at scala.collection.LinearSeqOptimized$class.foreach(LinearSeqOptimized.scala:61)
at scala.collection.immutable.List.foreach(List.scala:45)
at kafka.producer.BrokerPartitionInfo.updateInfo(BrokerPartitionInfo.scala:75)
at kafka.producer.async.DefaultEventHandler$$anonfun$handle$1.apply$mcV$sp(DefaultEventHandler.scala:54)
at kafka.utils.Utils$.swallow(Utils.scala:428)
at kafka.utils.Logging$class.swallowError(Logging.scala:102)
at kafka.utils.Utils$.swallowError(Utils.scala:39)
at kafka.producer.async.DefaultEventHandler.handle(DefaultEventHandler.scala:54)
at kafka.producer.Producer.send(Producer.scala:70)
at kafka.integration.LazyInitProducerTest.testMultiProduceResend(LazyInitProducerTest.scala:165)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at junit.framework.TestCase.runTest(TestCase.java:164)
at junit.framework.TestCase.runBare(TestCase.java:130)
at junit.framework.TestResult$1.protect(TestResult.java:110)
at junit.framework.TestResult.runProtected(TestResult.java:128)
at junit.framework.TestResult.run(TestResult.java:113)
at junit.framework.TestCase.run(TestCase.java:120)
at junit.framework.TestSuite.runTest(TestSuite.java:228)
at junit.framework.TestSuite.run(TestSuite.java:223)
at junit.framework.TestSuite.runTest(TestSuite.java:228)
at junit.framework.TestSuite.run(TestSuite.java:223)
at org.scalatest.junit.JUnit3Suite.run(JUnit3Suite.scala:309)
at org.scalatest.tools.ScalaTestFramework$ScalaTestRunner.run(ScalaTestFramework.scala:40)
at sbt.TestRunner.run(TestFramework.scala:53)
at sbt.TestRunner.runTest$1(TestFramework.scala:67)
at sbt.TestRunner.run(TestFramework.scala:76)
at sbt.TestFramework$$anonfun$10$$anonfun$apply$11.runTest$2(TestFramework.scala:194)
at sbt.TestFramework$$anonfun$10$$anonfun$apply$11$$anonfun$apply$12.apply(TestFramework.scala:205)
at sbt.TestFramework$$anonfun$10$$anonfun$apply$11$$anonfun$apply$12.apply(TestFramework.scala:205)
at sbt.NamedTestTask.run(TestFramework.scala:92)
at sbt.ScalaProject$$anonfun$sbt$ScalaProject$$toTask$1.apply(ScalaProject.scala:193)
at sbt.ScalaProject$$anonfun$sbt$ScalaProject$$toTask$1.apply(ScalaProject.scala:193)
at sbt.TaskManager$Task.invoke(TaskManager.scala:62)
at sbt.impl.RunTask.doRun$1(RunTask.scala:77)
at sbt.impl.RunTask.runTask(RunTask.scala:85)
at sbt.impl.RunTask.sbt$impl$RunTask$$runIfNotRoot(RunTask.scala:60)
at sbt.impl.RunTask$$anonfun$runTasksExceptRoot$2.apply(RunTask.scala:48)
at sbt.impl.RunTask$$anonfun$runTasksExceptRoot$2.apply(RunTask.scala:48)
at sbt.Distributor$Run$Worker$$anonfun$2.apply(ParallelRunner.scala:131)
at sbt.Distributor$Run$Worker$$anonfun$2.apply(ParallelRunner.scala:131)
at sbt.Control$.trapUnit(Control.scala:19)
at sbt.Distributor$Run$Worker.run(ParallelRunner.scala:131)
[0m[[0minfo[0m] [0mTest Passed: testMultiProduceResend(kafka.integration.LazyInitProducerTest)[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.integration.LazyInitProducerTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.consumer.TopicFilterTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testWhitelists[0m
[0m[[0minfo[0m] [0mTest Passed: testWhitelists[0m
[0m[[0minfo[0m] [0mTest Starting: testBlacklists[0m
[0m[[0minfo[0m] [0mTest Passed: testBlacklists[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.consumer.TopicFilterTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.server.ServerShutdownTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testCleanShutdown(kafka.server.ServerShutdownTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testCleanShutdown(kafka.server.ServerShutdownTest)[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.server.ServerShutdownTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.javaapi.message.ByteBufferMessageSetTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testWrittenEqualsRead[0m
[2012-07-26 20:13:17,638] WARN Session 0x138c4eb58740013 for server null, unexpected error, closing socket connection and attempting reconnect (org.apache.zookeeper.ClientCnxn:1188)
java.net.ConnectException: Connection refused
at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:567)
at org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1146)
[0m[[0minfo[0m] [0mTest Passed: testWrittenEqualsRead[0m
[0m[[0minfo[0m] [0mTest Starting: testIteratorIsConsistent[0m
[0m[[0minfo[0m] [0mTest Passed: testIteratorIsConsistent[0m
[0m[[0minfo[0m] [0mTest Starting: testSizeInBytes[0m
[0m[[0minfo[0m] [0mTest Passed: testSizeInBytes[0m
[0m[[0minfo[0m] [0mTest Starting: testValidBytes[0m
[0m[[0minfo[0m] [0mTest Passed: testValidBytes[0m
[0m[[0minfo[0m] [0mTest Starting: testEquals[0m
[0m[[0minfo[0m] [0mTest Passed: testEquals[0m
[0m[[0minfo[0m] [0mTest Starting: testIteratorIsConsistentWithCompression[0m
[0m[[0minfo[0m] [0mTest Passed: testIteratorIsConsistentWithCompression[0m
[0m[[0minfo[0m] [0mTest Starting: testSizeInBytesWithCompression[0m
[0m[[0minfo[0m] [0mTest Passed: testSizeInBytesWithCompression[0m
[0m[[0minfo[0m] [0mTest Starting: testValidBytesWithCompression[0m
[0m[[0minfo[0m] [0mTest Passed: testValidBytesWithCompression[0m
[0m[[0minfo[0m] [0mTest Starting: testEqualsWithCompression[0m
[0m[[0minfo[0m] [0mTest Passed: testEqualsWithCompression[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.javaapi.message.ByteBufferMessageSetTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.log4j.KafkaLog4jAppenderTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testKafkaLog4jConfigs(kafka.log4j.KafkaLog4jAppenderTest)[0m
log4j:WARN No appenders could be found for logger (org.I0Itec.zkclient.ZkEventThread).
log4j:WARN Please initialize the log4j system properly.
[0m[[0minfo[0m] [0mTest Passed: testKafkaLog4jConfigs(kafka.log4j.KafkaLog4jAppenderTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testZkConnectLog4jAppends(kafka.log4j.KafkaLog4jAppenderTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testZkConnectLog4jAppends(kafka.log4j.KafkaLog4jAppenderTest)[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.log4j.KafkaLog4jAppenderTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.message.CompressionUtilTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testSimpleCompressDecompress[0m
[0m[[0minfo[0m] [0mTest Passed: testSimpleCompressDecompress[0m
[0m[[0minfo[0m] [0mTest Starting: testComplexCompressDecompress[0m
[0m[[0minfo[0m] [0mTest Passed: testComplexCompressDecompress[0m
[0m[[0minfo[0m] [0mTest Starting: testSnappyCompressDecompressExplicit[0m
[0m[[0minfo[0m] [0mTest Passed: testSnappyCompressDecompressExplicit[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.message.CompressionUtilTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.javaapi.consumer.ZookeeperConsumerConnectorTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testBasic(kafka.javaapi.consumer.ZookeeperConsumerConnectorTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testBasic(kafka.javaapi.consumer.ZookeeperConsumerConnectorTest)[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.javaapi.consumer.ZookeeperConsumerConnectorTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.producer.AsyncProducerTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testProducerQueueSize(kafka.producer.AsyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testProducerQueueSize(kafka.producer.AsyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testProduceAfterClosed(kafka.producer.AsyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testProduceAfterClosed(kafka.producer.AsyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testBatchSize(kafka.producer.AsyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testBatchSize(kafka.producer.AsyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testQueueTimeExpired(kafka.producer.AsyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testQueueTimeExpired(kafka.producer.AsyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testPartitionAndCollateEvents(kafka.producer.AsyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testPartitionAndCollateEvents(kafka.producer.AsyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testSerializeEvents(kafka.producer.AsyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testSerializeEvents(kafka.producer.AsyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testInvalidPartition(kafka.producer.AsyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testInvalidPartition(kafka.producer.AsyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testNoBroker(kafka.producer.AsyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testNoBroker(kafka.producer.AsyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testIncompatibleEncoder(kafka.producer.AsyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testIncompatibleEncoder(kafka.producer.AsyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testRandomPartitioner(kafka.producer.AsyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testRandomPartitioner(kafka.producer.AsyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testBrokerListAndAsync(kafka.producer.AsyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testBrokerListAndAsync(kafka.producer.AsyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testFailedSendRetryLogic(kafka.producer.AsyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testFailedSendRetryLogic(kafka.producer.AsyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testJavaProducer(kafka.producer.AsyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testJavaProducer(kafka.producer.AsyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testInvalidConfiguration(kafka.producer.AsyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testInvalidConfiguration(kafka.producer.AsyncProducerTest)[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.producer.AsyncProducerTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.controller.ControllerToBrokerRequestTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testLeaderAndISRRequest(kafka.controller.ControllerToBrokerRequestTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testLeaderAndISRRequest(kafka.controller.ControllerToBrokerRequestTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testLeaderAndISRResponse(kafka.controller.ControllerToBrokerRequestTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testLeaderAndISRResponse(kafka.controller.ControllerToBrokerRequestTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testStopReplicaRequest(kafka.controller.ControllerToBrokerRequestTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testStopReplicaRequest(kafka.controller.ControllerToBrokerRequestTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testStopReplicaResponse(kafka.controller.ControllerToBrokerRequestTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testStopReplicaResponse(kafka.controller.ControllerToBrokerRequestTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testEndToEndLeaderAndISRRequest(kafka.controller.ControllerToBrokerRequestTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testEndToEndLeaderAndISRRequest(kafka.controller.ControllerToBrokerRequestTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testEndToEndStopReplicaRequest(kafka.controller.ControllerToBrokerRequestTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testEndToEndStopReplicaRequest(kafka.controller.ControllerToBrokerRequestTest)[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.controller.ControllerToBrokerRequestTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.log.LogCorruptionTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testMessageSizeTooLarge(kafka.log.LogCorruptionTest)[0m
This is good
[0m[[0minfo[0m] [0mTest Passed: testMessageSizeTooLarge(kafka.log.LogCorruptionTest)[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.log.LogCorruptionTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.server.LogRecoveryTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testHWCheckpointNoFailuresSingleLogSegment(kafka.server.LogRecoveryTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testHWCheckpointNoFailuresSingleLogSegment(kafka.server.LogRecoveryTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testHWCheckpointWithFailuresSingleLogSegment(kafka.server.LogRecoveryTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testHWCheckpointWithFailuresSingleLogSegment(kafka.server.LogRecoveryTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testHWCheckpointNoFailuresMultipleLogSegments(kafka.server.LogRecoveryTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testHWCheckpointNoFailuresMultipleLogSegments(kafka.server.LogRecoveryTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testHWCheckpointWithFailuresMultipleLogSegments(kafka.server.LogRecoveryTest)[0m
[0m[[31merror[0m] [0mTest Failed: testHWCheckpointWithFailuresMultipleLogSegments(kafka.server.LogRecoveryTest)[0m
java.lang.AssertionError: expected:<120> but was:<150>
at org.junit.Assert.fail(Assert.java:69)
at org.junit.Assert.failNotEquals(Assert.java:314)
at org.junit.Assert.assertEquals(Assert.java:94)
at org.junit.Assert.assertEquals(Assert.java:104)
at kafka.server.LogRecoveryTest.testHWCheckpointWithFailuresMultipleLogSegments(LogRecoveryTest.scala:239)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at junit.framework.TestCase.runTest(TestCase.java:164)
at junit.framework.TestCase.runBare(TestCase.java:130)
at junit.framework.TestResult$1.protect(TestResult.java:110)
at junit.framework.TestResult.runProtected(TestResult.java:128)
at junit.framework.TestResult.run(TestResult.java:113)
at junit.framework.TestCase.run(TestCase.java:120)
at junit.framework.TestSuite.runTest(TestSuite.java:228)
at junit.framework.TestSuite.run(TestSuite.java:223)
at junit.framework.TestSuite.runTest(TestSuite.java:228)
at junit.framework.TestSuite.run(TestSuite.java:223)
at org.scalatest.junit.JUnit3Suite.run(JUnit3Suite.scala:309)
at org.scalatest.tools.ScalaTestFramework$ScalaTestRunner.run(ScalaTestFramework.scala:40)
at sbt.TestRunner.run(TestFramework.scala:53)
at sbt.TestRunner.runTest$1(TestFramework.scala:67)
at sbt.TestRunner.run(TestFramework.scala:76)
at sbt.TestFramework$$anonfun$10$$anonfun$apply$11.runTest$2(TestFramework.scala:194)
at sbt.TestFramework$$anonfun$10$$anonfun$apply$11$$anonfun$apply$12.apply(TestFramework.scala:205)
at sbt.TestFramework$$anonfun$10$$anonfun$apply$11$$anonfun$apply$12.apply(TestFramework.scala:205)
at sbt.NamedTestTask.run(TestFramework.scala:92)
at sbt.ScalaProject$$anonfun$sbt$ScalaProject$$toTask$1.apply(ScalaProject.scala:193)
at sbt.ScalaProject$$anonfun$sbt$ScalaProject$$toTask$1.apply(ScalaProject.scala:193)
at sbt.TaskManager$Task.invoke(TaskManager.scala:62)
at sbt.impl.RunTask.doRun$1(RunTask.scala:77)
at sbt.impl.RunTask.runTask(RunTask.scala:85)
at sbt.impl.RunTask.sbt$impl$RunTask$$runIfNotRoot(RunTask.scala:60)
at sbt.impl.RunTask$$anonfun$runTasksExceptRoot$2.apply(RunTask.scala:48)
at sbt.impl.RunTask$$anonfun$runTasksExceptRoot$2.apply(RunTask.scala:48)
at sbt.Distributor$Run$Worker$$anonfun$2.apply(ParallelRunner.scala:131)
at sbt.Distributor$Run$Worker$$anonfun$2.apply(ParallelRunner.scala:131)
at sbt.Control$.trapUnit(Control.scala:19)
at sbt.Distributor$Run$Worker.run(ParallelRunner.scala:131)
[0m[[0minfo[0m] [34m== core-kafka / kafka.server.LogRecoveryTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / test-finish ==[0m
[0m[[31merror[0m] [0mFailed: : Total 145, Failed 2, Errors 0, Passed 143, Skipped 0[0m
[0m[[0minfo[0m] [34m== core-kafka / test-finish ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / Test cleanup 1 ==[0m
[0m[[0minfo[0m] [0mDeleting directory /tmp/sbt_3327170[0m
[0m[[0minfo[0m] [34m== core-kafka / Test cleanup 1 ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / test-cleanup ==[0m
[0m[[0minfo[0m] [34m== core-kafka / test-cleanup ==[0m
[0m[[31merror[0m] [0mError running kafka.integration.LazyInitProducerTest: Test FAILED[0m
[0m[[31merror[0m] [0mError running kafka.server.LogRecoveryTest: Test FAILED[0m
[0m[[31merror[0m] [0mError running test: One or more subtasks failed[0m
[0m[[0minfo[0m] [0m[0m
[0m[[0minfo[0m] [0mTotal time: 245 s, completed Jul 26, 2012 8:14:13 PM[0m
[0m[[0minfo[0m] [0m[0m
[0m[[0minfo[0m] [0mTotal session time: 276 s, completed Jul 26, 2012 8:14:13 PM[0m
[0m[[31merror[0m] [0mError during build.[0m
Build step 'Execute shell' marked build as failure