You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@kafka.apache.org by Apache Jenkins Server <je...@builds.apache.org> on 2012/08/06 20:10:09 UTC
Build failed in Jenkins: Kafka-0.8 #17
See <https://builds.apache.org/job/Kafka-0.8/17/changes>
Changes:
[nehanarkhede] KAFKA-380 Enhance single_host_multi_brokers test with failure to trigger leader re-election in replication; patched by John Fung; reviewed by Neha Narkhede, Jun Rao and Joel Koshy
------------------------------------------
[...truncated 1987 lines...]
at junit.framework.TestCase.runBare(TestCase.java:130)
at junit.framework.TestResult$1.protect(TestResult.java:110)
at junit.framework.TestResult.runProtected(TestResult.java:128)
at junit.framework.TestResult.run(TestResult.java:113)
at junit.framework.TestCase.run(TestCase.java:120)
at junit.framework.TestSuite.runTest(TestSuite.java:228)
at junit.framework.TestSuite.run(TestSuite.java:223)
at junit.framework.TestSuite.runTest(TestSuite.java:228)
at junit.framework.TestSuite.run(TestSuite.java:223)
at org.scalatest.junit.JUnit3Suite.run(JUnit3Suite.scala:309)
at org.scalatest.tools.ScalaTestFramework$ScalaTestRunner.run(ScalaTestFramework.scala:40)
at sbt.TestRunner.run(TestFramework.scala:53)
at sbt.TestRunner.runTest$1(TestFramework.scala:67)
at sbt.TestRunner.run(TestFramework.scala:76)
at sbt.TestFramework$$anonfun$10$$anonfun$apply$11.runTest$2(TestFramework.scala:194)
at sbt.TestFramework$$anonfun$10$$anonfun$apply$11$$anonfun$apply$12.apply(TestFramework.scala:205)
at sbt.TestFramework$$anonfun$10$$anonfun$apply$11$$anonfun$apply$12.apply(TestFramework.scala:205)
at sbt.NamedTestTask.run(TestFramework.scala:92)
at sbt.ScalaProject$$anonfun$sbt$ScalaProject$$toTask$1.apply(ScalaProject.scala:193)
at sbt.ScalaProject$$anonfun$sbt$ScalaProject$$toTask$1.apply(ScalaProject.scala:193)
at sbt.TaskManager$Task.invoke(TaskManager.scala:62)
at sbt.impl.RunTask.doRun$1(RunTask.scala:77)
at sbt.impl.RunTask.runTask(RunTask.scala:85)
at sbt.impl.RunTask.sbt$impl$RunTask$$runIfNotRoot(RunTask.scala:60)
at sbt.impl.RunTask$$anonfun$runTasksExceptRoot$2.apply(RunTask.scala:48)
at sbt.impl.RunTask$$anonfun$runTasksExceptRoot$2.apply(RunTask.scala:48)
at sbt.Distributor$Run$Worker$$anonfun$2.apply(ParallelRunner.scala:131)
at sbt.Distributor$Run$Worker$$anonfun$2.apply(ParallelRunner.scala:131)
at sbt.Control$.trapUnit(Control.scala:19)
at sbt.Distributor$Run$Worker.run(ParallelRunner.scala:131)
[2012-08-06 18:09:51,684] WARN EndOfStreamException: Unable to read additional data from client sessionid 0x138fd1f554f0008, likely client has closed socket (org.apache.zookeeper.server.NIOServerCnxn:634)
[0m[[0minfo[0m] [0mTest Passed: testInvalidPartition(kafka.producer.AsyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testNoBroker(kafka.producer.AsyncProducerTest)[0m
[2012-08-06 18:09:51,702] ERROR Failed to collate messages by topic, partition due to (kafka.producer.async.DefaultEventHandler:99)
kafka.common.NoBrokersForPartitionException: Partition = null
at kafka.producer.async.DefaultEventHandler.kafka$producer$async$DefaultEventHandler$$getPartitionListForTopic(DefaultEventHandler.scala:145)
at kafka.producer.async.DefaultEventHandler$$anonfun$partitionAndCollate$1.apply(DefaultEventHandler.scala:102)
at kafka.producer.async.DefaultEventHandler$$anonfun$partitionAndCollate$1.apply(DefaultEventHandler.scala:101)
at scala.collection.LinearSeqOptimized$class.foreach(LinearSeqOptimized.scala:61)
at scala.collection.immutable.List.foreach(List.scala:45)
at scala.collection.generic.TraversableForwarder$class.foreach(TraversableForwarder.scala:44)
at scala.collection.mutable.ListBuffer.foreach(ListBuffer.scala:42)
at kafka.producer.async.DefaultEventHandler.partitionAndCollate(DefaultEventHandler.scala:101)
at kafka.producer.async.DefaultEventHandler.dispatchSerializedData(DefaultEventHandler.scala:66)
at kafka.producer.async.DefaultEventHandler.handle(DefaultEventHandler.scala:49)
at kafka.producer.AsyncProducerTest.testNoBroker(AsyncProducerTest.scala:296)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at junit.framework.TestCase.runTest(TestCase.java:164)
at junit.framework.TestCase.runBare(TestCase.java:130)
at junit.framework.TestResult$1.protect(TestResult.java:110)
at junit.framework.TestResult.runProtected(TestResult.java:128)
at junit.framework.TestResult.run(TestResult.java:113)
at junit.framework.TestCase.run(TestCase.java:120)
at junit.framework.TestSuite.runTest(TestSuite.java:228)
at junit.framework.TestSuite.run(TestSuite.java:223)
at junit.framework.TestSuite.runTest(TestSuite.java:228)
at junit.framework.TestSuite.run(TestSuite.java:223)
at org.scalatest.junit.JUnit3Suite.run(JUnit3Suite.scala:309)
at org.scalatest.tools.ScalaTestFramework$ScalaTestRunner.run(ScalaTestFramework.scala:40)
at sbt.TestRunner.run(TestFramework.scala:53)
at sbt.TestRunner.runTest$1(TestFramework.scala:67)
at sbt.TestRunner.run(TestFramework.scala:76)
at sbt.TestFramework$$anonfun$10$$anonfun$apply$11.runTest$2(TestFramework.scala:194)
at sbt.TestFramework$$anonfun$10$$anonfun$apply$11$$anonfun$apply$12.apply(TestFramework.scala:205)
at sbt.TestFramework$$anonfun$10$$anonfun$apply$11$$anonfun$apply$12.apply(TestFramework.scala:205)
at sbt.NamedTestTask.run(TestFramework.scala:92)
at sbt.ScalaProject$$anonfun$sbt$ScalaProject$$toTask$1.apply(ScalaProject.scala:193)
at sbt.ScalaProject$$anonfun$sbt$ScalaProject$$toTask$1.apply(ScalaProject.scala:193)
at sbt.TaskManager$Task.invoke(TaskManager.scala:62)
at sbt.impl.RunTask.doRun$1(RunTask.scala:77)
at sbt.impl.RunTask.runTask(RunTask.scala:85)
at sbt.impl.RunTask.sbt$impl$RunTask$$runIfNotRoot(RunTask.scala:60)
at sbt.impl.RunTask$$anonfun$runTasksExceptRoot$2.apply(RunTask.scala:48)
at sbt.impl.RunTask$$anonfun$runTasksExceptRoot$2.apply(RunTask.scala:48)
at sbt.Distributor$Run$Worker$$anonfun$2.apply(ParallelRunner.scala:131)
at sbt.Distributor$Run$Worker$$anonfun$2.apply(ParallelRunner.scala:131)
at sbt.Control$.trapUnit(Control.scala:19)
at sbt.Distributor$Run$Worker.run(ParallelRunner.scala:131)
[0m[[0minfo[0m] [0mTest Passed: testNoBroker(kafka.producer.AsyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testIncompatibleEncoder(kafka.producer.AsyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testIncompatibleEncoder(kafka.producer.AsyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testRandomPartitioner(kafka.producer.AsyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testRandomPartitioner(kafka.producer.AsyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testBrokerListAndAsync(kafka.producer.AsyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testBrokerListAndAsync(kafka.producer.AsyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testFailedSendRetryLogic(kafka.producer.AsyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testFailedSendRetryLogic(kafka.producer.AsyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testJavaProducer(kafka.producer.AsyncProducerTest)[0m
[2012-08-06 18:09:51,994] WARN EndOfStreamException: Unable to read additional data from client sessionid 0x138fd2090ce0011, likely client has closed socket (org.apache.zookeeper.server.NIOServerCnxn:634)
[0m[[0minfo[0m] [0mTest Passed: testJavaProducer(kafka.producer.AsyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testInvalidConfiguration(kafka.producer.AsyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testInvalidConfiguration(kafka.producer.AsyncProducerTest)[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.producer.AsyncProducerTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.javaapi.message.ByteBufferMessageSetTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testWrittenEqualsRead[0m
[0m[[0minfo[0m] [0mTest Passed: testWrittenEqualsRead[0m
[0m[[0minfo[0m] [0mTest Starting: testIteratorIsConsistent[0m
[0m[[0minfo[0m] [0mTest Passed: testIteratorIsConsistent[0m
[0m[[0minfo[0m] [0mTest Starting: testSizeInBytes[0m
[0m[[0minfo[0m] [0mTest Passed: testSizeInBytes[0m
[0m[[0minfo[0m] [0mTest Starting: testValidBytes[0m
[0m[[0minfo[0m] [0mTest Passed: testValidBytes[0m
[0m[[0minfo[0m] [0mTest Starting: testEquals[0m
[0m[[0minfo[0m] [0mTest Passed: testEquals[0m
[0m[[0minfo[0m] [0mTest Starting: testIteratorIsConsistentWithCompression[0m
[0m[[0minfo[0m] [0mTest Passed: testIteratorIsConsistentWithCompression[0m
[0m[[0minfo[0m] [0mTest Starting: testSizeInBytesWithCompression[0m
[0m[[0minfo[0m] [0mTest Passed: testSizeInBytesWithCompression[0m
[0m[[0minfo[0m] [0mTest Starting: testValidBytesWithCompression[0m
[0m[[0minfo[0m] [0mTest Passed: testValidBytesWithCompression[0m
[0m[[0minfo[0m] [0mTest Starting: testEqualsWithCompression[0m
[0m[[0minfo[0m] [0mTest Passed: testEqualsWithCompression[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.javaapi.message.ByteBufferMessageSetTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.log4j.KafkaLog4jAppenderTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testKafkaLog4jConfigs(kafka.log4j.KafkaLog4jAppenderTest)[0m
log4j:WARN No appenders could be found for logger (org.I0Itec.zkclient.ZkEventThread).
log4j:WARN Please initialize the log4j system properly.
[0m[[0minfo[0m] [0mTest Passed: testKafkaLog4jConfigs(kafka.log4j.KafkaLog4jAppenderTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testZkConnectLog4jAppends(kafka.log4j.KafkaLog4jAppenderTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testZkConnectLog4jAppends(kafka.log4j.KafkaLog4jAppenderTest)[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.log4j.KafkaLog4jAppenderTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.producer.ProducerTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testZKSendToNewTopic(kafka.producer.ProducerTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testZKSendToNewTopic(kafka.producer.ProducerTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testZKSendWithDeadBroker(kafka.producer.ProducerTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testZKSendWithDeadBroker(kafka.producer.ProducerTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testAsyncSendCanCorrectlyFailWithTimeout(kafka.producer.ProducerTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testAsyncSendCanCorrectlyFailWithTimeout(kafka.producer.ProducerTest)[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.producer.ProducerTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.admin.AdminTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testReplicaAssignment(kafka.admin.AdminTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testReplicaAssignment(kafka.admin.AdminTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testManualReplicaAssignment(kafka.admin.AdminTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testManualReplicaAssignment(kafka.admin.AdminTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testTopicCreationInZK(kafka.admin.AdminTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testTopicCreationInZK(kafka.admin.AdminTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testGetTopicMetadata(kafka.admin.AdminTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testGetTopicMetadata(kafka.admin.AdminTest)[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.admin.AdminTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.network.SocketServerTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: simpleRequest[0m
[0m[[0minfo[0m] [0mTest Passed: simpleRequest[0m
[0m[[0minfo[0m] [0mTest Starting: tooBigRequestIsRejected[0m
[0m[[0minfo[0m] [0mTest Passed: tooBigRequestIsRejected[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.network.SocketServerTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.message.MessageTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testFieldValues[0m
[0m[[0minfo[0m] [0mTest Passed: testFieldValues[0m
[0m[[0minfo[0m] [0mTest Starting: testChecksum[0m
[0m[[0minfo[0m] [0mTest Passed: testChecksum[0m
[0m[[0minfo[0m] [0mTest Starting: testEquality[0m
[0m[[0minfo[0m] [0mTest Passed: testEquality[0m
[0m[[0minfo[0m] [0mTest Starting: testIsHashable[0m
[0m[[0minfo[0m] [0mTest Passed: testIsHashable[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.message.MessageTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.producer.SyncProducerTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testReachableServer(kafka.producer.SyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testReachableServer(kafka.producer.SyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testEmptyProduceRequest(kafka.producer.SyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testEmptyProduceRequest(kafka.producer.SyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testSingleMessageSizeTooLarge(kafka.producer.SyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testSingleMessageSizeTooLarge(kafka.producer.SyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testCompressedMessageSizeTooLarge(kafka.producer.SyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testCompressedMessageSizeTooLarge(kafka.producer.SyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testProduceCorrectlyReceivesResponse(kafka.producer.SyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testProduceCorrectlyReceivesResponse(kafka.producer.SyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testProducerCanTimeout(kafka.producer.SyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testProducerCanTimeout(kafka.producer.SyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testProduceRequestForUnknownTopic(kafka.producer.SyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testProduceRequestForUnknownTopic(kafka.producer.SyncProducerTest)[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.producer.SyncProducerTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.controller.ControllerBasicTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testControllerFailOver(kafka.controller.ControllerBasicTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testControllerFailOver(kafka.controller.ControllerBasicTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testControllerCommandSend(kafka.controller.ControllerBasicTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testControllerCommandSend(kafka.controller.ControllerBasicTest)[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.controller.ControllerBasicTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / test-finish ==[0m
[0m[[31merror[0m] [0mFailed: : Total 136, Failed 1, Errors 0, Passed 135, Skipped 0[0m
[0m[[0minfo[0m] [34m== core-kafka / test-finish ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / Test cleanup 1 ==[0m
[0m[[0minfo[0m] [0mDeleting directory /var/tmp/sbt_efc390bd[0m
[0m[[0minfo[0m] [34m== core-kafka / Test cleanup 1 ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / test-cleanup ==[0m
[0m[[0minfo[0m] [34m== core-kafka / test-cleanup ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== hadoop producer / test-compile ==[0m
[0m[[0minfo[0m] [0m Source analysis: 0 new/modified, 0 indirectly invalidated, 0 removed.[0m
[0m[[0minfo[0m] [0mCompiling test sources...[0m
[0m[[0minfo[0m] [0mNothing to compile.[0m
[0m[[0minfo[0m] [0m Post-analysis: 0 classes.[0m
[0m[[0minfo[0m] [34m== hadoop producer / test-compile ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== java-examples / test-compile ==[0m
[0m[[0minfo[0m] [0m Source analysis: 0 new/modified, 0 indirectly invalidated, 0 removed.[0m
[0m[[0minfo[0m] [0mCompiling test sources...[0m
[0m[[0minfo[0m] [0mNothing to compile.[0m
[0m[[0minfo[0m] [0m Post-analysis: 0 classes.[0m
[0m[[0minfo[0m] [34m== java-examples / test-compile ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== hadoop producer / copy-resources ==[0m
[0m[[0minfo[0m] [34m== hadoop producer / copy-resources ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== hadoop producer / copy-test-resources ==[0m
[0m[[0minfo[0m] [34m== hadoop producer / copy-test-resources ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== hadoop consumer / copy-test-resources ==[0m
[0m[[0minfo[0m] [34m== hadoop consumer / copy-test-resources ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== perf / copy-resources ==[0m
[0m[[0minfo[0m] [34m== perf / copy-resources ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== java-examples / copy-test-resources ==[0m
[0m[[0minfo[0m] [34m== java-examples / copy-test-resources ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== java-examples / copy-resources ==[0m
[0m[[0minfo[0m] [34m== java-examples / copy-resources ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== hadoop consumer / test-compile ==[0m
[0m[[0minfo[0m] [0m Source analysis: 0 new/modified, 0 indirectly invalidated, 0 removed.[0m
[0m[[0minfo[0m] [0mCompiling test sources...[0m
[0m[[0minfo[0m] [0mNothing to compile.[0m
[0m[[0minfo[0m] [0m Post-analysis: 0 classes.[0m
[0m[[0minfo[0m] [34m== hadoop consumer / test-compile ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== hadoop consumer / copy-resources ==[0m
[0m[[0minfo[0m] [34m== hadoop consumer / copy-resources ==[0m
[0m[[31merror[0m] [0mError running kafka.message.CompressionUtilTest: Test FAILED[0m
[0m[[31merror[0m] [0mError running test: One or more subtasks failed[0m
[0m[[0minfo[0m] [0m[0m
[0m[[0minfo[0m] [0mTotal time: 495 s, completed Aug 6, 2012 6:10:16 PM[0m
[0m[[0minfo[0m] [0m[0m
[0m[[0minfo[0m] [0mTotal session time: 496 s, completed Aug 6, 2012 6:10:16 PM[0m
[0m[[31merror[0m] [0mError during build.[0m
Build step 'Execute shell' marked build as failure
Jenkins build is back to normal : Kafka-0.8 #18
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/Kafka-0.8/18/changes>