You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@kafka.apache.org by Apache Jenkins Server <je...@builds.apache.org> on 2012/11/19 07:09:42 UTC
Build failed in Jenkins: Kafka-0.8 #113
See <https://builds.apache.org/job/Kafka-0.8/113/changes>
Changes:
[junrao] move shutting down of fetcher thread out of critical path; patched by Jun Rao; reviewed by Neha Narkhede; KAFKA-612
------------------------------------------
[...truncated 2862 lines...]
at kafka.server.KafkaApis.kafka$server$KafkaApis$$readMessageSet(KafkaApis.scala:359)
at kafka.server.KafkaApis$$anonfun$kafka$server$KafkaApis$$readMessageSets$1.apply(KafkaApis.scala:325)
at kafka.server.KafkaApis$$anonfun$kafka$server$KafkaApis$$readMessageSets$1.apply(KafkaApis.scala:321)
at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
at scala.collection.immutable.Map$Map4.foreach(Map.scala:180)
at scala.collection.TraversableLike$class.map(TraversableLike.scala:206)
at scala.collection.immutable.Map$Map4.map(Map.scala:157)
at kafka.server.KafkaApis.kafka$server$KafkaApis$$readMessageSets(KafkaApis.scala:321)
at kafka.server.KafkaApis.handleFetchRequest(KafkaApis.scala:289)
at kafka.server.KafkaApis.handle(KafkaApis.scala:57)
at kafka.server.KafkaRequestHandler.run(KafkaRequestHandler.scala:41)
at java.lang.Thread.run(Thread.java:662)
[2012-11-19 06:09:14,085] ERROR [KafkaApi-0] error when processing request (test1,-1,0,10000) (kafka.server.KafkaApis:102)
kafka.common.UnknownTopicOrPartitionException: Topic test1 partition -1 doesn't exist on 0
at kafka.server.ReplicaManager.getLeaderReplicaIfLocal(ReplicaManager.scala:163)
at kafka.server.KafkaApis.kafka$server$KafkaApis$$readMessageSet(KafkaApis.scala:359)
at kafka.server.KafkaApis$$anonfun$kafka$server$KafkaApis$$readMessageSets$1.apply(KafkaApis.scala:325)
at kafka.server.KafkaApis$$anonfun$kafka$server$KafkaApis$$readMessageSets$1.apply(KafkaApis.scala:321)
at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
at scala.collection.immutable.Map$Map4.foreach(Map.scala:180)
at scala.collection.TraversableLike$class.map(TraversableLike.scala:206)
at scala.collection.immutable.Map$Map4.map(Map.scala:157)
at kafka.server.KafkaApis.kafka$server$KafkaApis$$readMessageSets(KafkaApis.scala:321)
at kafka.server.KafkaApis.handleFetchRequest(KafkaApis.scala:289)
at kafka.server.KafkaApis.handle(KafkaApis.scala:57)
at kafka.server.KafkaRequestHandler.run(KafkaRequestHandler.scala:41)
at java.lang.Thread.run(Thread.java:662)
[0m[[0minfo[0m] [0mTest Passed: testProduceAndMultiFetch(kafka.integration.PrimitiveApiTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testMultiProduce(kafka.integration.PrimitiveApiTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testMultiProduce(kafka.integration.PrimitiveApiTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testFetchRequestCanProperlySerialize(kafka.integration.PrimitiveApiTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testFetchRequestCanProperlySerialize(kafka.integration.PrimitiveApiTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testEmptyFetchRequest(kafka.integration.PrimitiveApiTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testEmptyFetchRequest(kafka.integration.PrimitiveApiTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testDefaultEncoderProducerAndFetch(kafka.integration.PrimitiveApiTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testDefaultEncoderProducerAndFetch(kafka.integration.PrimitiveApiTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testDefaultEncoderProducerAndFetchWithCompression(kafka.integration.PrimitiveApiTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testDefaultEncoderProducerAndFetchWithCompression(kafka.integration.PrimitiveApiTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testProduceAndMultiFetchWithCompression(kafka.integration.PrimitiveApiTest)[0m
[2012-11-19 06:09:18,635] ERROR [KafkaApi-0] error when processing request (test2,0,-1,10000) (kafka.server.KafkaApis:102)
kafka.common.OffsetOutOfRangeException: Request for offset -1 but we only have log segments in the range 0 to 2.
at kafka.log.Log.read(Log.scala:371)
at kafka.server.KafkaApis.kafka$server$KafkaApis$$readMessageSet(KafkaApis.scala:368)
at kafka.server.KafkaApis$$anonfun$kafka$server$KafkaApis$$readMessageSets$1.apply(KafkaApis.scala:325)
at kafka.server.KafkaApis$$anonfun$kafka$server$KafkaApis$$readMessageSets$1.apply(KafkaApis.scala:321)
at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
at scala.collection.immutable.Map$Map4.foreach(Map.scala:180)
at scala.collection.TraversableLike$class.map(TraversableLike.scala:206)
at scala.collection.immutable.Map$Map4.map(Map.scala:157)
at kafka.server.KafkaApis.kafka$server$KafkaApis$$readMessageSets(KafkaApis.scala:321)
at kafka.server.KafkaApis.handleFetchRequest(KafkaApis.scala:289)
at kafka.server.KafkaApis.handle(KafkaApis.scala:57)
at kafka.server.KafkaRequestHandler.run(KafkaRequestHandler.scala:41)
at java.lang.Thread.run(Thread.java:662)
[2012-11-19 06:09:18,636] ERROR [KafkaApi-0] error when processing request (test3,0,-1,10000) (kafka.server.KafkaApis:102)
kafka.common.OffsetOutOfRangeException: Request for offset -1 but we only have log segments in the range 0 to 2.
at kafka.log.Log.read(Log.scala:371)
at kafka.server.KafkaApis.kafka$server$KafkaApis$$readMessageSet(KafkaApis.scala:368)
at kafka.server.KafkaApis$$anonfun$kafka$server$KafkaApis$$readMessageSets$1.apply(KafkaApis.scala:325)
at kafka.server.KafkaApis$$anonfun$kafka$server$KafkaApis$$readMessageSets$1.apply(KafkaApis.scala:321)
at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
at scala.collection.immutable.Map$Map4.foreach(Map.scala:180)
at scala.collection.TraversableLike$class.map(TraversableLike.scala:206)
at scala.collection.immutable.Map$Map4.map(Map.scala:157)
at kafka.server.KafkaApis.kafka$server$KafkaApis$$readMessageSets(KafkaApis.scala:321)
at kafka.server.KafkaApis.handleFetchRequest(KafkaApis.scala:289)
at kafka.server.KafkaApis.handle(KafkaApis.scala:57)
at kafka.server.KafkaRequestHandler.run(KafkaRequestHandler.scala:41)
at java.lang.Thread.run(Thread.java:662)
[2012-11-19 06:09:18,636] ERROR [KafkaApi-0] error when processing request (test4,0,-1,10000) (kafka.server.KafkaApis:102)
kafka.common.OffsetOutOfRangeException: Request for offset -1 but we only have log segments in the range 0 to 2.
at kafka.log.Log.read(Log.scala:371)
at kafka.server.KafkaApis.kafka$server$KafkaApis$$readMessageSet(KafkaApis.scala:368)
at kafka.server.KafkaApis$$anonfun$kafka$server$KafkaApis$$readMessageSets$1.apply(KafkaApis.scala:325)
at kafka.server.KafkaApis$$anonfun$kafka$server$KafkaApis$$readMessageSets$1.apply(KafkaApis.scala:321)
at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
at scala.collection.immutable.Map$Map4.foreach(Map.scala:180)
at scala.collection.TraversableLike$class.map(TraversableLike.scala:206)
at scala.collection.immutable.Map$Map4.map(Map.scala:157)
at kafka.server.KafkaApis.kafka$server$KafkaApis$$readMessageSets(KafkaApis.scala:321)
at kafka.server.KafkaApis.handleFetchRequest(KafkaApis.scala:289)
at kafka.server.KafkaApis.handle(KafkaApis.scala:57)
at kafka.server.KafkaRequestHandler.run(KafkaRequestHandler.scala:41)
at java.lang.Thread.run(Thread.java:662)
[2012-11-19 06:09:18,637] ERROR [KafkaApi-0] error when processing request (test1,0,-1,10000) (kafka.server.KafkaApis:102)
kafka.common.OffsetOutOfRangeException: Request for offset -1 but we only have log segments in the range 0 to 2.
at kafka.log.Log.read(Log.scala:371)
at kafka.server.KafkaApis.kafka$server$KafkaApis$$readMessageSet(KafkaApis.scala:368)
at kafka.server.KafkaApis$$anonfun$kafka$server$KafkaApis$$readMessageSets$1.apply(KafkaApis.scala:325)
at kafka.server.KafkaApis$$anonfun$kafka$server$KafkaApis$$readMessageSets$1.apply(KafkaApis.scala:321)
at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
at scala.collection.immutable.Map$Map4.foreach(Map.scala:180)
at scala.collection.TraversableLike$class.map(TraversableLike.scala:206)
at scala.collection.immutable.Map$Map4.map(Map.scala:157)
at kafka.server.KafkaApis.kafka$server$KafkaApis$$readMessageSets(KafkaApis.scala:321)
at kafka.server.KafkaApis.handleFetchRequest(KafkaApis.scala:289)
at kafka.server.KafkaApis.handle(KafkaApis.scala:57)
at kafka.server.KafkaRequestHandler.run(KafkaRequestHandler.scala:41)
at java.lang.Thread.run(Thread.java:662)
[2012-11-19 06:09:18,639] ERROR [KafkaApi-0] error when processing request (test2,-1,0,10000) (kafka.server.KafkaApis:102)
kafka.common.UnknownTopicOrPartitionException: Topic test2 partition -1 doesn't exist on 0
at kafka.server.ReplicaManager.getLeaderReplicaIfLocal(ReplicaManager.scala:163)
at kafka.server.KafkaApis.kafka$server$KafkaApis$$readMessageSet(KafkaApis.scala:359)
at kafka.server.KafkaApis$$anonfun$kafka$server$KafkaApis$$readMessageSets$1.apply(KafkaApis.scala:325)
at kafka.server.KafkaApis$$anonfun$kafka$server$KafkaApis$$readMessageSets$1.apply(KafkaApis.scala:321)
at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
at scala.collection.immutable.Map$Map4.foreach(Map.scala:180)
at scala.collection.TraversableLike$class.map(TraversableLike.scala:206)
at scala.collection.immutable.Map$Map4.map(Map.scala:157)
at kafka.server.KafkaApis.kafka$server$KafkaApis$$readMessageSets(KafkaApis.scala:321)
at kafka.server.KafkaApis.handleFetchRequest(KafkaApis.scala:289)
at kafka.server.KafkaApis.handle(KafkaApis.scala:57)
at kafka.server.KafkaRequestHandler.run(KafkaRequestHandler.scala:41)
at java.lang.Thread.run(Thread.java:662)
[2012-11-19 06:09:18,640] ERROR [KafkaApi-0] error when processing request (test3,-1,0,10000) (kafka.server.KafkaApis:102)
kafka.common.UnknownTopicOrPartitionException: Topic test3 partition -1 doesn't exist on 0
at kafka.server.ReplicaManager.getLeaderReplicaIfLocal(ReplicaManager.scala:163)
at kafka.server.KafkaApis.kafka$server$KafkaApis$$readMessageSet(KafkaApis.scala:359)
at kafka.server.KafkaApis$$anonfun$kafka$server$KafkaApis$$readMessageSets$1.apply(KafkaApis.scala:325)
at kafka.server.KafkaApis$$anonfun$kafka$server$KafkaApis$$readMessageSets$1.apply(KafkaApis.scala:321)
at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
at scala.collection.immutable.Map$Map4.foreach(Map.scala:180)
at scala.collection.TraversableLike$class.map(TraversableLike.scala:206)
at scala.collection.immutable.Map$Map4.map(Map.scala:157)
at kafka.server.KafkaApis.kafka$server$KafkaApis$$readMessageSets(KafkaApis.scala:321)
at kafka.server.KafkaApis.handleFetchRequest(KafkaApis.scala:289)
at kafka.server.KafkaApis.handle(KafkaApis.scala:57)
at kafka.server.KafkaRequestHandler.run(KafkaRequestHandler.scala:41)
at java.lang.Thread.run(Thread.java:662)
[2012-11-19 06:09:18,640] ERROR [KafkaApi-0] error when processing request (test4,-1,0,10000) (kafka.server.KafkaApis:102)
kafka.common.UnknownTopicOrPartitionException: Topic test4 partition -1 doesn't exist on 0
at kafka.server.ReplicaManager.getLeaderReplicaIfLocal(ReplicaManager.scala:163)
at kafka.server.KafkaApis.kafka$server$KafkaApis$$readMessageSet(KafkaApis.scala:359)
at kafka.server.KafkaApis$$anonfun$kafka$server$KafkaApis$$readMessageSets$1.apply(KafkaApis.scala:325)
at kafka.server.KafkaApis$$anonfun$kafka$server$KafkaApis$$readMessageSets$1.apply(KafkaApis.scala:321)
at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
at scala.collection.immutable.Map$Map4.foreach(Map.scala:180)
at scala.collection.TraversableLike$class.map(TraversableLike.scala:206)
at scala.collection.immutable.Map$Map4.map(Map.scala:157)
at kafka.server.KafkaApis.kafka$server$KafkaApis$$readMessageSets(KafkaApis.scala:321)
at kafka.server.KafkaApis.handleFetchRequest(KafkaApis.scala:289)
at kafka.server.KafkaApis.handle(KafkaApis.scala:57)
at kafka.server.KafkaRequestHandler.run(KafkaRequestHandler.scala:41)
at java.lang.Thread.run(Thread.java:662)
[2012-11-19 06:09:18,641] ERROR [KafkaApi-0] error when processing request (test1,-1,0,10000) (kafka.server.KafkaApis:102)
kafka.common.UnknownTopicOrPartitionException: Topic test1 partition -1 doesn't exist on 0
at kafka.server.ReplicaManager.getLeaderReplicaIfLocal(ReplicaManager.scala:163)
at kafka.server.KafkaApis.kafka$server$KafkaApis$$readMessageSet(KafkaApis.scala:359)
at kafka.server.KafkaApis$$anonfun$kafka$server$KafkaApis$$readMessageSets$1.apply(KafkaApis.scala:325)
at kafka.server.KafkaApis$$anonfun$kafka$server$KafkaApis$$readMessageSets$1.apply(KafkaApis.scala:321)
at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
at scala.collection.immutable.Map$Map4.foreach(Map.scala:180)
at scala.collection.TraversableLike$class.map(TraversableLike.scala:206)
at scala.collection.immutable.Map$Map4.map(Map.scala:157)
at kafka.server.KafkaApis.kafka$server$KafkaApis$$readMessageSets(KafkaApis.scala:321)
at kafka.server.KafkaApis.handleFetchRequest(KafkaApis.scala:289)
at kafka.server.KafkaApis.handle(KafkaApis.scala:57)
at kafka.server.KafkaRequestHandler.run(KafkaRequestHandler.scala:41)
at java.lang.Thread.run(Thread.java:662)
[0m[[0minfo[0m] [0mTest Passed: testProduceAndMultiFetchWithCompression(kafka.integration.PrimitiveApiTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testMultiProduceWithCompression(kafka.integration.PrimitiveApiTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testMultiProduceWithCompression(kafka.integration.PrimitiveApiTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testConsumerEmptyTopic(kafka.integration.PrimitiveApiTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testConsumerEmptyTopic(kafka.integration.PrimitiveApiTest)[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.integration.PrimitiveApiTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.log4j.KafkaLog4jAppenderTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testKafkaLog4jConfigs(kafka.log4j.KafkaLog4jAppenderTest)[0m
log4j:WARN No appenders could be found for logger (org.I0Itec.zkclient.ZkEventThread).
log4j:WARN Please initialize the log4j system properly.
[0m[[0minfo[0m] [0mTest Passed: testKafkaLog4jConfigs(kafka.log4j.KafkaLog4jAppenderTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testLog4jAppends(kafka.log4j.KafkaLog4jAppenderTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testLog4jAppends(kafka.log4j.KafkaLog4jAppenderTest)[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.log4j.KafkaLog4jAppenderTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.metrics.KafkaTimerTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testKafkaTimer(kafka.metrics.KafkaTimerTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testKafkaTimer(kafka.metrics.KafkaTimerTest)[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.metrics.KafkaTimerTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.message.ByteBufferMessageSetTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testWrittenEqualsRead[0m
[0m[[0minfo[0m] [0mTest Passed: testWrittenEqualsRead[0m
[0m[[0minfo[0m] [0mTest Starting: testIteratorIsConsistent[0m
[0m[[0minfo[0m] [0mTest Passed: testIteratorIsConsistent[0m
[0m[[0minfo[0m] [0mTest Starting: testSizeInBytes[0m
[0m[[0minfo[0m] [0mTest Passed: testSizeInBytes[0m
[0m[[0minfo[0m] [0mTest Starting: testEquals[0m
[0m[[0minfo[0m] [0mTest Passed: testEquals[0m
[0m[[0minfo[0m] [0mTest Starting: testWriteTo[0m
[0m[[0minfo[0m] [0mTest Passed: testWriteTo[0m
[0m[[0minfo[0m] [0mTest Starting: testValidBytes[0m
[0m[[0minfo[0m] [0mTest Passed: testValidBytes[0m
[0m[[0minfo[0m] [0mTest Starting: testValidBytesWithCompression[0m
[0m[[0minfo[0m] [0mTest Passed: testValidBytesWithCompression[0m
[0m[[0minfo[0m] [0mTest Starting: testIterator[0m
[0m[[0minfo[0m] [0mTest Passed: testIterator[0m
[0m[[0minfo[0m] [0mTest Starting: testOffsetAssignment[0m
[0m[[0minfo[0m] [0mTest Passed: testOffsetAssignment[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.message.ByteBufferMessageSetTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.producer.ProducerTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testUpdateBrokerPartitionInfo(kafka.producer.ProducerTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testUpdateBrokerPartitionInfo(kafka.producer.ProducerTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testSendToNewTopic(kafka.producer.ProducerTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testSendToNewTopic(kafka.producer.ProducerTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testSendWithDeadBroker(kafka.producer.ProducerTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testSendWithDeadBroker(kafka.producer.ProducerTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testAsyncSendCanCorrectlyFailWithTimeout(kafka.producer.ProducerTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testAsyncSendCanCorrectlyFailWithTimeout(kafka.producer.ProducerTest)[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.producer.ProducerTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.integration.FetcherTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testFetcher(kafka.integration.FetcherTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testFetcher(kafka.integration.FetcherTest)[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.integration.FetcherTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / Test cleanup 1 ==[0m
[0m[[0minfo[0m] [0mDeleting directory /tmp/sbt_b5aaac46[0m
[0m[[0minfo[0m] [34m== core-kafka / Test cleanup 1 ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / test-finish ==[0m
[0m[[31merror[0m] [0mFailed: : Total 167, Failed 2, Errors 0, Passed 165, Skipped 0[0m
[0m[[0minfo[0m] [34m== core-kafka / test-finish ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / test-cleanup ==[0m
[0m[[0minfo[0m] [34m== core-kafka / test-cleanup ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== hadoop consumer / copy-test-resources ==[0m
[0m[[0minfo[0m] [34m== hadoop consumer / copy-test-resources ==[0m
[0m[[31merror[0m] [0mError running kafka.server.LogRecoveryTest: Test FAILED[0m
[0m[[31merror[0m] [0mError running kafka.admin.AdminTest: Test FAILED[0m
[0m[[31merror[0m] [0mError running test: One or more subtasks failed[0m
[0m[[0minfo[0m] [0m[0m
[0m[[0minfo[0m] [0mTotal time: 233 s, completed Nov 19, 2012 6:09:42 AM[0m
[0m[[0minfo[0m] [0m[0m
[0m[[0minfo[0m] [0mTotal session time: 233 s, completed Nov 19, 2012 6:09:42 AM[0m
[0m[[31merror[0m] [0mError during build.[0m
Build step 'Execute shell' marked build as failure
Build failed in Jenkins: Kafka-0.8 #114
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/Kafka-0.8/114/changes>
Changes:
[jkreps] KAFKA-544. Follow-up items on key-retention. Addresses misc. comments from Joel, see ticket for details.
------------------------------------------
[...truncated 2024 lines...]
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at junit.framework.TestCase.runTest(TestCase.java:164)
at junit.framework.TestCase.runBare(TestCase.java:130)
at junit.framework.TestResult$1.protect(TestResult.java:110)
at junit.framework.TestResult.runProtected(TestResult.java:128)
at junit.framework.TestResult.run(TestResult.java:113)
at junit.framework.TestCase.run(TestCase.java:120)
at junit.framework.TestSuite.runTest(TestSuite.java:228)
at junit.framework.TestSuite.run(TestSuite.java:223)
at junit.framework.TestSuite.runTest(TestSuite.java:228)
at junit.framework.TestSuite.run(TestSuite.java:223)
at org.scalatest.junit.JUnit3Suite.run(JUnit3Suite.scala:309)
at org.scalatest.tools.ScalaTestFramework$ScalaTestRunner.run(ScalaTestFramework.scala:40)
at sbt.TestRunner.run(TestFramework.scala:53)
at sbt.TestRunner.runTest$1(TestFramework.scala:67)
at sbt.TestRunner.run(TestFramework.scala:76)
at sbt.TestFramework$$anonfun$10$$anonfun$apply$11.runTest$2(TestFramework.scala:194)
at sbt.TestFramework$$anonfun$10$$anonfun$apply$11$$anonfun$apply$12.apply(TestFramework.scala:205)
at sbt.TestFramework$$anonfun$10$$anonfun$apply$11$$anonfun$apply$12.apply(TestFramework.scala:205)
at sbt.NamedTestTask.run(TestFramework.scala:92)
at sbt.ScalaProject$$anonfun$sbt$ScalaProject$$toTask$1.apply(ScalaProject.scala:193)
at sbt.ScalaProject$$anonfun$sbt$ScalaProject$$toTask$1.apply(ScalaProject.scala:193)
at sbt.TaskManager$Task.invoke(TaskManager.scala:62)
at sbt.impl.RunTask.doRun$1(RunTask.scala:77)
at sbt.impl.RunTask.runTask(RunTask.scala:85)
at sbt.impl.RunTask.sbt$impl$RunTask$$runIfNotRoot(RunTask.scala:60)
at sbt.impl.RunTask$$anonfun$runTasksExceptRoot$2.apply(RunTask.scala:48)
at sbt.impl.RunTask$$anonfun$runTasksExceptRoot$2.apply(RunTask.scala:48)
at sbt.Distributor$Run$Worker$$anonfun$2.apply(ParallelRunner.scala:131)
at sbt.Distributor$Run$Worker$$anonfun$2.apply(ParallelRunner.scala:131)
at sbt.Control$.trapUnit(Control.scala:19)
at sbt.Distributor$Run$Worker.run(ParallelRunner.scala:131)
[0m[[0minfo[0m] [34m== core-kafka / kafka.admin.AdminTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.integration.AutoOffsetResetTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testResetToEarliestWhenOffsetTooHigh(kafka.integration.AutoOffsetResetTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testResetToEarliestWhenOffsetTooHigh(kafka.integration.AutoOffsetResetTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testResetToEarliestWhenOffsetTooLow(kafka.integration.AutoOffsetResetTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testResetToEarliestWhenOffsetTooLow(kafka.integration.AutoOffsetResetTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testResetToLatestWhenOffsetTooHigh(kafka.integration.AutoOffsetResetTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testResetToLatestWhenOffsetTooHigh(kafka.integration.AutoOffsetResetTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testResetToLatestWhenOffsetTooLow(kafka.integration.AutoOffsetResetTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testResetToLatestWhenOffsetTooLow(kafka.integration.AutoOffsetResetTest)[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.integration.AutoOffsetResetTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.server.ServerShutdownTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testCleanShutdown(kafka.server.ServerShutdownTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testCleanShutdown(kafka.server.ServerShutdownTest)[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.server.ServerShutdownTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.producer.AsyncProducerTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testProducerQueueSize(kafka.producer.AsyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testProducerQueueSize(kafka.producer.AsyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testProduceAfterClosed(kafka.producer.AsyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testProduceAfterClosed(kafka.producer.AsyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testBatchSize(kafka.producer.AsyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testBatchSize(kafka.producer.AsyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testQueueTimeExpired(kafka.producer.AsyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testQueueTimeExpired(kafka.producer.AsyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testPartitionAndCollateEvents(kafka.producer.AsyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testPartitionAndCollateEvents(kafka.producer.AsyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testSerializeEvents(kafka.producer.AsyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testSerializeEvents(kafka.producer.AsyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testInvalidPartition(kafka.producer.AsyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testInvalidPartition(kafka.producer.AsyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testNoBroker(kafka.producer.AsyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testNoBroker(kafka.producer.AsyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testIncompatibleEncoder(kafka.producer.AsyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testIncompatibleEncoder(kafka.producer.AsyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testRandomPartitioner(kafka.producer.AsyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testRandomPartitioner(kafka.producer.AsyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testBrokerListAndAsync(kafka.producer.AsyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testBrokerListAndAsync(kafka.producer.AsyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testFailedSendRetryLogic(kafka.producer.AsyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testFailedSendRetryLogic(kafka.producer.AsyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testJavaProducer(kafka.producer.AsyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testJavaProducer(kafka.producer.AsyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Starting: testInvalidConfiguration(kafka.producer.AsyncProducerTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testInvalidConfiguration(kafka.producer.AsyncProducerTest)[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.producer.AsyncProducerTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.javaapi.consumer.ZookeeperConsumerConnectorTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testBasic(kafka.javaapi.consumer.ZookeeperConsumerConnectorTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testBasic(kafka.javaapi.consumer.ZookeeperConsumerConnectorTest)[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.javaapi.consumer.ZookeeperConsumerConnectorTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.message.MessageCompressionTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testSimpleCompressDecompress[0m
java.lang.reflect.InvocationTargetException
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at org.xerial.snappy.SnappyLoader.loadNativeLibrary(SnappyLoader.java:317)
at org.xerial.snappy.SnappyLoader.load(SnappyLoader.java:219)
at org.xerial.snappy.Snappy.<clinit>(Snappy.java:44)
at org.xerial.snappy.SnappyOutputStream.<init>(SnappyOutputStream.java:79)
at org.xerial.snappy.SnappyOutputStream.<init>(SnappyOutputStream.java:66)
at kafka.message.MessageCompressionTest.isSnappyAvailable(MessageCompressionTest.scala:57)
at kafka.message.MessageCompressionTest.testSimpleCompressDecompress(MessageCompressionTest.scala:31)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at org.junit.internal.runners.TestMethodRunner.executeMethodBody(TestMethodRunner.java:99)
at org.junit.internal.runners.TestMethodRunner.runUnprotected(TestMethodRunner.java:81)
at org.junit.internal.runners.BeforeAndAfterRunner.runProtected(BeforeAndAfterRunner.java:34)
at org.junit.internal.runners.TestMethodRunner.runMethod(TestMethodRunner.java:75)
at org.junit.internal.runners.TestMethodRunner.run(TestMethodRunner.java:45)
at org.junit.internal.runners.TestClassMethodsRunner.invokeTestMethod(TestClassMethodsRunner.java:71)
at org.junit.internal.runners.TestClassMethodsRunner.run(TestClassMethodsRunner.java:35)
at org.junit.internal.runners.TestClassRunner$1.runUnprotected(TestClassRunner.java:42)
at org.junit.internal.runners.BeforeAndAfterRunner.runProtected(BeforeAndAfterRunner.java:34)
at org.junit.internal.runners.TestClassRunner.run(TestClassRunner.java:52)
at org.junit.internal.runners.CompositeRunner.run(CompositeRunner.java:29)
at org.junit.runner.JUnitCore.run(JUnitCore.java:121)
at org.junit.runner.JUnitCore.run(JUnitCore.java:100)
at org.junit.runner.JUnitCore.run(JUnitCore.java:91)
at org.scalatest.junit.JUnitSuite$class.run(JUnitSuite.scala:261)
at kafka.message.MessageCompressionTest.run(MessageCompressionTest.scala:26)
at org.scalatest.tools.ScalaTestFramework$ScalaTestRunner.run(ScalaTestFramework.scala:40)
at sbt.TestRunner.run(TestFramework.scala:53)
at sbt.TestRunner.runTest$1(TestFramework.scala:67)
at sbt.TestRunner.run(TestFramework.scala:76)
at sbt.TestFramework$$anonfun$10$$anonfun$apply$11.runTest$2(TestFramework.scala:194)
at sbt.TestFramework$$anonfun$10$$anonfun$apply$11$$anonfun$apply$12.apply(TestFramework.scala:205)
at sbt.TestFramework$$anonfun$10$$anonfun$apply$11$$anonfun$apply$12.apply(TestFramework.scala:205)
at sbt.NamedTestTask.run(TestFramework.scala:92)
at sbt.ScalaProject$$anonfun$sbt$ScalaProject$$toTask$1.apply(ScalaProject.scala:193)
at sbt.ScalaProject$$anonfun$sbt$ScalaProject$$toTask$1.apply(ScalaProject.scala:193)
at sbt.TaskManager$Task.invoke(TaskManager.scala:62)
at sbt.impl.RunTask.doRun$1(RunTask.scala:77)
at sbt.impl.RunTask.runTask(RunTask.scala:85)
at sbt.impl.RunTask.sbt$impl$RunTask$$runIfNotRoot(RunTask.scala:60)
at sbt.impl.RunTask$$anonfun$runTasksExceptRoot$2.apply(RunTask.scala:48)
at sbt.impl.RunTask$$anonfun$runTasksExceptRoot$2.apply(RunTask.scala:48)
at sbt.Distributor$Run$Worker$$anonfun$2.apply(ParallelRunner.scala:131)
at sbt.Distributor$Run$Worker$$anonfun$2.apply(ParallelRunner.scala:131)
at sbt.Control$.trapUnit(Control.scala:19)
at sbt.Distributor$Run$Worker.run(ParallelRunner.scala:131)
Caused by: java.lang.UnsatisfiedLinkError: no snappyjava in java.library.path
at java.lang.ClassLoader.loadLibrary(ClassLoader.java:1734)
at java.lang.Runtime.loadLibrary0(Runtime.java:823)
at java.lang.System.loadLibrary(System.java:1028)
at org.xerial.snappy.SnappyNativeLoader.loadLibrary(SnappyNativeLoader.java:52)
... 51 more
[0m[[0minfo[0m] [0mTest Passed: testSimpleCompressDecompress[0m
[0m[[0minfo[0m] [0mTest Starting: testComplexCompressDecompress[0m
[0m[[0minfo[0m] [0mTest Passed: testComplexCompressDecompress[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.message.MessageCompressionTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.log.LogTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testTimeBasedLogRoll[0m
[0m[[0minfo[0m] [0mTest Passed: testTimeBasedLogRoll[0m
[0m[[0minfo[0m] [0mTest Starting: testSizeBasedLogRoll[0m
[0m[[0minfo[0m] [0mTest Passed: testSizeBasedLogRoll[0m
[0m[[0minfo[0m] [0mTest Starting: testLoadEmptyLog[0m
[0m[[0minfo[0m] [0mTest Passed: testLoadEmptyLog[0m
[0m[[0minfo[0m] [0mTest Starting: testAppendAndRead[0m
[0m[[0minfo[0m] [0mTest Passed: testAppendAndRead[0m
[0m[[0minfo[0m] [0mTest Starting: testReadOutOfRange[0m
[0m[[0minfo[0m] [0mTest Passed: testReadOutOfRange[0m
[0m[[0minfo[0m] [0mTest Starting: testLogRolls[0m
[0m[[0minfo[0m] [0mTest Passed: testLogRolls[0m
[0m[[0minfo[0m] [0mTest Starting: testCompressedMessages[0m
[0m[[0minfo[0m] [0mTest Passed: testCompressedMessages[0m
[0m[[0minfo[0m] [0mTest Starting: testFindSegment[0m
[0m[[0minfo[0m] [0mTest Passed: testFindSegment[0m
[0m[[0minfo[0m] [0mTest Starting: testEdgeLogRollsStartingAtZero[0m
[0m[[0minfo[0m] [0mTest Passed: testEdgeLogRollsStartingAtZero[0m
[0m[[0minfo[0m] [0mTest Starting: testEdgeLogRollsStartingAtNonZero[0m
[0m[[0minfo[0m] [0mTest Passed: testEdgeLogRollsStartingAtNonZero[0m
[0m[[0minfo[0m] [0mTest Starting: testMessageSizeCheck[0m
[0m[[0minfo[0m] [0mTest Passed: testMessageSizeCheck[0m
[0m[[0minfo[0m] [0mTest Starting: testLogRecoversToCorrectOffset[0m
[0m[[0minfo[0m] [0mTest Passed: testLogRecoversToCorrectOffset[0m
[0m[[0minfo[0m] [0mTest Starting: testTruncateTo[0m
[0m[[0minfo[0m] [0mTest Passed: testTruncateTo[0m
[0m[[0minfo[0m] [0mTest Starting: testIndexResizingAtTruncation[0m
[0m[[0minfo[0m] [0mTest Passed: testIndexResizingAtTruncation[0m
[0m[[0minfo[0m] [0mTest Starting: testAppendWithoutOffsetAssignment[0m
[0m[[0minfo[0m] [0mTest Passed: testAppendWithoutOffsetAssignment[0m
[0m[[0minfo[0m] [0mTest Starting: testReopenThenTruncate[0m
[0m[[0minfo[0m] [0mTest Passed: testReopenThenTruncate[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.log.LogTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.log.FileMessageSetTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testWrittenEqualsRead[0m
[0m[[0minfo[0m] [0mTest Passed: testWrittenEqualsRead[0m
[0m[[0minfo[0m] [0mTest Starting: testIteratorIsConsistent[0m
[0m[[0minfo[0m] [0mTest Passed: testIteratorIsConsistent[0m
[0m[[0minfo[0m] [0mTest Starting: testSizeInBytes[0m
[0m[[0minfo[0m] [0mTest Passed: testSizeInBytes[0m
[0m[[0minfo[0m] [0mTest Starting: testWriteTo[0m
[0m[[0minfo[0m] [0mTest Passed: testWriteTo[0m
[0m[[0minfo[0m] [0mTest Starting: testFileSize[0m
[0m[[0minfo[0m] [0mTest Passed: testFileSize[0m
[0m[[0minfo[0m] [0mTest Starting: testIterationOverPartialAndTruncation[0m
[0m[[0minfo[0m] [0mTest Passed: testIterationOverPartialAndTruncation[0m
[0m[[0minfo[0m] [0mTest Starting: testIterationDoesntChangePosition[0m
[0m[[0minfo[0m] [0mTest Passed: testIterationDoesntChangePosition[0m
[0m[[0minfo[0m] [0mTest Starting: testRead[0m
[0m[[0minfo[0m] [0mTest Passed: testRead[0m
[0m[[0minfo[0m] [0mTest Starting: testSearch[0m
[0m[[0minfo[0m] [0mTest Passed: testSearch[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.log.FileMessageSetTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.server.ReplicaFetchTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testReplicaFetcherThread(kafka.server.ReplicaFetchTest)[0m
[0m[[0minfo[0m] [0mTest Passed: testReplicaFetcherThread(kafka.server.ReplicaFetchTest)[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.server.ReplicaFetchTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.javaapi.message.ByteBufferMessageSetTest ==[0m
[0m[[0minfo[0m] [0mTest Starting: testWrittenEqualsRead[0m
[0m[[0minfo[0m] [0mTest Passed: testWrittenEqualsRead[0m
[0m[[0minfo[0m] [0mTest Starting: testIteratorIsConsistent[0m
[0m[[0minfo[0m] [0mTest Passed: testIteratorIsConsistent[0m
[0m[[0minfo[0m] [0mTest Starting: testSizeInBytes[0m
[0m[[0minfo[0m] [0mTest Passed: testSizeInBytes[0m
[0m[[0minfo[0m] [0mTest Starting: testEquals[0m
[0m[[0minfo[0m] [0mTest Passed: testEquals[0m
[0m[[0minfo[0m] [0mTest Starting: testIteratorIsConsistentWithCompression[0m
[0m[[0minfo[0m] [0mTest Passed: testIteratorIsConsistentWithCompression[0m
[0m[[0minfo[0m] [0mTest Starting: testSizeInBytesWithCompression[0m
[0m[[0minfo[0m] [0mTest Passed: testSizeInBytesWithCompression[0m
[0m[[0minfo[0m] [0mTest Starting: testEqualsWithCompression[0m
[0m[[0minfo[0m] [0mTest Passed: testEqualsWithCompression[0m
[0m[[0minfo[0m] [34m== core-kafka / kafka.javaapi.message.ByteBufferMessageSetTest ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / test-finish ==[0m
[0m[[31merror[0m] [0mFailed: : Total 167, Failed 1, Errors 0, Passed 166, Skipped 0[0m
[0m[[0minfo[0m] [34m== core-kafka / test-finish ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / Test cleanup 1 ==[0m
[0m[[0minfo[0m] [0mDeleting directory /var/tmp/sbt_32c34d6e[0m
[0m[[0minfo[0m] [34m== core-kafka / Test cleanup 1 ==[0m
[0m[[0minfo[0m] [34m[0m
[0m[[0minfo[0m] [34m== core-kafka / test-cleanup ==[0m
[0m[[0minfo[0m] [34m== core-kafka / test-cleanup ==[0m
[0m[[31merror[0m] [0mError running kafka.admin.AdminTest: Test FAILED[0m
[0m[[31merror[0m] [0mError running test: One or more subtasks failed[0m
[0m[[0minfo[0m] [0m[0m
[0m[[0minfo[0m] [0mTotal time: 305 s, completed Nov 26, 2012 9:09:22 PM[0m
[0m[[0minfo[0m] [0m[0m
[0m[[0minfo[0m] [0mTotal session time: 306 s, completed Nov 26, 2012 9:09:22 PM[0m
[0m[[31merror[0m] [0mError during build.[0m
Build step 'Execute shell' marked build as failure
Re: Build failed in Jenkins: Kafka-0.8 #113
Posted by Jay Kreps <ja...@gmail.com>.
Hey Jun,
I think AdminTest.testShutdownBroker started failing consistently after
this checkin.
-Jay
On Sun, Nov 18, 2012 at 10:09 PM, Apache Jenkins Server <
jenkins@builds.apache.org> wrote:
> See <https://builds.apache.org/job/Kafka-0.8/113/changes>
>
> Changes:
>
> [junrao] move shutting down of fetcher thread out of critical path;
> patched by Jun Rao; reviewed by Neha Narkhede; KAFKA-612
>
> ------------------------------------------
> [...truncated 2862 lines...]
> at
> kafka.server.KafkaApis.kafka$server$KafkaApis$$readMessageSet(KafkaApis.scala:359)
> at
> kafka.server.KafkaApis$$anonfun$kafka$server$KafkaApis$$readMessageSets$1.apply(KafkaApis.scala:325)
> at
> kafka.server.KafkaApis$$anonfun$kafka$server$KafkaApis$$readMessageSets$1.apply(KafkaApis.scala:321)
> at
> scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
> at
> scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
> at scala.collection.immutable.Map$Map4.foreach(Map.scala:180)
> at
> scala.collection.TraversableLike$class.map(TraversableLike.scala:206)
> at scala.collection.immutable.Map$Map4.map(Map.scala:157)
> at
> kafka.server.KafkaApis.kafka$server$KafkaApis$$readMessageSets(KafkaApis.scala:321)
> at kafka.server.KafkaApis.handleFetchRequest(KafkaApis.scala:289)
> at kafka.server.KafkaApis.handle(KafkaApis.scala:57)
> at
> kafka.server.KafkaRequestHandler.run(KafkaRequestHandler.scala:41)
> at java.lang.Thread.run(Thread.java:662)
> [2012-11-19 06:09:14,085] ERROR [KafkaApi-0] error when processing request
> (test1,-1,0,10000) (kafka.server.KafkaApis:102)
> kafka.common.UnknownTopicOrPartitionException: Topic test1 partition -1
> doesn't exist on 0
> at
> kafka.server.ReplicaManager.getLeaderReplicaIfLocal(ReplicaManager.scala:163)
> at
> kafka.server.KafkaApis.kafka$server$KafkaApis$$readMessageSet(KafkaApis.scala:359)
> at
> kafka.server.KafkaApis$$anonfun$kafka$server$KafkaApis$$readMessageSets$1.apply(KafkaApis.scala:325)
> at
> kafka.server.KafkaApis$$anonfun$kafka$server$KafkaApis$$readMessageSets$1.apply(KafkaApis.scala:321)
> at
> scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
> at
> scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
> at scala.collection.immutable.Map$Map4.foreach(Map.scala:180)
> at
> scala.collection.TraversableLike$class.map(TraversableLike.scala:206)
> at scala.collection.immutable.Map$Map4.map(Map.scala:157)
> at
> kafka.server.KafkaApis.kafka$server$KafkaApis$$readMessageSets(KafkaApis.scala:321)
> at kafka.server.KafkaApis.handleFetchRequest(KafkaApis.scala:289)
> at kafka.server.KafkaApis.handle(KafkaApis.scala:57)
> at
> kafka.server.KafkaRequestHandler.run(KafkaRequestHandler.scala:41)
> at java.lang.Thread.run(Thread.java:662)
> [0m[ [0minfo [0m] [0mTest Passed:
> testProduceAndMultiFetch(kafka.integration.PrimitiveApiTest) [0m
> [0m[ [0minfo [0m] [0mTest Starting:
> testMultiProduce(kafka.integration.PrimitiveApiTest) [0m
> [0m[ [0minfo [0m] [0mTest Passed:
> testMultiProduce(kafka.integration.PrimitiveApiTest) [0m
> [0m[ [0minfo [0m] [0mTest Starting:
> testFetchRequestCanProperlySerialize(kafka.integration.PrimitiveApiTest) [0m
> [0m[ [0minfo [0m] [0mTest Passed:
> testFetchRequestCanProperlySerialize(kafka.integration.PrimitiveApiTest) [0m
> [0m[ [0minfo [0m] [0mTest Starting:
> testEmptyFetchRequest(kafka.integration.PrimitiveApiTest) [0m
> [0m[ [0minfo [0m] [0mTest Passed:
> testEmptyFetchRequest(kafka.integration.PrimitiveApiTest) [0m
> [0m[ [0minfo [0m] [0mTest Starting:
> testDefaultEncoderProducerAndFetch(kafka.integration.PrimitiveApiTest) [0m
> [0m[ [0minfo [0m] [0mTest Passed:
> testDefaultEncoderProducerAndFetch(kafka.integration.PrimitiveApiTest) [0m
> [0m[ [0minfo [0m] [0mTest Starting:
> testDefaultEncoderProducerAndFetchWithCompression(kafka.integration.PrimitiveApiTest)
> [0m
> [0m[ [0minfo [0m] [0mTest Passed:
> testDefaultEncoderProducerAndFetchWithCompression(kafka.integration.PrimitiveApiTest)
> [0m
> [0m[ [0minfo [0m] [0mTest Starting:
> testProduceAndMultiFetchWithCompression(kafka.integration.PrimitiveApiTest)
> [0m
> [2012-11-19 06:09:18,635] ERROR [KafkaApi-0] error when processing request
> (test2,0,-1,10000) (kafka.server.KafkaApis:102)
> kafka.common.OffsetOutOfRangeException: Request for offset -1 but we only
> have log segments in the range 0 to 2.
> at kafka.log.Log.read(Log.scala:371)
> at
> kafka.server.KafkaApis.kafka$server$KafkaApis$$readMessageSet(KafkaApis.scala:368)
> at
> kafka.server.KafkaApis$$anonfun$kafka$server$KafkaApis$$readMessageSets$1.apply(KafkaApis.scala:325)
> at
> kafka.server.KafkaApis$$anonfun$kafka$server$KafkaApis$$readMessageSets$1.apply(KafkaApis.scala:321)
> at
> scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
> at
> scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
> at scala.collection.immutable.Map$Map4.foreach(Map.scala:180)
> at
> scala.collection.TraversableLike$class.map(TraversableLike.scala:206)
> at scala.collection.immutable.Map$Map4.map(Map.scala:157)
> at
> kafka.server.KafkaApis.kafka$server$KafkaApis$$readMessageSets(KafkaApis.scala:321)
> at kafka.server.KafkaApis.handleFetchRequest(KafkaApis.scala:289)
> at kafka.server.KafkaApis.handle(KafkaApis.scala:57)
> at
> kafka.server.KafkaRequestHandler.run(KafkaRequestHandler.scala:41)
> at java.lang.Thread.run(Thread.java:662)
> [2012-11-19 06:09:18,636] ERROR [KafkaApi-0] error when processing request
> (test3,0,-1,10000) (kafka.server.KafkaApis:102)
> kafka.common.OffsetOutOfRangeException: Request for offset -1 but we only
> have log segments in the range 0 to 2.
> at kafka.log.Log.read(Log.scala:371)
> at
> kafka.server.KafkaApis.kafka$server$KafkaApis$$readMessageSet(KafkaApis.scala:368)
> at
> kafka.server.KafkaApis$$anonfun$kafka$server$KafkaApis$$readMessageSets$1.apply(KafkaApis.scala:325)
> at
> kafka.server.KafkaApis$$anonfun$kafka$server$KafkaApis$$readMessageSets$1.apply(KafkaApis.scala:321)
> at
> scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
> at
> scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
> at scala.collection.immutable.Map$Map4.foreach(Map.scala:180)
> at
> scala.collection.TraversableLike$class.map(TraversableLike.scala:206)
> at scala.collection.immutable.Map$Map4.map(Map.scala:157)
> at
> kafka.server.KafkaApis.kafka$server$KafkaApis$$readMessageSets(KafkaApis.scala:321)
> at kafka.server.KafkaApis.handleFetchRequest(KafkaApis.scala:289)
> at kafka.server.KafkaApis.handle(KafkaApis.scala:57)
> at
> kafka.server.KafkaRequestHandler.run(KafkaRequestHandler.scala:41)
> at java.lang.Thread.run(Thread.java:662)
> [2012-11-19 06:09:18,636] ERROR [KafkaApi-0] error when processing request
> (test4,0,-1,10000) (kafka.server.KafkaApis:102)
> kafka.common.OffsetOutOfRangeException: Request for offset -1 but we only
> have log segments in the range 0 to 2.
> at kafka.log.Log.read(Log.scala:371)
> at
> kafka.server.KafkaApis.kafka$server$KafkaApis$$readMessageSet(KafkaApis.scala:368)
> at
> kafka.server.KafkaApis$$anonfun$kafka$server$KafkaApis$$readMessageSets$1.apply(KafkaApis.scala:325)
> at
> kafka.server.KafkaApis$$anonfun$kafka$server$KafkaApis$$readMessageSets$1.apply(KafkaApis.scala:321)
> at
> scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
> at
> scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
> at scala.collection.immutable.Map$Map4.foreach(Map.scala:180)
> at
> scala.collection.TraversableLike$class.map(TraversableLike.scala:206)
> at scala.collection.immutable.Map$Map4.map(Map.scala:157)
> at
> kafka.server.KafkaApis.kafka$server$KafkaApis$$readMessageSets(KafkaApis.scala:321)
> at kafka.server.KafkaApis.handleFetchRequest(KafkaApis.scala:289)
> at kafka.server.KafkaApis.handle(KafkaApis.scala:57)
> at
> kafka.server.KafkaRequestHandler.run(KafkaRequestHandler.scala:41)
> at java.lang.Thread.run(Thread.java:662)
> [2012-11-19 06:09:18,637] ERROR [KafkaApi-0] error when processing request
> (test1,0,-1,10000) (kafka.server.KafkaApis:102)
> kafka.common.OffsetOutOfRangeException: Request for offset -1 but we only
> have log segments in the range 0 to 2.
> at kafka.log.Log.read(Log.scala:371)
> at
> kafka.server.KafkaApis.kafka$server$KafkaApis$$readMessageSet(KafkaApis.scala:368)
> at
> kafka.server.KafkaApis$$anonfun$kafka$server$KafkaApis$$readMessageSets$1.apply(KafkaApis.scala:325)
> at
> kafka.server.KafkaApis$$anonfun$kafka$server$KafkaApis$$readMessageSets$1.apply(KafkaApis.scala:321)
> at
> scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
> at
> scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
> at scala.collection.immutable.Map$Map4.foreach(Map.scala:180)
> at
> scala.collection.TraversableLike$class.map(TraversableLike.scala:206)
> at scala.collection.immutable.Map$Map4.map(Map.scala:157)
> at
> kafka.server.KafkaApis.kafka$server$KafkaApis$$readMessageSets(KafkaApis.scala:321)
> at kafka.server.KafkaApis.handleFetchRequest(KafkaApis.scala:289)
> at kafka.server.KafkaApis.handle(KafkaApis.scala:57)
> at
> kafka.server.KafkaRequestHandler.run(KafkaRequestHandler.scala:41)
> at java.lang.Thread.run(Thread.java:662)
> [2012-11-19 06:09:18,639] ERROR [KafkaApi-0] error when processing request
> (test2,-1,0,10000) (kafka.server.KafkaApis:102)
> kafka.common.UnknownTopicOrPartitionException: Topic test2 partition -1
> doesn't exist on 0
> at
> kafka.server.ReplicaManager.getLeaderReplicaIfLocal(ReplicaManager.scala:163)
> at
> kafka.server.KafkaApis.kafka$server$KafkaApis$$readMessageSet(KafkaApis.scala:359)
> at
> kafka.server.KafkaApis$$anonfun$kafka$server$KafkaApis$$readMessageSets$1.apply(KafkaApis.scala:325)
> at
> kafka.server.KafkaApis$$anonfun$kafka$server$KafkaApis$$readMessageSets$1.apply(KafkaApis.scala:321)
> at
> scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
> at
> scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
> at scala.collection.immutable.Map$Map4.foreach(Map.scala:180)
> at
> scala.collection.TraversableLike$class.map(TraversableLike.scala:206)
> at scala.collection.immutable.Map$Map4.map(Map.scala:157)
> at
> kafka.server.KafkaApis.kafka$server$KafkaApis$$readMessageSets(KafkaApis.scala:321)
> at kafka.server.KafkaApis.handleFetchRequest(KafkaApis.scala:289)
> at kafka.server.KafkaApis.handle(KafkaApis.scala:57)
> at
> kafka.server.KafkaRequestHandler.run(KafkaRequestHandler.scala:41)
> at java.lang.Thread.run(Thread.java:662)
> [2012-11-19 06:09:18,640] ERROR [KafkaApi-0] error when processing request
> (test3,-1,0,10000) (kafka.server.KafkaApis:102)
> kafka.common.UnknownTopicOrPartitionException: Topic test3 partition -1
> doesn't exist on 0
> at
> kafka.server.ReplicaManager.getLeaderReplicaIfLocal(ReplicaManager.scala:163)
> at
> kafka.server.KafkaApis.kafka$server$KafkaApis$$readMessageSet(KafkaApis.scala:359)
> at
> kafka.server.KafkaApis$$anonfun$kafka$server$KafkaApis$$readMessageSets$1.apply(KafkaApis.scala:325)
> at
> kafka.server.KafkaApis$$anonfun$kafka$server$KafkaApis$$readMessageSets$1.apply(KafkaApis.scala:321)
> at
> scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
> at
> scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
> at scala.collection.immutable.Map$Map4.foreach(Map.scala:180)
> at
> scala.collection.TraversableLike$class.map(TraversableLike.scala:206)
> at scala.collection.immutable.Map$Map4.map(Map.scala:157)
> at
> kafka.server.KafkaApis.kafka$server$KafkaApis$$readMessageSets(KafkaApis.scala:321)
> at kafka.server.KafkaApis.handleFetchRequest(KafkaApis.scala:289)
> at kafka.server.KafkaApis.handle(KafkaApis.scala:57)
> at
> kafka.server.KafkaRequestHandler.run(KafkaRequestHandler.scala:41)
> at java.lang.Thread.run(Thread.java:662)
> [2012-11-19 06:09:18,640] ERROR [KafkaApi-0] error when processing request
> (test4,-1,0,10000) (kafka.server.KafkaApis:102)
> kafka.common.UnknownTopicOrPartitionException: Topic test4 partition -1
> doesn't exist on 0
> at
> kafka.server.ReplicaManager.getLeaderReplicaIfLocal(ReplicaManager.scala:163)
> at
> kafka.server.KafkaApis.kafka$server$KafkaApis$$readMessageSet(KafkaApis.scala:359)
> at
> kafka.server.KafkaApis$$anonfun$kafka$server$KafkaApis$$readMessageSets$1.apply(KafkaApis.scala:325)
> at
> kafka.server.KafkaApis$$anonfun$kafka$server$KafkaApis$$readMessageSets$1.apply(KafkaApis.scala:321)
> at
> scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
> at
> scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
> at scala.collection.immutable.Map$Map4.foreach(Map.scala:180)
> at
> scala.collection.TraversableLike$class.map(TraversableLike.scala:206)
> at scala.collection.immutable.Map$Map4.map(Map.scala:157)
> at
> kafka.server.KafkaApis.kafka$server$KafkaApis$$readMessageSets(KafkaApis.scala:321)
> at kafka.server.KafkaApis.handleFetchRequest(KafkaApis.scala:289)
> at kafka.server.KafkaApis.handle(KafkaApis.scala:57)
> at
> kafka.server.KafkaRequestHandler.run(KafkaRequestHandler.scala:41)
> at java.lang.Thread.run(Thread.java:662)
> [2012-11-19 06:09:18,641] ERROR [KafkaApi-0] error when processing request
> (test1,-1,0,10000) (kafka.server.KafkaApis:102)
> kafka.common.UnknownTopicOrPartitionException: Topic test1 partition -1
> doesn't exist on 0
> at
> kafka.server.ReplicaManager.getLeaderReplicaIfLocal(ReplicaManager.scala:163)
> at
> kafka.server.KafkaApis.kafka$server$KafkaApis$$readMessageSet(KafkaApis.scala:359)
> at
> kafka.server.KafkaApis$$anonfun$kafka$server$KafkaApis$$readMessageSets$1.apply(KafkaApis.scala:325)
> at
> kafka.server.KafkaApis$$anonfun$kafka$server$KafkaApis$$readMessageSets$1.apply(KafkaApis.scala:321)
> at
> scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
> at
> scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
> at scala.collection.immutable.Map$Map4.foreach(Map.scala:180)
> at
> scala.collection.TraversableLike$class.map(TraversableLike.scala:206)
> at scala.collection.immutable.Map$Map4.map(Map.scala:157)
> at
> kafka.server.KafkaApis.kafka$server$KafkaApis$$readMessageSets(KafkaApis.scala:321)
> at kafka.server.KafkaApis.handleFetchRequest(KafkaApis.scala:289)
> at kafka.server.KafkaApis.handle(KafkaApis.scala:57)
> at
> kafka.server.KafkaRequestHandler.run(KafkaRequestHandler.scala:41)
> at java.lang.Thread.run(Thread.java:662)
> [0m[ [0minfo [0m] [0mTest Passed:
> testProduceAndMultiFetchWithCompression(kafka.integration.PrimitiveApiTest)
> [0m
> [0m[ [0minfo [0m] [0mTest Starting:
> testMultiProduceWithCompression(kafka.integration.PrimitiveApiTest) [0m
> [0m[ [0minfo [0m] [0mTest Passed:
> testMultiProduceWithCompression(kafka.integration.PrimitiveApiTest) [0m
> [0m[ [0minfo [0m] [0mTest Starting:
> testConsumerEmptyTopic(kafka.integration.PrimitiveApiTest) [0m
> [0m[ [0minfo [0m] [0mTest Passed:
> testConsumerEmptyTopic(kafka.integration.PrimitiveApiTest) [0m
> [0m[ [0minfo [0m] [34m== core-kafka / kafka.integration.PrimitiveApiTest
> == [0m
> [0m[ [0minfo [0m] [34m [0m
> [0m[ [0minfo [0m] [34m== core-kafka / kafka.log4j.KafkaLog4jAppenderTest
> == [0m
> [0m[ [0minfo [0m] [0mTest Starting:
> testKafkaLog4jConfigs(kafka.log4j.KafkaLog4jAppenderTest) [0m
> log4j:WARN No appenders could be found for logger
> (org.I0Itec.zkclient.ZkEventThread).
> log4j:WARN Please initialize the log4j system properly.
> [0m[ [0minfo [0m] [0mTest Passed:
> testKafkaLog4jConfigs(kafka.log4j.KafkaLog4jAppenderTest) [0m
> [0m[ [0minfo [0m] [0mTest Starting:
> testLog4jAppends(kafka.log4j.KafkaLog4jAppenderTest) [0m
> [0m[ [0minfo [0m] [0mTest Passed:
> testLog4jAppends(kafka.log4j.KafkaLog4jAppenderTest) [0m
> [0m[ [0minfo [0m] [34m== core-kafka / kafka.log4j.KafkaLog4jAppenderTest
> == [0m
> [0m[ [0minfo [0m] [34m [0m
> [0m[ [0minfo [0m] [34m== core-kafka / kafka.metrics.KafkaTimerTest == [0m
> [0m[ [0minfo [0m] [0mTest Starting:
> testKafkaTimer(kafka.metrics.KafkaTimerTest) [0m
> [0m[ [0minfo [0m] [0mTest Passed:
> testKafkaTimer(kafka.metrics.KafkaTimerTest) [0m
> [0m[ [0minfo [0m] [34m== core-kafka / kafka.metrics.KafkaTimerTest == [0m
> [0m[ [0minfo [0m] [34m [0m
> [0m[ [0minfo [0m] [34m== core-kafka /
> kafka.message.ByteBufferMessageSetTest == [0m
> [0m[ [0minfo [0m] [0mTest Starting: testWrittenEqualsRead [0m
> [0m[ [0minfo [0m] [0mTest Passed: testWrittenEqualsRead [0m
> [0m[ [0minfo [0m] [0mTest Starting: testIteratorIsConsistent [0m
> [0m[ [0minfo [0m] [0mTest Passed: testIteratorIsConsistent [0m
> [0m[ [0minfo [0m] [0mTest Starting: testSizeInBytes [0m
> [0m[ [0minfo [0m] [0mTest Passed: testSizeInBytes [0m
> [0m[ [0minfo [0m] [0mTest Starting: testEquals [0m
> [0m[ [0minfo [0m] [0mTest Passed: testEquals [0m
> [0m[ [0minfo [0m] [0mTest Starting: testWriteTo [0m
> [0m[ [0minfo [0m] [0mTest Passed: testWriteTo [0m
> [0m[ [0minfo [0m] [0mTest Starting: testValidBytes [0m
> [0m[ [0minfo [0m] [0mTest Passed: testValidBytes [0m
> [0m[ [0minfo [0m] [0mTest Starting: testValidBytesWithCompression [0m
> [0m[ [0minfo [0m] [0mTest Passed: testValidBytesWithCompression [0m
> [0m[ [0minfo [0m] [0mTest Starting: testIterator [0m
> [0m[ [0minfo [0m] [0mTest Passed: testIterator [0m
> [0m[ [0minfo [0m] [0mTest Starting: testOffsetAssignment [0m
> [0m[ [0minfo [0m] [0mTest Passed: testOffsetAssignment [0m
> [0m[ [0minfo [0m] [34m== core-kafka /
> kafka.message.ByteBufferMessageSetTest == [0m
> [0m[ [0minfo [0m] [34m [0m
> [0m[ [0minfo [0m] [34m== core-kafka / kafka.producer.ProducerTest == [0m
> [0m[ [0minfo [0m] [0mTest Starting:
> testUpdateBrokerPartitionInfo(kafka.producer.ProducerTest) [0m
> [0m[ [0minfo [0m] [0mTest Passed:
> testUpdateBrokerPartitionInfo(kafka.producer.ProducerTest) [0m
> [0m[ [0minfo [0m] [0mTest Starting:
> testSendToNewTopic(kafka.producer.ProducerTest) [0m
> [0m[ [0minfo [0m] [0mTest Passed:
> testSendToNewTopic(kafka.producer.ProducerTest) [0m
> [0m[ [0minfo [0m] [0mTest Starting:
> testSendWithDeadBroker(kafka.producer.ProducerTest) [0m
> [0m[ [0minfo [0m] [0mTest Passed:
> testSendWithDeadBroker(kafka.producer.ProducerTest) [0m
> [0m[ [0minfo [0m] [0mTest Starting:
> testAsyncSendCanCorrectlyFailWithTimeout(kafka.producer.ProducerTest) [0m
> [0m[ [0minfo [0m] [0mTest Passed:
> testAsyncSendCanCorrectlyFailWithTimeout(kafka.producer.ProducerTest) [0m
> [0m[ [0minfo [0m] [34m== core-kafka / kafka.producer.ProducerTest == [0m
> [0m[ [0minfo [0m] [34m [0m
> [0m[ [0minfo [0m] [34m== core-kafka / kafka.integration.FetcherTest ==
> [0m
> [0m[ [0minfo [0m] [0mTest Starting:
> testFetcher(kafka.integration.FetcherTest) [0m
> [0m[ [0minfo [0m] [0mTest Passed:
> testFetcher(kafka.integration.FetcherTest) [0m
> [0m[ [0minfo [0m] [34m== core-kafka / kafka.integration.FetcherTest ==
> [0m
> [0m[ [0minfo [0m] [34m [0m
> [0m[ [0minfo [0m] [34m== core-kafka / Test cleanup 1 == [0m
> [0m[ [0minfo [0m] [0mDeleting directory /tmp/sbt_b5aaac46 [0m
> [0m[ [0minfo [0m] [34m== core-kafka / Test cleanup 1 == [0m
> [0m[ [0minfo [0m] [34m [0m
> [0m[ [0minfo [0m] [34m== core-kafka / test-finish == [0m
> [0m[ [31merror [0m] [0mFailed: : Total 167, Failed 2, Errors 0, Passed
> 165, Skipped 0 [0m
> [0m[ [0minfo [0m] [34m== core-kafka / test-finish == [0m
> [0m[ [0minfo [0m] [34m [0m
> [0m[ [0minfo [0m] [34m== core-kafka / test-cleanup == [0m
> [0m[ [0minfo [0m] [34m== core-kafka / test-cleanup == [0m
> [0m[ [0minfo [0m] [34m [0m
> [0m[ [0minfo [0m] [34m== hadoop consumer / copy-test-resources == [0m
> [0m[ [0minfo [0m] [34m== hadoop consumer / copy-test-resources == [0m
> [0m[ [31merror [0m] [0mError running kafka.server.LogRecoveryTest: Test
> FAILED [0m
> [0m[ [31merror [0m] [0mError running kafka.admin.AdminTest: Test FAILED
> [0m
> [0m[ [31merror [0m] [0mError running test: One or more subtasks failed
> [0m
> [0m[ [0minfo [0m] [0m [0m
> [0m[ [0minfo [0m] [0mTotal time: 233 s, completed Nov 19, 2012 6:09:42
> AM [0m
> [0m[ [0minfo [0m] [0m [0m
> [0m[ [0minfo [0m] [0mTotal session time: 233 s, completed Nov 19, 2012
> 6:09:42 AM [0m
> [0m[ [31merror [0m] [0mError during build. [0m
> Build step 'Execute shell' marked build as failure
>