You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@kafka.apache.org by Apache Jenkins Server <je...@builds.apache.org> on 2012/09/13 07:13:39 UTC

Build failed in Jenkins: Kafka-0.8 #42

See <https://builds.apache.org/job/Kafka-0.8/42/changes>

Changes:

[junrao] Improve Kafka internal metrics; patched by Jun Rao; reviewed by Joel Koshy and Neha Narkhede; KAFKA-203

------------------------------------------
[...truncated 834 lines...]
	at kafka.server.ReplicaManager.becomeLeaderOrFollower(ReplicaManager.scala:144)
	at kafka.server.KafkaApis.handleLeaderAndISRRequest(KafkaApis.scala:73)
	at kafka.server.KafkaApis.handle(KafkaApis.scala:60)
	at kafka.server.KafkaRequestHandler.run(KafkaRequestHandler.scala:40)
	at java.lang.Thread.run(Thread.java:662)
[info] Test Passed: testUpdateBrokerPartitionInfo(kafka.producer.ProducerTest)
[info] Test Starting: testSendToNewTopic(kafka.producer.ProducerTest)
[info] Test Passed: testSendToNewTopic(kafka.producer.ProducerTest)
[info] Test Starting: testSendWithDeadBroker(kafka.producer.ProducerTest)
[info] Test Passed: testSendWithDeadBroker(kafka.producer.ProducerTest)
[info] Test Starting: testAsyncSendCanCorrectlyFailWithTimeout(kafka.producer.ProducerTest)
[info] Test Passed: testAsyncSendCanCorrectlyFailWithTimeout(kafka.producer.ProducerTest)
[info] == core-kafka / kafka.producer.ProducerTest ==
[info] 
[info] == core-kafka / kafka.consumer.TopicFilterTest ==
[info] Test Starting: testWhitelists
[info] Test Passed: testWhitelists
[info] Test Starting: testBlacklists
[info] Test Passed: testBlacklists
[info] == core-kafka / kafka.consumer.TopicFilterTest ==
[info] 
[info] == core-kafka / kafka.integration.TopicMetadataTest ==
[info] Test Starting: testTopicMetadataRequest(kafka.integration.TopicMetadataTest)
[info] Test Passed: testTopicMetadataRequest(kafka.integration.TopicMetadataTest)
[info] Test Starting: testBasicTopicMetadata(kafka.integration.TopicMetadataTest)
[info] Test Passed: testBasicTopicMetadata(kafka.integration.TopicMetadataTest)
[info] Test Starting: testAutoCreateTopic(kafka.integration.TopicMetadataTest)
[info] Test Passed: testAutoCreateTopic(kafka.integration.TopicMetadataTest)
[info] == core-kafka / kafka.integration.TopicMetadataTest ==
[info] 
[info] == core-kafka / kafka.server.HighwatermarkPersistenceTest ==
[info] Test Starting: testHighWatermarkPersistenceSinglePartition(kafka.server.HighwatermarkPersistenceTest)
[info] Test Passed: testHighWatermarkPersistenceSinglePartition(kafka.server.HighwatermarkPersistenceTest)
[info] Test Starting: testHighWatermarkPersistenceMultiplePartitions(kafka.server.HighwatermarkPersistenceTest)
[info] Test Passed: testHighWatermarkPersistenceMultiplePartitions(kafka.server.HighwatermarkPersistenceTest)
[info] == core-kafka / kafka.server.HighwatermarkPersistenceTest ==
[info] 
[info] == core-kafka / unit.kafka.metrics.KafkaTimerTest ==
[info] Test Starting: testKafkaTimer(unit.kafka.metrics.KafkaTimerTest)
[info] Test Passed: testKafkaTimer(unit.kafka.metrics.KafkaTimerTest)
[info] == core-kafka / unit.kafka.metrics.KafkaTimerTest ==
[info] 
[info] == core-kafka / kafka.server.LogRecoveryTest ==
[info] Test Starting: testHWCheckpointNoFailuresSingleLogSegment(kafka.server.LogRecoveryTest)
[info] Test Passed: testHWCheckpointNoFailuresSingleLogSegment(kafka.server.LogRecoveryTest)
[info] Test Starting: testHWCheckpointWithFailuresSingleLogSegment(kafka.server.LogRecoveryTest)
[info] Test Passed: testHWCheckpointWithFailuresSingleLogSegment(kafka.server.LogRecoveryTest)
[info] Test Starting: testHWCheckpointNoFailuresMultipleLogSegments(kafka.server.LogRecoveryTest)
[info] Test Passed: testHWCheckpointNoFailuresMultipleLogSegments(kafka.server.LogRecoveryTest)
[info] Test Starting: testHWCheckpointWithFailuresMultipleLogSegments(kafka.server.LogRecoveryTest)
[info] Test Passed: testHWCheckpointWithFailuresMultipleLogSegments(kafka.server.LogRecoveryTest)
[info] == core-kafka / kafka.server.LogRecoveryTest ==
[info] 
[info] == core-kafka / kafka.consumer.ZookeeperConsumerConnectorTest ==
[info] Test Starting: testBasic(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] Test Passed: testBasic(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] Test Starting: testCompression(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] Test Passed: testCompression(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] Test Starting: testCompressionSetConsumption(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] Test Passed: testCompressionSetConsumption(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] Test Starting: testConsumerDecoder(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] Test Passed: testConsumerDecoder(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] Test Starting: testLeaderSelectionForPartition(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] Test Passed: testLeaderSelectionForPartition(kafka.consumer.ZookeeperConsumerConnectorTest)
[info] == core-kafka / kafka.consumer.ZookeeperConsumerConnectorTest ==
[info] 
[info] == core-kafka / kafka.log.LogTest ==
[info] Test Starting: testTimeBasedLogRoll
[info] Test Passed: testTimeBasedLogRoll
[info] Test Starting: testSizeBasedLogRoll
[info] Test Passed: testSizeBasedLogRoll
[info] Test Starting: testLoadEmptyLog
[info] Test Passed: testLoadEmptyLog
[info] Test Starting: testLoadInvalidLogsFails
[info] Test Passed: testLoadInvalidLogsFails
[info] Test Starting: testAppendAndRead
[info] Test Passed: testAppendAndRead
[info] Test Starting: testReadOutOfRange
[info] Test Passed: testReadOutOfRange
[info] Test Starting: testLogRolls
[info] Test Passed: testLogRolls
[info] Test Starting: testFindSegment
[info] Test Passed: testFindSegment
[info] Test Starting: testEdgeLogRolls
[info] Test Passed: testEdgeLogRolls
[info] Test Starting: testMessageSizeCheck
[info] Test Passed: testMessageSizeCheck
[info] == core-kafka / kafka.log.LogTest ==
[info] 
[info] == core-kafka / kafka.log.LogCorruptionTest ==
[info] Test Starting: testMessageSizeTooLarge(kafka.log.LogCorruptionTest)
[info] Test Passed: testMessageSizeTooLarge(kafka.log.LogCorruptionTest)
[info] == core-kafka / kafka.log.LogCorruptionTest ==
[info] 
[info] == core-kafka / kafka.consumer.FetcherTest ==
[info] Test Starting: testFetcher(kafka.consumer.FetcherTest)
[info] Test Passed: testFetcher(kafka.consumer.FetcherTest)
[info] == core-kafka / kafka.consumer.FetcherTest ==
[info] 
[info] == core-kafka / kafka.network.SocketServerTest ==
[info] Test Starting: simpleRequest
[info] Test Passed: simpleRequest
[info] Test Starting: tooBigRequestIsRejected
[info] Test Passed: tooBigRequestIsRejected
[info] == core-kafka / kafka.network.SocketServerTest ==
[info] 
[info] == core-kafka / kafka.server.SimpleFetchTest ==
[info] Test Starting: testNonReplicaSeesHwWhenFetching(kafka.server.SimpleFetchTest)
[info] Test Passed: testNonReplicaSeesHwWhenFetching(kafka.server.SimpleFetchTest)
[info] Test Starting: testReplicaSeesLeoWhenFetching(kafka.server.SimpleFetchTest)
[info] Test Passed: testReplicaSeesLeoWhenFetching(kafka.server.SimpleFetchTest)
[info] == core-kafka / kafka.server.SimpleFetchTest ==
[info] 
[info] == core-kafka / kafka.integration.LazyInitProducerTest ==
[info] Test Starting: testProduceAndMultiFetch(kafka.integration.LazyInitProducerTest)
[info] Test Passed: testProduceAndMultiFetch(kafka.integration.LazyInitProducerTest)
[info] Test Starting: testMultiProduce(kafka.integration.LazyInitProducerTest)
[info] Test Passed: testMultiProduce(kafka.integration.LazyInitProducerTest)
[info] Test Starting: testProduceAndFetch(kafka.integration.LazyInitProducerTest)
[info] Test Passed: testProduceAndFetch(kafka.integration.LazyInitProducerTest)
[info] Test Starting: testMultiProduceResend(kafka.integration.LazyInitProducerTest)
[info] Test Passed: testMultiProduceResend(kafka.integration.LazyInitProducerTest)
[info] == core-kafka / kafka.integration.LazyInitProducerTest ==
[info] 
[info] == core-kafka / kafka.log.LogManagerTest ==
[info] Test Starting: testCreateLog(kafka.log.LogManagerTest)
[info] Test Passed: testCreateLog(kafka.log.LogManagerTest)
[info] Test Starting: testGetLog(kafka.log.LogManagerTest)
[info] Test Passed: testGetLog(kafka.log.LogManagerTest)
[info] Test Starting: testCleanupExpiredSegments(kafka.log.LogManagerTest)
[info] Test Passed: testCleanupExpiredSegments(kafka.log.LogManagerTest)
[info] Test Starting: testCleanupSegmentsToMaintainSize(kafka.log.LogManagerTest)
[info] Test Passed: testCleanupSegmentsToMaintainSize(kafka.log.LogManagerTest)
[info] Test Starting: testTimeBasedFlush(kafka.log.LogManagerTest)
[info] Test Passed: testTimeBasedFlush(kafka.log.LogManagerTest)
[info] Test Starting: testConfigurablePartitions(kafka.log.LogManagerTest)
[info] Test Passed: testConfigurablePartitions(kafka.log.LogManagerTest)
[info] == core-kafka / kafka.log.LogManagerTest ==
[info] 
[info] == core-kafka / kafka.server.ReplicaFetchTest ==
[info] Test Starting: testReplicaFetcherThread(kafka.server.ReplicaFetchTest)
[info] Test Passed: testReplicaFetcherThread(kafka.server.ReplicaFetchTest)
[info] == core-kafka / kafka.server.ReplicaFetchTest ==
[info] 
[info] == core-kafka / kafka.producer.AsyncProducerTest ==
[info] Test Starting: testProducerQueueSize(kafka.producer.AsyncProducerTest)
[info] Test Passed: testProducerQueueSize(kafka.producer.AsyncProducerTest)
[info] Test Starting: testProduceAfterClosed(kafka.producer.AsyncProducerTest)
[info] Test Passed: testProduceAfterClosed(kafka.producer.AsyncProducerTest)
[info] Test Starting: testBatchSize(kafka.producer.AsyncProducerTest)
[info] Test Passed: testBatchSize(kafka.producer.AsyncProducerTest)
[info] Test Starting: testQueueTimeExpired(kafka.producer.AsyncProducerTest)
[info] Test Passed: testQueueTimeExpired(kafka.producer.AsyncProducerTest)
[info] Test Starting: testPartitionAndCollateEvents(kafka.producer.AsyncProducerTest)
[info] Test Passed: testPartitionAndCollateEvents(kafka.producer.AsyncProducerTest)
[info] Test Starting: testSerializeEvents(kafka.producer.AsyncProducerTest)
[info] Test Passed: testSerializeEvents(kafka.producer.AsyncProducerTest)
[info] Test Starting: testInvalidPartition(kafka.producer.AsyncProducerTest)
[info] Test Passed: testInvalidPartition(kafka.producer.AsyncProducerTest)
[info] Test Starting: testNoBroker(kafka.producer.AsyncProducerTest)
[info] Test Passed: testNoBroker(kafka.producer.AsyncProducerTest)
[info] Test Starting: testIncompatibleEncoder(kafka.producer.AsyncProducerTest)
[info] Test Passed: testIncompatibleEncoder(kafka.producer.AsyncProducerTest)
[info] Test Starting: testRandomPartitioner(kafka.producer.AsyncProducerTest)
[info] Test Passed: testRandomPartitioner(kafka.producer.AsyncProducerTest)
[info] Test Starting: testBrokerListAndAsync(kafka.producer.AsyncProducerTest)
[info] Test Passed: testBrokerListAndAsync(kafka.producer.AsyncProducerTest)
[info] Test Starting: testFailedSendRetryLogic(kafka.producer.AsyncProducerTest)
[info] Test Passed: testFailedSendRetryLogic(kafka.producer.AsyncProducerTest)
[info] Test Starting: testJavaProducer(kafka.producer.AsyncProducerTest)
[info] Test Passed: testJavaProducer(kafka.producer.AsyncProducerTest)
[info] Test Starting: testInvalidConfiguration(kafka.producer.AsyncProducerTest)
[info] Test Passed: testInvalidConfiguration(kafka.producer.AsyncProducerTest)
[info] == core-kafka / kafka.producer.AsyncProducerTest ==
[info] 
[info] == core-kafka / kafka.log.SegmentListTest ==
[info] Test Starting: testAppend
[info] Test Passed: testAppend
[info] Test Starting: testTrunc
[info] Test Passed: testTrunc
[info] Test Starting: testTruncLast
[info] Test Passed: testTruncLast
[info] Test Starting: testTruncBeyondList
[info] Test Passed: testTruncBeyondList
[info] == core-kafka / kafka.log.SegmentListTest ==
[info] 
[info] == perf / copy-resources ==
[info] == perf / copy-resources ==
[info] 
[info] == hadoop consumer / test-compile ==
[info]   Source analysis: 0 new/modified, 0 indirectly invalidated, 0 removed.
[info] Compiling test sources...
[info] Nothing to compile.
[info]   Post-analysis: 0 classes.
[info] == hadoop consumer / test-compile ==
[info] 
[info] == hadoop producer / copy-test-resources ==
[info] == hadoop producer / copy-test-resources ==
[info] 
[info] == hadoop consumer / copy-resources ==
[info] == hadoop consumer / copy-resources ==
[info] 
[info] == java-examples / copy-test-resources ==
[info] == java-examples / copy-test-resources ==
[info] 
[info] == java-examples / copy-resources ==
[info] == java-examples / copy-resources ==
[info] 
[info] == perf / test-compile ==
[info]   Source analysis: 0 new/modified, 0 indirectly invalidated, 0 removed.
[info] Compiling test sources...
[info] Nothing to compile.
[info]   Post-analysis: 0 classes.
[info] == perf / test-compile ==
[info] 
[info] == hadoop consumer / copy-test-resources ==
[info] == hadoop consumer / copy-test-resources ==
[info] 
[info] == perf / copy-test-resources ==
[info] == perf / copy-test-resources ==
[info] 
[info] == hadoop producer / copy-resources ==
[info] == hadoop producer / copy-resources ==
[info] 
[info] == hadoop producer / test-compile ==
[info]   Source analysis: 0 new/modified, 0 indirectly invalidated, 0 removed.
[info] Compiling test sources...
[info] Nothing to compile.
[info]   Post-analysis: 0 classes.
[info] == hadoop producer / test-compile ==
[info] 
[info] == core-kafka / test-finish ==
[error] Failed: : Total 138, Failed 1, Errors 0, Passed 137, Skipped 0
[info] == core-kafka / test-finish ==
[info] 
[info] == core-kafka / Test cleanup 1 ==
[info] Deleting directory /tmp/sbt_281659f6
[info] == core-kafka / Test cleanup 1 ==
[info] 
[info] == core-kafka / test-cleanup ==
[info] == core-kafka / test-cleanup ==
[error] Error running kafka.integration.PrimitiveApiTest: Test FAILED
[error] Error running test: One or more subtasks failed
[info] 
[info] Total time: 211 s, completed Sep 13, 2012 5:13:38 AM
[info] 
[info] Total session time: 211 s, completed Sep 13, 2012 5:13:38 AM
[error] Error during build.
Build step 'Execute shell' marked build as failure

Build failed in Jenkins: Kafka-0.8 #44

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/Kafka-0.8/44/changes>

Changes:

[nehanarkhede] KAFKA-449: Leader election test

------------------------------------------
[...truncated 586 lines...]
[info] == core-kafka / kafka.consumer.ZookeeperConsumerConnectorTest ==
[info] 
[info] == core-kafka / kafka.message.CompressionUtilTest ==
[info] Test Starting: testSimpleCompressDecompress
[info] Test Passed: testSimpleCompressDecompress
[info] Test Starting: testComplexCompressDecompress
[info] Test Passed: testComplexCompressDecompress
[info] Test Starting: testSnappyCompressDecompressExplicit
java.lang.reflect.InvocationTargetException
	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
	at java.lang.reflect.Method.invoke(Method.java:597)
	at org.xerial.snappy.SnappyLoader.loadNativeLibrary(SnappyLoader.java:317)
	at org.xerial.snappy.SnappyLoader.load(SnappyLoader.java:219)
	at org.xerial.snappy.Snappy.<clinit>(Snappy.java:44)
	at org.xerial.snappy.SnappyOutputStream.<init>(SnappyOutputStream.java:79)
	at org.xerial.snappy.SnappyOutputStream.<init>(SnappyOutputStream.java:66)
	at kafka.message.SnappyCompression.<init>(CompressionUtils.scala:61)
	at kafka.message.CompressionFactory$.apply(CompressionUtils.scala:82)
	at kafka.message.CompressionUtils$.compress(CompressionUtils.scala:109)
	at kafka.message.CompressionUtilTest.testSnappyCompressDecompressExplicit(CompressionUtilsTest.scala:65)
	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
	at java.lang.reflect.Method.invoke(Method.java:597)
	at org.junit.internal.runners.TestMethodRunner.executeMethodBody(TestMethodRunner.java:99)
	at org.junit.internal.runners.TestMethodRunner.runUnprotected(TestMethodRunner.java:81)
	at org.junit.internal.runners.BeforeAndAfterRunner.runProtected(BeforeAndAfterRunner.java:34)
	at org.junit.internal.runners.TestMethodRunner.runMethod(TestMethodRunner.java:75)
	at org.junit.internal.runners.TestMethodRunner.run(TestMethodRunner.java:45)
	at org.junit.internal.runners.TestClassMethodsRunner.invokeTestMethod(TestClassMethodsRunner.java:71)
	at org.junit.internal.runners.TestClassMethodsRunner.run(TestClassMethodsRunner.java:35)
	at org.junit.internal.runners.TestClassRunner$1.runUnprotected(TestClassRunner.java:42)
	at org.junit.internal.runners.BeforeAndAfterRunner.runProtected(BeforeAndAfterRunner.java:34)
	at org.junit.internal.runners.TestClassRunner.run(TestClassRunner.java:52)
	at org.junit.internal.runners.CompositeRunner.run(CompositeRunner.java:29)
	at org.junit.runner.JUnitCore.run(JUnitCore.java:121)
	at org.junit.runner.JUnitCore.run(JUnitCore.java:100)
	at org.junit.runner.JUnitCore.run(JUnitCore.java:91)
	at org.scalatest.junit.JUnitSuite$class.run(JUnitSuite.scala:261)
	at kafka.message.CompressionUtilTest.run(CompressionUtilsTest.scala:25)
	at org.scalatest.tools.ScalaTestFramework$ScalaTestRunner.run(ScalaTestFramework.scala:40)
	at sbt.TestRunner.run(TestFramework.scala:53)
	at sbt.TestRunner.runTest$1(TestFramework.scala:67)
	at sbt.TestRunner.run(TestFramework.scala:76)
	at sbt.TestFramework$$anonfun$10$$anonfun$apply$11.runTest$2(TestFramework.scala:194)
	at sbt.TestFramework$$anonfun$10$$anonfun$apply$11$$anonfun$apply$12.apply(TestFramework.scala:205)
	at sbt.TestFramework$$anonfun$10$$anonfun$apply$11$$anonfun$apply$12.apply(TestFramework.scala:205)
	at sbt.NamedTestTask.run(TestFramework.scala:92)
	at sbt.ScalaProject$$anonfun$sbt$ScalaProject$$toTask$1.apply(ScalaProject.scala:193)
	at sbt.ScalaProject$$anonfun$sbt$ScalaProject$$toTask$1.apply(ScalaProject.scala:193)
	at sbt.TaskManager$Task.invoke(TaskManager.scala:62)
	at sbt.impl.RunTask.doRun$1(RunTask.scala:77)
	at sbt.impl.RunTask.runTask(RunTask.scala:85)
	at sbt.impl.RunTask.sbt$impl$RunTask$$runIfNotRoot(RunTask.scala:60)
	at sbt.impl.RunTask$$anonfun$runTasksExceptRoot$2.apply(RunTask.scala:48)
	at sbt.impl.RunTask$$anonfun$runTasksExceptRoot$2.apply(RunTask.scala:48)
	at sbt.Distributor$Run$Worker$$anonfun$2.apply(ParallelRunner.scala:131)
	at sbt.Distributor$Run$Worker$$anonfun$2.apply(ParallelRunner.scala:131)
	at sbt.Control$.trapUnit(Control.scala:19)
	at sbt.Distributor$Run$Worker.run(ParallelRunner.scala:131)
Caused by: java.lang.UnsatisfiedLinkError: no snappyjava in java.library.path
	at java.lang.ClassLoader.loadLibrary(ClassLoader.java:1734)
	at java.lang.Runtime.loadLibrary0(Runtime.java:823)
	at java.lang.System.loadLibrary(System.java:1028)
	at org.xerial.snappy.SnappyNativeLoader.loadLibrary(SnappyNativeLoader.java:52)
	... 53 more
[error] Test Failed: testSnappyCompressDecompressExplicit
org.xerial.snappy.SnappyError: [FAILED_TO_LOAD_NATIVE_LIBRARY] null
	at org.xerial.snappy.SnappyLoader.load(SnappyLoader.java:229)
	at org.xerial.snappy.Snappy.<clinit>(Snappy.java:44)
	at org.xerial.snappy.SnappyOutputStream.<init>(SnappyOutputStream.java:79)
	at org.xerial.snappy.SnappyOutputStream.<init>(SnappyOutputStream.java:66)
	at kafka.message.SnappyCompression.<init>(CompressionUtils.scala:61)
	at kafka.message.CompressionFactory$.apply(CompressionUtils.scala:82)
	at kafka.message.CompressionUtils$.compress(CompressionUtils.scala:109)
	at kafka.message.CompressionUtilTest.testSnappyCompressDecompressExplicit(CompressionUtilsTest.scala:65)
	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
	at java.lang.reflect.Method.invoke(Method.java:597)
	at org.junit.internal.runners.TestMethodRunner.executeMethodBody(TestMethodRunner.java:99)
	at org.junit.internal.runners.TestMethodRunner.runUnprotected(TestMethodRunner.java:81)
	at org.junit.internal.runners.BeforeAndAfterRunner.runProtected(BeforeAndAfterRunner.java:34)
	at org.junit.internal.runners.TestMethodRunner.runMethod(TestMethodRunner.java:75)
	at org.junit.internal.runners.TestMethodRunner.run(TestMethodRunner.java:45)
	at org.junit.internal.runners.TestClassMethodsRunner.invokeTestMethod(TestClassMethodsRunner.java:71)
	at org.junit.internal.runners.TestClassMethodsRunner.run(TestClassMethodsRunner.java:35)
	at org.junit.internal.runners.TestClassRunner$1.runUnprotected(TestClassRunner.java:42)
	at org.junit.internal.runners.BeforeAndAfterRunner.runProtected(BeforeAndAfterRunner.java:34)
	at org.junit.internal.runners.TestClassRunner.run(TestClassRunner.java:52)
	at org.junit.internal.runners.CompositeRunner.run(CompositeRunner.java:29)
	at org.junit.runner.JUnitCore.run(JUnitCore.java:121)
	at org.junit.runner.JUnitCore.run(JUnitCore.java:100)
	at org.junit.runner.JUnitCore.run(JUnitCore.java:91)
	at org.scalatest.junit.JUnitSuite$class.run(JUnitSuite.scala:261)
	at kafka.message.CompressionUtilTest.run(CompressionUtilsTest.scala:25)
	at org.scalatest.tools.ScalaTestFramework$ScalaTestRunner.run(ScalaTestFramework.scala:40)
	at sbt.TestRunner.run(TestFramework.scala:53)
	at sbt.TestRunner.runTest$1(TestFramework.scala:67)
	at sbt.TestRunner.run(TestFramework.scala:76)
	at sbt.TestFramework$$anonfun$10$$anonfun$apply$11.runTest$2(TestFramework.scala:194)
	at sbt.TestFramework$$anonfun$10$$anonfun$apply$11$$anonfun$apply$12.apply(TestFramework.scala:205)
	at sbt.TestFramework$$anonfun$10$$anonfun$apply$11$$anonfun$apply$12.apply(TestFramework.scala:205)
	at sbt.NamedTestTask.run(TestFramework.scala:92)
	at sbt.ScalaProject$$anonfun$sbt$ScalaProject$$toTask$1.apply(ScalaProject.scala:193)
	at sbt.ScalaProject$$anonfun$sbt$ScalaProject$$toTask$1.apply(ScalaProject.scala:193)
	at sbt.TaskManager$Task.invoke(TaskManager.scala:62)
	at sbt.impl.RunTask.doRun$1(RunTask.scala:77)
	at sbt.impl.RunTask.runTask(RunTask.scala:85)
	at sbt.impl.RunTask.sbt$impl$RunTask$$runIfNotRoot(RunTask.scala:60)
	at sbt.impl.RunTask$$anonfun$runTasksExceptRoot$2.apply(RunTask.scala:48)
	at sbt.impl.RunTask$$anonfun$runTasksExceptRoot$2.apply(RunTask.scala:48)
	at sbt.Distributor$Run$Worker$$anonfun$2.apply(ParallelRunner.scala:131)
	at sbt.Distributor$Run$Worker$$anonfun$2.apply(ParallelRunner.scala:131)
	at sbt.Control$.trapUnit(Control.scala:19)
	at sbt.Distributor$Run$Worker.run(ParallelRunner.scala:131)
[info] == core-kafka / kafka.message.CompressionUtilTest ==
[info] 
[info] == core-kafka / kafka.zk.ZKEphemeralTest ==
[info] Test Starting: testEphemeralNodeCleanup(kafka.zk.ZKEphemeralTest)
[info] Test Passed: testEphemeralNodeCleanup(kafka.zk.ZKEphemeralTest)
[info] == core-kafka / kafka.zk.ZKEphemeralTest ==
[info] 
[info] == core-kafka / kafka.integration.TopicMetadataTest ==
[info] Test Starting: testTopicMetadataRequest(kafka.integration.TopicMetadataTest)
[info] Test Passed: testTopicMetadataRequest(kafka.integration.TopicMetadataTest)
[info] Test Starting: testBasicTopicMetadata(kafka.integration.TopicMetadataTest)
[info] Test Passed: testBasicTopicMetadata(kafka.integration.TopicMetadataTest)
[info] Test Starting: testAutoCreateTopic(kafka.integration.TopicMetadataTest)
[info] Test Passed: testAutoCreateTopic(kafka.integration.TopicMetadataTest)
[info] == core-kafka / kafka.integration.TopicMetadataTest ==
[info] 
[info] == core-kafka / kafka.message.ByteBufferMessageSetTest ==
[info] Test Starting: testWrittenEqualsRead
[info] Test Passed: testWrittenEqualsRead
[info] Test Starting: testIteratorIsConsistent
[info] Test Passed: testIteratorIsConsistent
[info] Test Starting: testSizeInBytes
[info] Test Passed: testSizeInBytes
[info] Test Starting: testWriteTo
[info] Test Passed: testWriteTo
[info] Test Starting: testEquals
[info] Test Passed: testEquals
[info] Test Starting: testSmallFetchSize
[info] Test Passed: testSmallFetchSize
[info] Test Starting: testValidBytes
[info] Test Passed: testValidBytes
[info] Test Starting: testValidBytesWithCompression
[info] Test Passed: testValidBytesWithCompression
[info] Test Starting: testIterator
[info] Test Passed: testIterator
[info] == core-kafka / kafka.message.ByteBufferMessageSetTest ==
[info] 
[info] == core-kafka / kafka.log.LogManagerTest ==
[info] Test Starting: testCreateLog(kafka.log.LogManagerTest)
[info] Test Passed: testCreateLog(kafka.log.LogManagerTest)
[info] Test Starting: testGetLog(kafka.log.LogManagerTest)
[info] Test Passed: testGetLog(kafka.log.LogManagerTest)
[info] Test Starting: testCleanupExpiredSegments(kafka.log.LogManagerTest)
[info] Test Passed: testCleanupExpiredSegments(kafka.log.LogManagerTest)
[info] Test Starting: testCleanupSegmentsToMaintainSize(kafka.log.LogManagerTest)
[info] Test Passed: testCleanupSegmentsToMaintainSize(kafka.log.LogManagerTest)
[info] Test Starting: testTimeBasedFlush(kafka.log.LogManagerTest)
[info] Test Passed: testTimeBasedFlush(kafka.log.LogManagerTest)
[info] == core-kafka / kafka.log.LogManagerTest ==
[info] 
[info] == core-kafka / kafka.server.ReplicaFetchTest ==
[info] Test Starting: testReplicaFetcherThread(kafka.server.ReplicaFetchTest)
[info] Test Passed: testReplicaFetcherThread(kafka.server.ReplicaFetchTest)
[info] == core-kafka / kafka.server.ReplicaFetchTest ==
[info] 
[info] == core-kafka / kafka.server.LogRecoveryTest ==
[info] Test Starting: testHWCheckpointNoFailuresSingleLogSegment(kafka.server.LogRecoveryTest)
[info] Test Passed: testHWCheckpointNoFailuresSingleLogSegment(kafka.server.LogRecoveryTest)
[info] Test Starting: testHWCheckpointWithFailuresSingleLogSegment(kafka.server.LogRecoveryTest)
[info] Test Passed: testHWCheckpointWithFailuresSingleLogSegment(kafka.server.LogRecoveryTest)
[info] Test Starting: testHWCheckpointNoFailuresMultipleLogSegments(kafka.server.LogRecoveryTest)
[info] Test Passed: testHWCheckpointNoFailuresMultipleLogSegments(kafka.server.LogRecoveryTest)
[info] Test Starting: testHWCheckpointWithFailuresMultipleLogSegments(kafka.server.LogRecoveryTest)
[info] Test Passed: testHWCheckpointWithFailuresMultipleLogSegments(kafka.server.LogRecoveryTest)
[info] == core-kafka / kafka.server.LogRecoveryTest ==
[info] 
[info] == core-kafka / kafka.admin.AdminTest ==
[info] Test Starting: testReplicaAssignment(kafka.admin.AdminTest)
[info] Test Passed: testReplicaAssignment(kafka.admin.AdminTest)
[info] Test Starting: testManualReplicaAssignment(kafka.admin.AdminTest)
[info] Test Passed: testManualReplicaAssignment(kafka.admin.AdminTest)
[info] Test Starting: testTopicCreationInZK(kafka.admin.AdminTest)
[info] Test Passed: testTopicCreationInZK(kafka.admin.AdminTest)
[info] Test Starting: testGetTopicMetadata(kafka.admin.AdminTest)
[info] Test Passed: testGetTopicMetadata(kafka.admin.AdminTest)
[info] == core-kafka / kafka.admin.AdminTest ==
[info] 
[info] == core-kafka / kafka.javaapi.consumer.ZookeeperConsumerConnectorTest ==
[info] Test Starting: testBasic(kafka.javaapi.consumer.ZookeeperConsumerConnectorTest)
[info] Test Passed: testBasic(kafka.javaapi.consumer.ZookeeperConsumerConnectorTest)
[info] == core-kafka / kafka.javaapi.consumer.ZookeeperConsumerConnectorTest ==
[info] 
[info] == core-kafka / kafka.message.MessageTest ==
[info] Test Starting: testFieldValues
[info] Test Passed: testFieldValues
[info] Test Starting: testChecksum
[info] Test Passed: testChecksum
[info] Test Starting: testEquality
[info] Test Passed: testEquality
[info] Test Starting: testIsHashable
[info] Test Passed: testIsHashable
[info] == core-kafka / kafka.message.MessageTest ==
[info] 
[info] == core-kafka / kafka.server.ISRExpirationTest ==
[info] Test Starting: testISRExpirationForStuckFollowers(kafka.server.ISRExpirationTest)
[info] Test Passed: testISRExpirationForStuckFollowers(kafka.server.ISRExpirationTest)
[info] Test Starting: testISRExpirationForSlowFollowers(kafka.server.ISRExpirationTest)
[info] Test Passed: testISRExpirationForSlowFollowers(kafka.server.ISRExpirationTest)
[info] == core-kafka / kafka.server.ISRExpirationTest ==
[info] 
[info] == core-kafka / kafka.producer.SyncProducerTest ==
[info] Test Starting: testMessageSizeTooLarge(kafka.producer.SyncProducerTest)
[info] Test Passed: testMessageSizeTooLarge(kafka.producer.SyncProducerTest)
[info] Test Starting: testReachableServer(kafka.producer.SyncProducerTest)
[info] Test Passed: testReachableServer(kafka.producer.SyncProducerTest)
[info] Test Starting: testEmptyProduceRequest(kafka.producer.SyncProducerTest)
[info] Test Passed: testEmptyProduceRequest(kafka.producer.SyncProducerTest)
[info] Test Starting: testProduceCorrectlyReceivesResponse(kafka.producer.SyncProducerTest)
[info] Test Passed: testProduceCorrectlyReceivesResponse(kafka.producer.SyncProducerTest)
[info] Test Starting: testProducerCanTimeout(kafka.producer.SyncProducerTest)
[info] Test Passed: testProducerCanTimeout(kafka.producer.SyncProducerTest)
[info] == core-kafka / kafka.producer.SyncProducerTest ==
[info] 
[info] == core-kafka / Test cleanup 1 ==
[info] Deleting directory /var/tmp/sbt_11a20027
[info] == core-kafka / Test cleanup 1 ==
[info] 
[info] == core-kafka / test-finish ==
[error] Failed: : Total 137, Failed 1, Errors 0, Passed 136, Skipped 0
[info] == core-kafka / test-finish ==
[info] 
[info] == core-kafka / test-cleanup ==
[info] == core-kafka / test-cleanup ==
[error] Error running kafka.message.CompressionUtilTest: Test FAILED
[error] Error running test: One or more subtasks failed
[info] 
[info] Total time: 221 s, completed Sep 14, 2012 6:15:11 PM
[info] 
[info] Total session time: 223 s, completed Sep 14, 2012 6:15:11 PM
[error] Error during build.
Build step 'Execute shell' marked build as failure

Build failed in Jenkins: Kafka-0.8 #43

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/Kafka-0.8/43/changes>

Changes:

[junrao] revisit broker config in 0.8; patched by Swapnil Ghike; reviewed by Jun Rao; KAFKA-325

------------------------------------------
[...truncated 5 lines...]
A         core/src/test/scala/unit/kafka/utils/TopicTest.scala
U         core/src/test/scala/unit/kafka/utils/TestUtils.scala
U         core/src/test/scala/unit/kafka/log/LogManagerTest.scala
U         core/src/test/scala/unit/kafka/integration/TopicMetadataTest.scala
U         core/src/test/scala/unit/kafka/server/SimpleFetchTest.scala
U         core/src/test/scala/unit/kafka/network/SocketServerTest.scala
U         core/src/main/scala/kafka/cluster/Replica.scala
U         core/src/main/scala/kafka/cluster/Partition.scala
D         core/src/main/scala/kafka/log/LogStats.scala
U         core/src/main/scala/kafka/log/Log.scala
U         core/src/main/scala/kafka/log/LogManager.scala
U         core/src/main/scala/kafka/producer/SyncProducer.scala
U         core/src/main/scala/kafka/producer/Producer.scala
D         core/src/main/scala/kafka/producer/async/AsyncProducerStatsMBean.scala
U         core/src/main/scala/kafka/producer/async/AsyncProducerStats.scala
U         core/src/main/scala/kafka/producer/async/DefaultEventHandler.scala
U         core/src/main/scala/kafka/producer/async/ProducerSendThread.scala
U         core/src/main/scala/kafka/message/FileMessageSet.scala
U         core/src/main/scala/kafka/admin/CreateTopicCommand.scala
D         core/src/main/scala/kafka/network/SocketServerStats.scala
U         core/src/main/scala/kafka/network/SocketServer.scala
U         core/src/main/scala/kafka/network/RequestChannel.scala
U         core/src/main/scala/kafka/consumer/ConsumerIterator.scala
U         core/src/main/scala/kafka/consumer/SimpleConsumer.scala
U         core/src/main/scala/kafka/consumer/ConsumerConnector.scala
U         core/src/main/scala/kafka/consumer/PartitionTopicInfo.scala
U         core/src/main/scala/kafka/consumer/ConsumerTopicStat.scala
U         core/src/main/scala/kafka/consumer/ZookeeperConsumerConnector.scala
U         core/src/main/scala/kafka/utils/Utils.scala
U         core/src/main/scala/kafka/utils/Pool.scala
A         core/src/main/scala/kafka/utils/Topic.scala
U         core/src/main/scala/kafka/metrics/KafkaMetricsGroup.scala
U         core/src/main/scala/kafka/server/KafkaConfig.scala
U         core/src/main/scala/kafka/server/KafkaServer.scala
U         core/src/main/scala/kafka/server/RequestPurgatory.scala
U         core/src/main/scala/kafka/server/ReplicaManager.scala
U         core/src/main/scala/kafka/server/KafkaController.scala
U         core/src/main/scala/kafka/server/KafkaRequestHandler.scala
U         core/src/main/scala/kafka/server/KafkaApis.scala
U         core/src/main/scala/kafka/server/AbstractFetcherThread.scala
U         core/src/main/scala/kafka/api/StopReplicaRequest.scala
U         core/src/main/scala/kafka/api/OffsetRequest.scala
U         core/src/main/scala/kafka/api/FetchRequest.scala
U         core/src/main/scala/kafka/api/RequestKeys.scala
U         core/src/main/scala/kafka/api/LeaderAndIsrRequest.scala
U         core/src/main/scala/kafka/api/TopicMetadataRequest.scala
U         core/src/main/scala/kafka/api/ProducerRequest.scala
U         core/src/main/scala/kafka/javaapi/ProducerRequest.scala
At revision 1384629
[Kafka-0.8] $ /bin/bash -xe /var/tmp/hudson3795267082552113480.sh
+ ./sbt update
[info] Building project Kafka 0.8.0 against Scala 2.8.0
[info]    using KafkaProject with sbt 0.7.5 and Scala 2.7.7
[info] 
[info] == core-kafka / update ==
[info] :: retrieving :: kafka#core-kafka_2.8.0 [sync]
[info] 	confs: [compile, runtime, test, provided, system, optional, sources, javadoc]
[info] 	0 artifacts copied, 12 already retrieved (0kB/87ms)
[info] == core-kafka / update ==
[info] 
[info] == hadoop producer / update ==
[info] :: retrieving :: kafka#hadoop-producer_2.8.0 [sync]
[info] 	confs: [compile, runtime, test, provided, system, optional, sources, javadoc]
[info] 	0 artifacts copied, 40 already retrieved (0kB/66ms)
[info] == hadoop producer / update ==
[info] 
[info] == hadoop consumer / update ==
[info] :: retrieving :: kafka#hadoop-consumer_2.8.0 [sync]
[info] 	confs: [compile, runtime, test, provided, system, optional, sources, javadoc]
[info] 	0 artifacts copied, 37 already retrieved (0kB/63ms)
[info] == hadoop consumer / update ==
[info] 
[info] == contrib / update ==
[warn] No dependency configuration found, using defaults.
[info] :: retrieving :: kafka#contrib_2.8.0 [sync]
[info] 	confs: [default]
[info] 	0 artifacts copied, 0 already retrieved (0kB/2ms)
[info] == contrib / update ==
[info] 
[info] == perf / update ==
[info] :: retrieving :: kafka#perf_2.8.0 [sync]
[info] 	confs: [compile, runtime, test, provided, system, optional, sources, javadoc]
[info] 	0 artifacts copied, 5 already retrieved (0kB/41ms)
[info] == perf / update ==
[info] 
[info] == java-examples / update ==
[info] :: retrieving :: kafka#java-examples_2.8.0 [sync]
[info] 	confs: [compile, runtime, test, provided, system, optional, sources, javadoc]
[info] 	0 artifacts copied, 5 already retrieved (0kB/27ms)
[info] == java-examples / update ==
[info] 
[info] == Kafka / update ==
[info] :: retrieving :: kafka#kafka_2.8.0 [sync]
[info] 	confs: [compile, runtime, test, provided, system, optional, sources, javadoc]
[info] 	0 artifacts copied, 8 already retrieved (0kB/223ms)
[info] == Kafka / update ==
[success] Successful.
[info] 
[info] Total time: 27 s, completed Sep 14, 2012 2:12:20 AM
[info] 
[info] Total session time: 32 s, completed Sep 14, 2012 2:12:20 AM
[success] Build completed successfully.
[Kafka-0.8] $ /bin/bash -xe /var/tmp/hudson3758644180412537828.sh
+ ./sbt test
[info] Building project Kafka 0.8.0 against Scala 2.8.0
[info]    using KafkaProject with sbt 0.7.5 and Scala 2.7.7
[info] 
[info] == core-kafka / compile ==
[info]   Source analysis: 39 new/modified, 83 indirectly invalidated, 3 removed.
[info] Compiling main sources...
[warn] there were unchecked warnings; re-run with -unchecked for details
[warn] one warning found
java.io.IOException: Cannot run program "javac": error=12, Not enough space
	at java.lang.ProcessBuilder.start(ProcessBuilder.java:460)
	at sbt.SimpleProcessBuilder.run(ProcessImpl.scala:381)
	at sbt.AbstractProcessBuilder.run(ProcessImpl.scala:132)
	at sbt.AbstractProcessBuilder$$anonfun$runBuffered$1.apply(ProcessImpl.scala:165)
	at sbt.AbstractProcessBuilder$$anonfun$runBuffered$1.apply(ProcessImpl.scala:165)
	at sbt.BufferedLogger.bufferAll(Logger.scala:179)
	at sbt.AbstractProcessBuilder.runBuffered(ProcessImpl.scala:165)
	at sbt.AbstractProcessBuilder.$bang(ProcessImpl.scala:160)
	at sbt.Compile.externalJavac$1(Compile.scala:94)
	at sbt.Compile$$anonfun$forkJavac$2.apply(Compile.scala:95)
	at sbt.Compile$$anonfun$forkJavac$2.apply(Compile.scala:95)
	at sbt.WithArgumentFile$$anonfun$withArgumentFile$1.apply(Compile.scala:117)
	at sbt.WithArgumentFile$$anonfun$withArgumentFile$1.apply(Compile.scala:114)
	at xsbt.FileUtilities$.withTemporaryDirectory(FileUtilities.scala:169)
	at sbt.WithArgumentFile$class.withArgumentFile(Compile.scala:114)
	at sbt.Compile.withArgumentFile(Compile.scala:71)
	at sbt.Compile.forkJavac(Compile.scala:95)
	at sbt.Compile.processJava(Compile.scala:86)
	at sbt.CompilerCore$$anonfun$2.apply(Compile.scala:28)
	at sbt.CompilerCore$$anonfun$2.apply(Compile.scala:28)
	at sbt.CompilerCore$$anonfun$process$1$1.apply(Compile.scala:22)
	at sbt.CompilerCore$$anonfun$process$1$1.apply(Compile.scala:22)
	at sbt.CompilerCore$$anonfun$doCompile$3.apply(Compile.scala:46)
	at sbt.CompilerCore$$anonfun$doCompile$3.apply(Compile.scala:42)
	at scala.Option.orElse(Option.scala:102)
	at sbt.CompilerCore.doCompile(Compile.scala:41)
	at sbt.CompilerCore.apply(Compile.scala:29)
	at sbt.AbstractCompileConditional.run$1(Conditional.scala:341)
	at sbt.AbstractCompileConditional$$anonfun$3.apply(Conditional.scala:344)
	at sbt.AbstractCompileConditional$$anonfun$3.apply(Conditional.scala:344)
	at sbt.classfile.Analyze$.apply(Analyze.scala:110)
	at sbt.AbstractCompileConditional.execute(Conditional.scala:344)
	at sbt.Conditional$class.run(Conditional.scala:43)
	at sbt.AbstractCompileConditional.run(Conditional.scala:231)
	at sbt.BasicScalaProject.sbt$BasicScalaProject$$doCompile(DefaultProject.scala:260)
	at sbt.BasicScalaProject$$anonfun$compileAction$1.apply(DefaultProject.scala:274)
	at sbt.BasicScalaProject$$anonfun$compileAction$1.apply(DefaultProject.scala:274)
	at sbt.TaskManager$Task.invoke(TaskManager.scala:62)
	at sbt.impl.RunTask.doRun$1(RunTask.scala:77)
	at sbt.impl.RunTask.runTask(RunTask.scala:85)
	at sbt.impl.RunTask.sbt$impl$RunTask$$runIfNotRoot(RunTask.scala:60)
	at sbt.impl.RunTask$$anonfun$runTasksExceptRoot$2.apply(RunTask.scala:48)
	at sbt.impl.RunTask$$anonfun$runTasksExceptRoot$2.apply(RunTask.scala:48)
	at sbt.Distributor$Run$Worker$$anonfun$2.apply(ParallelRunner.scala:131)
	at sbt.Distributor$Run$Worker$$anonfun$2.apply(ParallelRunner.scala:131)
	at sbt.Control$.trapUnit(Control.scala:19)
	at sbt.Distributor$Run$Worker.run(ParallelRunner.scala:131)
Caused by: java.io.IOException: error=12, Not enough space
	at java.lang.UNIXProcess.forkAndExec(Native Method)
	at java.lang.UNIXProcess.<init>(UNIXProcess.java:53)
	at java.lang.ProcessImpl.start(ProcessImpl.java:65)
	at java.lang.ProcessBuilder.start(ProcessBuilder.java:453)
	at sbt.SimpleProcessBuilder.run(ProcessImpl.scala:381)
	at sbt.AbstractProcessBuilder.run(ProcessImpl.scala:132)
	at sbt.AbstractProcessBuilder$$anonfun$runBuffered$1.apply(ProcessImpl.scala:165)
	at sbt.AbstractProcessBuilder$$anonfun$runBuffered$1.apply(ProcessImpl.scala:165)
	at sbt.BufferedLogger.bufferAll(Logger.scala:179)
	at sbt.AbstractProcessBuilder.runBuffered(ProcessImpl.scala:165)
	at sbt.AbstractProcessBuilder.$bang(ProcessImpl.scala:160)
	at sbt.Compile.externalJavac$1(Compile.scala:94)
	at sbt.Compile$$anonfun$forkJavac$2.apply(Compile.scala:95)
	at sbt.Compile$$anonfun$forkJavac$2.apply(Compile.scala:95)
	at sbt.WithArgumentFile$$anonfun$withArgumentFile$1.apply(Compile.scala:117)
	at sbt.WithArgumentFile$$anonfun$withArgumentFile$1.apply(Compile.scala:114)
	at xsbt.FileUtilities$.withTemporaryDirectory(FileUtilities.scala:169)
	at sbt.WithArgumentFile$class.withArgumentFile(Compile.scala:114)
	at sbt.Compile.withArgumentFile(Compile.scala:71)
	at sbt.Compile.forkJavac(Compile.scala:95)
	at sbt.Compile.processJava(Compile.scala:86)
	at sbt.CompilerCore$$anonfun$2.apply(Compile.scala:28)
	at sbt.CompilerCore$$anonfun$2.apply(Compile.scala:28)
	at sbt.CompilerCore$$anonfun$process$1$1.apply(Compile.scala:22)
	at sbt.CompilerCore$$anonfun$process$1$1.apply(Compile.scala:22)
	at sbt.CompilerCore$$anonfun$doCompile$3.apply(Compile.scala:46)
	at sbt.CompilerCore$$anonfun$doCompile$3.apply(Compile.scala:42)
	at scala.Option.orElse(Option.scala:102)
	at sbt.CompilerCore.doCompile(Compile.scala:41)
	at sbt.CompilerCore.apply(Compile.scala:29)
	at sbt.AbstractCompileConditional.run$1(Conditional.scala:341)
	at sbt.AbstractCompileConditional$$anonfun$3.apply(Conditional.scala:344)
	at sbt.AbstractCompileConditional$$anonfun$3.apply(Conditional.scala:344)
	at sbt.classfile.Analyze$.apply(Analyze.scala:110)
	at sbt.AbstractCompileConditional.execute(Conditional.scala:344)
	at sbt.Conditional$class.run(Conditional.scala:43)
	at sbt.AbstractCompileConditional.run(Conditional.scala:231)
	at sbt.BasicScalaProject.sbt$BasicScalaProject$$doCompile(DefaultProject.scala:260)
	at sbt.BasicScalaProject$$anonfun$compileAction$1.apply(DefaultProject.scala:274)
	at sbt.BasicScalaProject$$anonfun$compileAction$1.apply(DefaultProject.scala:274)
	at sbt.TaskManager$Task.invoke(TaskManager.scala:62)
	at sbt.impl.RunTask.doRun$1(RunTask.scala:77)
	at sbt.impl.RunTask.runTask(RunTask.scala:85)
	at sbt.impl.RunTask.sbt$impl$RunTask$$runIfNotRoot(RunTask.scala:60)
	at sbt.impl.RunTask$$anonfun$runTasksExceptRoot$2.apply(RunTask.scala:48)
	at sbt.impl.RunTask$$anonfun$runTasksExceptRoot$2.apply(RunTask.scala:48)
	at sbt.Distributor$Run$Worker$$anonfun$2.apply(ParallelRunner.scala:131)
	at sbt.Distributor$Run$Worker$$anonfun$2.apply(ParallelRunner.scala:131)
	at sbt.Control$.trapUnit(Control.scala:19)
	at sbt.Distributor$Run$Worker.run(ParallelRunner.scala:131)
[info] == core-kafka / compile ==
[info] 
[info] == core-kafka / copy-test-resources ==
[info] == core-kafka / copy-test-resources ==
[info] 
[info] == core-kafka / copy-resources ==
[info] == core-kafka / copy-resources ==
[info] 
[info] == perf / copy-test-resources ==
[info] == perf / copy-test-resources ==
[info] 
[info] == hadoop consumer / copy-resources ==
[info] == hadoop consumer / copy-resources ==
[info] 
[info] == hadoop consumer / copy-test-resources ==
[info] == hadoop consumer / copy-test-resources ==
[info] 
[info] == java-examples / copy-test-resources ==
[info] == java-examples / copy-test-resources ==
[info] 
[info] == perf / copy-resources ==
[info] == perf / copy-resources ==
[info] 
[info] == hadoop producer / copy-test-resources ==
[info] == hadoop producer / copy-test-resources ==
[info] 
[info] == hadoop producer / copy-resources ==
[info] == hadoop producer / copy-resources ==
[info] 
[info] == java-examples / copy-resources ==
[info] == java-examples / copy-resources ==
[error] Error running compile: java.io.IOException: Cannot run program "javac": error=12, Not enough space
[info] 
[info] Total time: 106 s, completed Sep 14, 2012 2:14:17 AM
[info] 
[info] Total session time: 109 s, completed Sep 14, 2012 2:14:17 AM
[error] Error during build.
Build step 'Execute shell' marked build as failure