You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@kafka.apache.org by Apache Jenkins Server <je...@builds.apache.org> on 2019/08/26 03:57:56 UTC
Build failed in Jenkins: kafka-trunk-jdk8 #3870
See <https://builds.apache.org/job/kafka-trunk-jdk8/3870/display/redirect?page=changes>
Changes:
[gwen] KAFKA-8391; Temporarily ignore flaky Connect rebalance integration tests
[rhauch] MINOR: Fix the doc of scheduled.rebalance.max.delay.ms config property
------------------------------------------
[...truncated 2.59 MB...]
org.apache.kafka.connect.file.FileStreamSourceConnectorTest > testTaskClass PASSED
org.apache.kafka.connect.file.FileStreamSourceConnectorTest > testInvalidBatchSize STARTED
org.apache.kafka.connect.file.FileStreamSourceConnectorTest > testInvalidBatchSize PASSED
org.apache.kafka.connect.file.FileStreamSourceConnectorTest > testConnectorConfigValidation STARTED
org.apache.kafka.connect.file.FileStreamSourceConnectorTest > testConnectorConfigValidation PASSED
org.apache.kafka.connect.file.FileStreamSourceConnectorTest > testMultipleSourcesInvalid STARTED
org.apache.kafka.connect.file.FileStreamSourceConnectorTest > testMultipleSourcesInvalid PASSED
> Task :streams:streams-scala:test
org.apache.kafka.streams.scala.kstream.KTableTest > filter a KTable should filter records satisfying the predicate STARTED
org.apache.kafka.streams.scala.kstream.KTableTest > filter a KTable should filter records satisfying the predicate PASSED
org.apache.kafka.streams.scala.kstream.KTableTest > filterNot a KTable should filter records not satisfying the predicate STARTED
org.apache.kafka.streams.scala.kstream.KTableTest > filterNot a KTable should filter records not satisfying the predicate PASSED
org.apache.kafka.streams.scala.kstream.KTableTest > join 2 KTables should join correctly records STARTED
org.apache.kafka.streams.scala.kstream.KTableTest > join 2 KTables should join correctly records PASSED
org.apache.kafka.streams.scala.kstream.KTableTest > join 2 KTables with a Materialized should join correctly records and state store STARTED
org.apache.kafka.streams.scala.kstream.KTableTest > join 2 KTables with a Materialized should join correctly records and state store PASSED
org.apache.kafka.streams.scala.kstream.KTableTest > windowed KTable#suppress should correctly suppress results using Suppressed.untilTimeLimit STARTED
org.apache.kafka.streams.scala.kstream.KTableTest > windowed KTable#suppress should correctly suppress results using Suppressed.untilTimeLimit PASSED
org.apache.kafka.streams.scala.kstream.KTableTest > windowed KTable#suppress should correctly suppress results using Suppressed.untilWindowCloses STARTED
org.apache.kafka.streams.scala.kstream.KTableTest > windowed KTable#suppress should correctly suppress results using Suppressed.untilWindowCloses PASSED
org.apache.kafka.streams.scala.kstream.KTableTest > session windowed KTable#suppress should correctly suppress results using Suppressed.untilWindowCloses STARTED
org.apache.kafka.streams.scala.kstream.KTableTest > session windowed KTable#suppress should correctly suppress results using Suppressed.untilWindowCloses PASSED
org.apache.kafka.streams.scala.kstream.KTableTest > non-windowed KTable#suppress should correctly suppress results using Suppressed.untilTimeLimit STARTED
org.apache.kafka.streams.scala.kstream.KTableTest > non-windowed KTable#suppress should correctly suppress results using Suppressed.untilTimeLimit PASSED
org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed should create a Consumed with Serdes STARTED
org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed should create a Consumed with Serdes PASSED
org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed with timestampExtractor and resetPolicy should create a Consumed with Serdes, timestampExtractor and resetPolicy STARTED
org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed with timestampExtractor and resetPolicy should create a Consumed with Serdes, timestampExtractor and resetPolicy PASSED
org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed with timestampExtractor should create a Consumed with Serdes and timestampExtractor STARTED
org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed with timestampExtractor should create a Consumed with Serdes and timestampExtractor PASSED
org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed with resetPolicy should create a Consumed with Serdes and resetPolicy STARTED
org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed with resetPolicy should create a Consumed with Serdes and resetPolicy PASSED
org.apache.kafka.streams.scala.kstream.KStreamTest > filter a KStream should filter records satisfying the predicate STARTED
org.apache.kafka.streams.scala.kstream.KStreamTest > filter a KStream should filter records satisfying the predicate PASSED
org.apache.kafka.streams.scala.kstream.KStreamTest > filterNot a KStream should filter records not satisfying the predicate STARTED
org.apache.kafka.streams.scala.kstream.KStreamTest > filterNot a KStream should filter records not satisfying the predicate PASSED
org.apache.kafka.streams.scala.kstream.KStreamTest > foreach a KStream should run foreach actions on records STARTED
org.apache.kafka.streams.scala.kstream.KStreamTest > foreach a KStream should run foreach actions on records PASSED
org.apache.kafka.streams.scala.kstream.KStreamTest > peek a KStream should run peek actions on records STARTED
org.apache.kafka.streams.scala.kstream.KStreamTest > peek a KStream should run peek actions on records PASSED
org.apache.kafka.streams.scala.kstream.KStreamTest > selectKey a KStream should select a new key STARTED
org.apache.kafka.streams.scala.kstream.KStreamTest > selectKey a KStream should select a new key PASSED
org.apache.kafka.streams.scala.kstream.KStreamTest > join 2 KStreams should join correctly records STARTED
org.apache.kafka.streams.scala.kstream.KStreamTest > join 2 KStreams should join correctly records PASSED
org.apache.kafka.streams.scala.kstream.JoinedTest > Create a Joined should create a Joined with Serdes STARTED
org.apache.kafka.streams.scala.kstream.JoinedTest > Create a Joined should create a Joined with Serdes PASSED
org.apache.kafka.streams.scala.kstream.JoinedTest > Create a Joined should create a Joined with Serdes and repartition topic name STARTED
org.apache.kafka.streams.scala.kstream.JoinedTest > Create a Joined should create a Joined with Serdes and repartition topic name PASSED
org.apache.kafka.streams.scala.kstream.ProducedTest > Create a Produced should create a Produced with Serdes STARTED
org.apache.kafka.streams.scala.kstream.ProducedTest > Create a Produced should create a Produced with Serdes PASSED
org.apache.kafka.streams.scala.kstream.ProducedTest > Create a Produced with timestampExtractor and resetPolicy should create a Consumed with Serdes, timestampExtractor and resetPolicy STARTED
org.apache.kafka.streams.scala.kstream.ProducedTest > Create a Produced with timestampExtractor and resetPolicy should create a Consumed with Serdes, timestampExtractor and resetPolicy PASSED
org.apache.kafka.streams.scala.kstream.GroupedTest > Create a Grouped should create a Grouped with Serdes STARTED
org.apache.kafka.streams.scala.kstream.GroupedTest > Create a Grouped should create a Grouped with Serdes PASSED
org.apache.kafka.streams.scala.kstream.GroupedTest > Create a Grouped with repartition topic name should create a Grouped with Serdes, and repartition topic name STARTED
org.apache.kafka.streams.scala.kstream.GroupedTest > Create a Grouped with repartition topic name should create a Grouped with Serdes, and repartition topic name PASSED
org.apache.kafka.streams.scala.kstream.SuppressedTest > Suppressed.untilWindowCloses should produce the correct suppression STARTED
org.apache.kafka.streams.scala.kstream.SuppressedTest > Suppressed.untilWindowCloses should produce the correct suppression PASSED
org.apache.kafka.streams.scala.kstream.SuppressedTest > Suppressed.untilTimeLimit should produce the correct suppression STARTED
org.apache.kafka.streams.scala.kstream.SuppressedTest > Suppressed.untilTimeLimit should produce the correct suppression PASSED
org.apache.kafka.streams.scala.kstream.SuppressedTest > BufferConfig.maxRecords should produce the correct buffer config STARTED
org.apache.kafka.streams.scala.kstream.SuppressedTest > BufferConfig.maxRecords should produce the correct buffer config PASSED
org.apache.kafka.streams.scala.kstream.SuppressedTest > BufferConfig.maxBytes should produce the correct buffer config STARTED
org.apache.kafka.streams.scala.kstream.SuppressedTest > BufferConfig.maxBytes should produce the correct buffer config PASSED
org.apache.kafka.streams.scala.kstream.SuppressedTest > BufferConfig.unbounded should produce the correct buffer config STARTED
org.apache.kafka.streams.scala.kstream.SuppressedTest > BufferConfig.unbounded should produce the correct buffer config PASSED
org.apache.kafka.streams.scala.kstream.SuppressedTest > BufferConfig should support very long chains of factory methods STARTED
org.apache.kafka.streams.scala.kstream.SuppressedTest > BufferConfig should support very long chains of factory methods PASSED
org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialized should create a Materialized with Serdes STARTED
org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialized should create a Materialized with Serdes PASSED
org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize with a store name should create a Materialized with Serdes and a store name STARTED
org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize with a store name should create a Materialized with Serdes and a store name PASSED
org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize with a window store supplier should create a Materialized with Serdes and a store supplier STARTED
org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize with a window store supplier should create a Materialized with Serdes and a store supplier PASSED
org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize with a key value store supplier should create a Materialized with Serdes and a store supplier STARTED
org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize with a key value store supplier should create a Materialized with Serdes and a store supplier PASSED
org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize with a session store supplier should create a Materialized with Serdes and a store supplier STARTED
org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize with a session store supplier should create a Materialized with Serdes and a store supplier PASSED
org.apache.kafka.streams.scala.WordCountTest > testShouldCountWordsMaterialized STARTED
org.apache.kafka.streams.scala.WordCountTest > testShouldCountWordsMaterialized PASSED
org.apache.kafka.streams.scala.WordCountTest > testShouldCountWordsJava STARTED
org.apache.kafka.streams.scala.WordCountTest > testShouldCountWordsJava PASSED
org.apache.kafka.streams.scala.WordCountTest > testShouldCountWords STARTED
org.apache.kafka.streams.scala.WordCountTest > testShouldCountWords PASSED
org.apache.kafka.streams.scala.StreamToTableJoinScalaIntegrationTestImplicitSerdes > testShouldCountClicksPerRegionWithNamedRepartitionTopic STARTED
org.apache.kafka.streams.scala.StreamToTableJoinScalaIntegrationTestImplicitSerdes > testShouldCountClicksPerRegionWithNamedRepartitionTopic PASSED
org.apache.kafka.streams.scala.StreamToTableJoinScalaIntegrationTestImplicitSerdes > testShouldCountClicksPerRegionJava STARTED
org.apache.kafka.streams.scala.StreamToTableJoinScalaIntegrationTestImplicitSerdes > testShouldCountClicksPerRegionJava PASSED
org.apache.kafka.streams.scala.StreamToTableJoinScalaIntegrationTestImplicitSerdes > testShouldCountClicksPerRegion STARTED
org.apache.kafka.streams.scala.StreamToTableJoinScalaIntegrationTestImplicitSerdes > testShouldCountClicksPerRegion PASSED
org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaJoin STARTED
org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaJoin PASSED
org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaSimple STARTED
org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaSimple PASSED
org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaAggregate STARTED
org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaAggregate PASSED
org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaProperties STARTED
org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaProperties PASSED
org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaTransform STARTED
org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaTransform PASSED
FAILURE: Build failed with an exception.
* What went wrong:
Execution failed for task ':core:test'.
> There were failing tests. See the report at: file://<https://builds.apache.org/job/kafka-trunk-jdk8/ws/core/build/reports/tests/test/index.html>
* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
* Get more help at https://help.gradle.org
Deprecated Gradle features were used in this build, making it incompatible with Gradle 6.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/5.4.1/userguide/command_line_interface.html#sec:command_line_warnings
BUILD FAILED in 3h 26m 18s
182 actionable tasks: 155 executed, 27 up-to-date
Build step 'Execute shell' marked build as failure
[FINDBUGS] Collecting findbugs analysis files...
Setting GRADLE_4_10_3_HOME=/home/jenkins/tools/gradle/4.10.3
[FINDBUGS] Searching for all files in <https://builds.apache.org/job/kafka-trunk-jdk8/ws/> that match the pattern **/build/reports/*bugs/*.xml
[FINDBUGS] Parsing 17 files in <https://builds.apache.org/job/kafka-trunk-jdk8/ws/>
[FINDBUGS] Successfully parsed file <https://builds.apache.org/job/kafka-trunk-jdk8/ws/clients/build/reports/spotbugs/main.xml> with 0 unique warnings and 0 duplicates.
[FINDBUGS] Successfully parsed file <https://builds.apache.org/job/kafka-trunk-jdk8/ws/connect/api/build/reports/spotbugs/main.xml> with 0 unique warnings and 0 duplicates.
[FINDBUGS] Successfully parsed file <https://builds.apache.org/job/kafka-trunk-jdk8/ws/connect/basic-auth-extension/build/reports/spotbugs/main.xml> with 0 unique warnings and 0 duplicates.
[FINDBUGS] Successfully parsed file <https://builds.apache.org/job/kafka-trunk-jdk8/ws/connect/file/build/reports/spotbugs/main.xml> with 0 unique warnings and 0 duplicates.
[FINDBUGS] Successfully parsed file <https://builds.apache.org/job/kafka-trunk-jdk8/ws/connect/json/build/reports/spotbugs/main.xml> with 0 unique warnings and 0 duplicates.
[FINDBUGS] Successfully parsed file <https://builds.apache.org/job/kafka-trunk-jdk8/ws/connect/runtime/build/reports/spotbugs/main.xml> with 0 unique warnings and 0 duplicates.
[FINDBUGS] Successfully parsed file <https://builds.apache.org/job/kafka-trunk-jdk8/ws/connect/transforms/build/reports/spotbugs/main.xml> with 0 unique warnings and 0 duplicates.
[FINDBUGS] Successfully parsed file <https://builds.apache.org/job/kafka-trunk-jdk8/ws/core/build/reports/spotbugs/main.xml> with 0 unique warnings and 0 duplicates.
[FINDBUGS] Successfully parsed file <https://builds.apache.org/job/kafka-trunk-jdk8/ws/examples/build/reports/spotbugs/main.xml> with 0 unique warnings and 0 duplicates.
[FINDBUGS] Successfully parsed file <https://builds.apache.org/job/kafka-trunk-jdk8/ws/generator/build/reports/spotbugs/main.xml> with 0 unique warnings and 0 duplicates.
[FINDBUGS] Successfully parsed file <https://builds.apache.org/job/kafka-trunk-jdk8/ws/jmh-benchmarks/build/reports/spotbugs/main.xml> with 0 unique warnings and 0 duplicates.
[FINDBUGS] Successfully parsed file <https://builds.apache.org/job/kafka-trunk-jdk8/ws/log4j-appender/build/reports/spotbugs/main.xml> with 0 unique warnings and 0 duplicates.
[FINDBUGS] Successfully parsed file <https://builds.apache.org/job/kafka-trunk-jdk8/ws/streams/build/reports/spotbugs/main.xml> with 0 unique warnings and 0 duplicates.
[FINDBUGS] Successfully parsed file <https://builds.apache.org/job/kafka-trunk-jdk8/ws/streams/examples/build/reports/spotbugs/main.xml> with 0 unique warnings and 0 duplicates.
[FINDBUGS] Successfully parsed file <https://builds.apache.org/job/kafka-trunk-jdk8/ws/streams/streams-scala/build/reports/spotbugs/main.xml> with 0 unique warnings and 0 duplicates.
[FINDBUGS] Successfully parsed file <https://builds.apache.org/job/kafka-trunk-jdk8/ws/streams/test-utils/build/reports/spotbugs/main.xml> with 0 unique warnings and 0 duplicates.
[FINDBUGS] Successfully parsed file <https://builds.apache.org/job/kafka-trunk-jdk8/ws/tools/build/reports/spotbugs/main.xml> with 0 unique warnings and 0 duplicates.
Setting GRADLE_4_10_3_HOME=/home/jenkins/tools/gradle/4.10.3
No credentials specified
Setting GRADLE_4_10_3_HOME=/home/jenkins/tools/gradle/4.10.3
<Git Blamer> Using GitBlamer to create author and commit information for all warnings.
<Git Blamer> GIT_COMMIT=d7f8ec8628ee18326d22f6a58c787629a8faf307, workspace=<https://builds.apache.org/job/kafka-trunk-jdk8/ws/>
[FINDBUGS] Computing warning deltas based on reference build #3868
Recording test results
Setting GRADLE_4_10_3_HOME=/home/jenkins/tools/gradle/4.10.3
Setting GRADLE_4_10_3_HOME=/home/jenkins/tools/gradle/4.10.3
Not sending mail to unregistered user gwen@confluent.io
Jenkins build is back to normal : kafka-trunk-jdk8 #3872
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/kafka-trunk-jdk8/3872/display/redirect?page=changes>
Build failed in Jenkins: kafka-trunk-jdk8 #3871
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/kafka-trunk-jdk8/3871/display/redirect?page=changes>
Changes:
[wangguoz] KAFKA-8412: Fix nullpointer exception thrown on flushing before closing
[harsha] KAFKA-8669: Add security providers in kafka security config (#7090)
------------------------------------------
[...truncated 8.51 MB...]
org.apache.kafka.connect.storage.OffsetStorageWriterTest > testAlreadyFlushing PASSED
org.apache.kafka.connect.storage.OffsetStorageWriterTest > testCancelBeforeAwaitFlush STARTED
org.apache.kafka.connect.storage.OffsetStorageWriterTest > testCancelBeforeAwaitFlush PASSED
org.apache.kafka.connect.storage.OffsetStorageWriterTest > testCancelAfterAwaitFlush STARTED
org.apache.kafka.connect.storage.OffsetStorageWriterTest > testCancelAfterAwaitFlush PASSED
org.apache.kafka.connect.storage.OffsetStorageWriterTest > testWriteFlush STARTED
org.apache.kafka.connect.storage.OffsetStorageWriterTest > testWriteFlush PASSED
org.apache.kafka.connect.storage.KafkaConfigBackingStoreTest > testPutConnectorConfig STARTED
org.apache.kafka.connect.storage.KafkaConfigBackingStoreTest > testPutConnectorConfig PASSED
org.apache.kafka.connect.storage.KafkaConfigBackingStoreTest > testPutTaskConfigs STARTED
org.apache.kafka.connect.storage.KafkaConfigBackingStoreTest > testPutTaskConfigs PASSED
org.apache.kafka.connect.storage.KafkaConfigBackingStoreTest > testPutTaskConfigsZeroTasks STARTED
org.apache.kafka.connect.storage.KafkaConfigBackingStoreTest > testPutTaskConfigsZeroTasks PASSED
org.apache.kafka.connect.storage.KafkaConfigBackingStoreTest > testRestoreTargetState STARTED
org.apache.kafka.connect.storage.KafkaConfigBackingStoreTest > testRestoreTargetState PASSED
org.apache.kafka.connect.storage.KafkaConfigBackingStoreTest > testBackgroundUpdateTargetState STARTED
org.apache.kafka.connect.storage.KafkaConfigBackingStoreTest > testBackgroundUpdateTargetState PASSED
org.apache.kafka.connect.storage.KafkaConfigBackingStoreTest > testBackgroundConnectorDeletion STARTED
org.apache.kafka.connect.storage.KafkaConfigBackingStoreTest > testBackgroundConnectorDeletion PASSED
org.apache.kafka.connect.storage.KafkaConfigBackingStoreTest > testRestore STARTED
org.apache.kafka.connect.storage.KafkaConfigBackingStoreTest > testRestore PASSED
org.apache.kafka.connect.storage.KafkaConfigBackingStoreTest > testRestoreConnectorDeletion STARTED
org.apache.kafka.connect.storage.KafkaConfigBackingStoreTest > testRestoreConnectorDeletion PASSED
org.apache.kafka.connect.storage.KafkaConfigBackingStoreTest > testRestoreZeroTasks STARTED
org.apache.kafka.connect.storage.KafkaConfigBackingStoreTest > testRestoreZeroTasks PASSED
org.apache.kafka.connect.storage.KafkaConfigBackingStoreTest > testPutTaskConfigsDoesNotResolveAllInconsistencies STARTED
org.apache.kafka.connect.storage.KafkaConfigBackingStoreTest > testPutTaskConfigsDoesNotResolveAllInconsistencies PASSED
org.apache.kafka.connect.storage.KafkaConfigBackingStoreTest > testStartStop STARTED
org.apache.kafka.connect.storage.KafkaConfigBackingStoreTest > testStartStop PASSED
org.apache.kafka.connect.storage.KafkaConfigBackingStoreTest > testRestoreTargetStateUnexpectedDeletion STARTED
org.apache.kafka.connect.storage.KafkaConfigBackingStoreTest > testRestoreTargetStateUnexpectedDeletion PASSED
org.apache.kafka.connect.storage.FileOffsetBackingStoreTest > testSaveRestore STARTED
org.apache.kafka.connect.storage.FileOffsetBackingStoreTest > testSaveRestore PASSED
org.apache.kafka.connect.storage.FileOffsetBackingStoreTest > testGetSet STARTED
org.apache.kafka.connect.storage.FileOffsetBackingStoreTest > testGetSet PASSED
org.apache.kafka.connect.integration.ErrorHandlingIntegrationTest > testSkipRetryAndDLQWithHeaders STARTED
org.apache.kafka.connect.integration.ErrorHandlingIntegrationTest > testSkipRetryAndDLQWithHeaders PASSED
org.apache.kafka.connect.integration.StartAndStopLatchTest > shouldReturnFalseWhenAwaitingForDependentLatchToComplete STARTED
org.apache.kafka.connect.integration.StartAndStopLatchTest > shouldReturnFalseWhenAwaitingForDependentLatchToComplete PASSED
org.apache.kafka.connect.integration.StartAndStopLatchTest > shouldReturnTrueWhenAwaitingForStartAndStopAndDependentLatch STARTED
org.apache.kafka.connect.integration.StartAndStopLatchTest > shouldReturnTrueWhenAwaitingForStartAndStopAndDependentLatch PASSED
org.apache.kafka.connect.integration.StartAndStopLatchTest > shouldReturnFalseWhenAwaitingForStartToNeverComplete STARTED
org.apache.kafka.connect.integration.StartAndStopLatchTest > shouldReturnFalseWhenAwaitingForStartToNeverComplete PASSED
org.apache.kafka.connect.integration.StartAndStopLatchTest > shouldReturnFalseWhenAwaitingForStopToNeverComplete STARTED
org.apache.kafka.connect.integration.StartAndStopLatchTest > shouldReturnFalseWhenAwaitingForStopToNeverComplete PASSED
org.apache.kafka.connect.integration.StartAndStopLatchTest > shouldReturnTrueWhenAwaitingForStartAndStopToComplete STARTED
org.apache.kafka.connect.integration.StartAndStopLatchTest > shouldReturnTrueWhenAwaitingForStartAndStopToComplete PASSED
org.apache.kafka.connect.integration.RebalanceSourceConnectorsIntegrationTest > testStartTwoConnectors STARTED
org.apache.kafka.connect.integration.RebalanceSourceConnectorsIntegrationTest > testStartTwoConnectors SKIPPED
org.apache.kafka.connect.integration.RebalanceSourceConnectorsIntegrationTest > testDeleteConnector STARTED
org.apache.kafka.connect.integration.RebalanceSourceConnectorsIntegrationTest > testDeleteConnector SKIPPED
org.apache.kafka.connect.integration.RebalanceSourceConnectorsIntegrationTest > testAddingWorker STARTED
org.apache.kafka.connect.integration.RebalanceSourceConnectorsIntegrationTest > testAddingWorker PASSED
org.apache.kafka.connect.integration.RebalanceSourceConnectorsIntegrationTest > testRemovingWorker STARTED
org.apache.kafka.connect.integration.RebalanceSourceConnectorsIntegrationTest > testRemovingWorker PASSED
org.apache.kafka.connect.integration.RebalanceSourceConnectorsIntegrationTest > testReconfigConnector STARTED
org.apache.kafka.connect.integration.RebalanceSourceConnectorsIntegrationTest > testReconfigConnector SKIPPED
org.apache.kafka.connect.integration.StartAndStopCounterTest > shouldFailToWaitForRestartThatNeverHappens STARTED
org.apache.kafka.connect.integration.StartAndStopCounterTest > shouldFailToWaitForRestartThatNeverHappens PASSED
org.apache.kafka.connect.integration.StartAndStopCounterTest > shouldRecordStarts STARTED
org.apache.kafka.connect.integration.StartAndStopCounterTest > shouldRecordStarts PASSED
org.apache.kafka.connect.integration.StartAndStopCounterTest > shouldExpectRestarts STARTED
org.apache.kafka.connect.integration.StartAndStopCounterTest > shouldExpectRestarts PASSED
org.apache.kafka.connect.integration.StartAndStopCounterTest > shouldRecordStops STARTED
org.apache.kafka.connect.integration.StartAndStopCounterTest > shouldRecordStops PASSED
org.apache.kafka.connect.integration.ExampleConnectIntegrationTest > testSourceConnector STARTED
org.apache.kafka.connect.integration.ExampleConnectIntegrationTest > testSourceConnector PASSED
org.apache.kafka.connect.integration.ExampleConnectIntegrationTest > testSinkConnector STARTED
org.apache.kafka.connect.integration.ExampleConnectIntegrationTest > testSinkConnector PASSED
org.apache.kafka.connect.integration.RestExtensionIntegrationTest > testImmediateRequestForListOfConnectors STARTED
org.apache.kafka.connect.integration.RestExtensionIntegrationTest > testImmediateRequestForListOfConnectors PASSED
org.apache.kafka.connect.integration.ConnectorCientPolicyIntegrationTest > testCreateWithNotAllowedOverridesForPrincipalPolicy STARTED
ERROR: Could not install GRADLE_4_10_3_HOME
java.lang.NullPointerException
at hudson.plugins.toolenv.ToolEnvBuildWrapper$1.buildEnvVars(ToolEnvBuildWrapper.java:46)
at hudson.model.AbstractBuild.getEnvironment(AbstractBuild.java:881)
at hudson.plugins.git.GitSCM.getParamExpandedRepos(GitSCM.java:484)
at hudson.plugins.git.GitSCM.compareRemoteRevisionWithImpl(GitSCM.java:693)
at hudson.plugins.git.GitSCM.compareRemoteRevisionWith(GitSCM.java:658)
at hudson.scm.SCM.compareRemoteRevisionWith(SCM.java:400)
at hudson.scm.SCM.poll(SCM.java:417)
at hudson.model.AbstractProject._poll(AbstractProject.java:1390)
at hudson.model.AbstractProject.poll(AbstractProject.java:1293)
at hudson.triggers.SCMTrigger$Runner.runPolling(SCMTrigger.java:603)
at hudson.triggers.SCMTrigger$Runner.run(SCMTrigger.java:649)
at hudson.util.SequentialExecutionQueue$QueueEntry.run(SequentialExecutionQueue.java:119)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
org.apache.kafka.connect.integration.ConnectorCientPolicyIntegrationTest > testCreateWithNotAllowedOverridesForPrincipalPolicy PASSED
org.apache.kafka.connect.integration.ConnectorCientPolicyIntegrationTest > testCreateWithAllowedOverridesForAllPolicy STARTED
org.apache.kafka.connect.integration.ConnectorCientPolicyIntegrationTest > testCreateWithAllowedOverridesForAllPolicy PASSED
org.apache.kafka.connect.integration.ConnectorCientPolicyIntegrationTest > testCreateWithAllowedOverridesForPrincipalPolicy STARTED
org.apache.kafka.connect.integration.ConnectorCientPolicyIntegrationTest > testCreateWithAllowedOverridesForPrincipalPolicy PASSED
org.apache.kafka.connect.integration.ConnectorCientPolicyIntegrationTest > testCreateWithOverridesForNonePolicy STARTED
Build timed out (after 600 minutes). Marking the build as failed.
Build was aborted
[FINDBUGS] Collecting findbugs analysis files...
Setting GRADLE_4_10_3_HOME=/home/jenkins/tools/gradle/4.10.3
org.apache.kafka.connect.integration.ConnectorCientPolicyIntegrationTest > testCreateWithOverridesForNonePolicy SKIPPED
> Task :connect:runtime:test FAILED
FAILURE: Build failed with an exception.
* What went wrong:
Execution failed for task ':connect:runtime:test'.
> Process 'Gradle Test Executor 19' finished with non-zero exit value 143
This problem might be caused by incorrect test process configuration.
Please refer to the test execution section in the User Manual at https://docs.gradle.org/5.4.1/userguide/java_testing.html#sec:test_execution
* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
* Get more help at https://help.gradle.org
Deprecated Gradle features were used in this build, making it incompatible with Gradle 6.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/5.4.1/userguide/command_line_interface.html#sec:command_line_warnings
BUILD FAILED in 3h 48m 15s
182 actionable tasks: 179 executed, 3 up-to-date
FAILURE: Build failed with an exception.
* What went wrong:
Process 'Gradle Worker Daemon 2' finished with non-zero exit value 143
* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
* Get more help at https://help.gradle.org
Deprecated Gradle features were used in this build, making it incompatible with Gradle 6.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/5.4.1/userguide/command_line_interface.html#sec:command_line_warnings
BUILD FAILED in 3h 48m 16s
[FINDBUGS] Searching for all files in <https://builds.apache.org/job/kafka-trunk-jdk8/ws/> that match the pattern **/build/reports/*bugs/*.xml
[FINDBUGS] Parsing 17 files in <https://builds.apache.org/job/kafka-trunk-jdk8/ws/>
[FINDBUGS] Successfully parsed file <https://builds.apache.org/job/kafka-trunk-jdk8/ws/clients/build/reports/spotbugs/main.xml> with 0 unique warnings and 0 duplicates.
[FINDBUGS] Successfully parsed file <https://builds.apache.org/job/kafka-trunk-jdk8/ws/connect/api/build/reports/spotbugs/main.xml> with 0 unique warnings and 0 duplicates.
[FINDBUGS] Successfully parsed file <https://builds.apache.org/job/kafka-trunk-jdk8/ws/connect/basic-auth-extension/build/reports/spotbugs/main.xml> with 0 unique warnings and 0 duplicates.
[FINDBUGS] Successfully parsed file <https://builds.apache.org/job/kafka-trunk-jdk8/ws/connect/file/build/reports/spotbugs/main.xml> with 0 unique warnings and 0 duplicates.
[FINDBUGS] Successfully parsed file <https://builds.apache.org/job/kafka-trunk-jdk8/ws/connect/json/build/reports/spotbugs/main.xml> with 0 unique warnings and 0 duplicates.
[FINDBUGS] Successfully parsed file <https://builds.apache.org/job/kafka-trunk-jdk8/ws/connect/runtime/build/reports/spotbugs/main.xml> with 0 unique warnings and 0 duplicates.
[FINDBUGS] Successfully parsed file <https://builds.apache.org/job/kafka-trunk-jdk8/ws/connect/transforms/build/reports/spotbugs/main.xml> with 0 unique warnings and 0 duplicates.
[FINDBUGS] Successfully parsed file <https://builds.apache.org/job/kafka-trunk-jdk8/ws/core/build/reports/spotbugs/main.xml> with 0 unique warnings and 0 duplicates.
[FINDBUGS] Successfully parsed file <https://builds.apache.org/job/kafka-trunk-jdk8/ws/examples/build/reports/spotbugs/main.xml> with 0 unique warnings and 0 duplicates.
[FINDBUGS] Successfully parsed file <https://builds.apache.org/job/kafka-trunk-jdk8/ws/generator/build/reports/spotbugs/main.xml> with 0 unique warnings and 0 duplicates.
[FINDBUGS] Successfully parsed file <https://builds.apache.org/job/kafka-trunk-jdk8/ws/jmh-benchmarks/build/reports/spotbugs/main.xml> with 0 unique warnings and 0 duplicates.
[FINDBUGS] Successfully parsed file <https://builds.apache.org/job/kafka-trunk-jdk8/ws/log4j-appender/build/reports/spotbugs/main.xml> with 0 unique warnings and 0 duplicates.
[FINDBUGS] Successfully parsed file <https://builds.apache.org/job/kafka-trunk-jdk8/ws/streams/build/reports/spotbugs/main.xml> with 0 unique warnings and 0 duplicates.
[FINDBUGS] Successfully parsed file <https://builds.apache.org/job/kafka-trunk-jdk8/ws/streams/examples/build/reports/spotbugs/main.xml> with 0 unique warnings and 0 duplicates.
[FINDBUGS] Successfully parsed file <https://builds.apache.org/job/kafka-trunk-jdk8/ws/streams/streams-scala/build/reports/spotbugs/main.xml> with 0 unique warnings and 0 duplicates.
[FINDBUGS] Successfully parsed file <https://builds.apache.org/job/kafka-trunk-jdk8/ws/streams/test-utils/build/reports/spotbugs/main.xml> with 0 unique warnings and 0 duplicates.
[FINDBUGS] Successfully parsed file <https://builds.apache.org/job/kafka-trunk-jdk8/ws/tools/build/reports/spotbugs/main.xml> with 0 unique warnings and 0 duplicates.
Setting GRADLE_4_10_3_HOME=/home/jenkins/tools/gradle/4.10.3
No credentials specified
Setting GRADLE_4_10_3_HOME=/home/jenkins/tools/gradle/4.10.3
<Git Blamer> Using GitBlamer to create author and commit information for all warnings.
<Git Blamer> GIT_COMMIT=d08bcae7f99a863b6d2058a1568384f904e539ed, workspace=<https://builds.apache.org/job/kafka-trunk-jdk8/ws/>
[FINDBUGS] Computing warning deltas based on reference build #3868
Recording test results
Setting GRADLE_4_10_3_HOME=/home/jenkins/tools/gradle/4.10.3
Setting GRADLE_4_10_3_HOME=/home/jenkins/tools/gradle/4.10.3
Not sending mail to unregistered user gwen@confluent.io
Not sending mail to unregistered user wangguoz@gmail.com