You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@kafka.apache.org by Apache Jenkins Server <je...@builds.apache.org> on 2019/08/13 02:59:24 UTC

Build failed in Jenkins: kafka-trunk-jdk11 #748

See <https://builds.apache.org/job/kafka-trunk-jdk11/748/display/redirect?page=changes>

Changes:

[jason] MINOR: Format AdminUtils::assignReplicasToBrokers java documentation

------------------------------------------
[...truncated 6.52 MB...]

org.apache.kafka.streams.scala.kstream.JoinedTest > Create a Joined should create a Joined with Serdes and repartition topic name STARTED

org.apache.kafka.streams.scala.kstream.JoinedTest > Create a Joined should create a Joined with Serdes and repartition topic name PASSED

org.apache.kafka.streams.scala.kstream.GroupedTest > Create a Grouped should create a Grouped with Serdes STARTED

org.apache.kafka.streams.scala.kstream.GroupedTest > Create a Grouped should create a Grouped with Serdes PASSED

org.apache.kafka.streams.scala.kstream.GroupedTest > Create a Grouped with repartition topic name should create a Grouped with Serdes, and repartition topic name STARTED

org.apache.kafka.streams.scala.kstream.GroupedTest > Create a Grouped with repartition topic name should create a Grouped with Serdes, and repartition topic name PASSED

org.apache.kafka.streams.scala.kstream.SuppressedTest > Suppressed.untilWindowCloses should produce the correct suppression STARTED

org.apache.kafka.streams.scala.kstream.SuppressedTest > Suppressed.untilWindowCloses should produce the correct suppression PASSED

org.apache.kafka.streams.scala.kstream.SuppressedTest > Suppressed.untilTimeLimit should produce the correct suppression STARTED

org.apache.kafka.streams.scala.kstream.SuppressedTest > Suppressed.untilTimeLimit should produce the correct suppression PASSED

org.apache.kafka.streams.scala.kstream.SuppressedTest > BufferConfig.maxRecords should produce the correct buffer config STARTED

org.apache.kafka.streams.scala.kstream.SuppressedTest > BufferConfig.maxRecords should produce the correct buffer config PASSED

org.apache.kafka.streams.scala.kstream.SuppressedTest > BufferConfig.maxBytes should produce the correct buffer config STARTED

org.apache.kafka.streams.scala.kstream.SuppressedTest > BufferConfig.maxBytes should produce the correct buffer config PASSED

org.apache.kafka.streams.scala.kstream.SuppressedTest > BufferConfig.unbounded should produce the correct buffer config STARTED

org.apache.kafka.streams.scala.kstream.SuppressedTest > BufferConfig.unbounded should produce the correct buffer config PASSED

org.apache.kafka.streams.scala.kstream.SuppressedTest > BufferConfig should support very long chains of factory methods STARTED

org.apache.kafka.streams.scala.kstream.SuppressedTest > BufferConfig should support very long chains of factory methods PASSED

org.apache.kafka.streams.scala.kstream.ProducedTest > Create a Produced should create a Produced with Serdes STARTED

org.apache.kafka.streams.scala.kstream.ProducedTest > Create a Produced should create a Produced with Serdes PASSED

org.apache.kafka.streams.scala.kstream.ProducedTest > Create a Produced with timestampExtractor and resetPolicy should create a Consumed with Serdes, timestampExtractor and resetPolicy STARTED

org.apache.kafka.streams.scala.kstream.ProducedTest > Create a Produced with timestampExtractor and resetPolicy should create a Consumed with Serdes, timestampExtractor and resetPolicy PASSED

org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed should create a Consumed with Serdes STARTED

org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed should create a Consumed with Serdes PASSED

org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed with timestampExtractor and resetPolicy should create a Consumed with Serdes, timestampExtractor and resetPolicy STARTED

org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed with timestampExtractor and resetPolicy should create a Consumed with Serdes, timestampExtractor and resetPolicy PASSED

org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed with timestampExtractor should create a Consumed with Serdes and timestampExtractor STARTED

org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed with timestampExtractor should create a Consumed with Serdes and timestampExtractor PASSED

org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed with resetPolicy should create a Consumed with Serdes and resetPolicy STARTED

org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed with resetPolicy should create a Consumed with Serdes and resetPolicy PASSED

org.apache.kafka.streams.scala.kstream.KStreamTest > filter a KStream should filter records satisfying the predicate STARTED

org.apache.kafka.streams.scala.kstream.KStreamTest > filter a KStream should filter records satisfying the predicate PASSED

org.apache.kafka.streams.scala.kstream.KStreamTest > filterNot a KStream should filter records not satisfying the predicate STARTED

org.apache.kafka.streams.scala.kstream.KStreamTest > filterNot a KStream should filter records not satisfying the predicate PASSED

org.apache.kafka.streams.scala.kstream.KStreamTest > foreach a KStream should run foreach actions on records STARTED

org.apache.kafka.streams.scala.kstream.KStreamTest > foreach a KStream should run foreach actions on records PASSED

org.apache.kafka.streams.scala.kstream.KStreamTest > peek a KStream should run peek actions on records STARTED

org.apache.kafka.streams.scala.kstream.KStreamTest > peek a KStream should run peek actions on records PASSED

org.apache.kafka.streams.scala.kstream.KStreamTest > selectKey a KStream should select a new key STARTED

org.apache.kafka.streams.scala.kstream.KStreamTest > selectKey a KStream should select a new key PASSED

org.apache.kafka.streams.scala.kstream.KStreamTest > join 2 KStreams should join correctly records STARTED

org.apache.kafka.streams.scala.kstream.KStreamTest > join 2 KStreams should join correctly records PASSED

org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialized should create a Materialized with Serdes STARTED

org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialized should create a Materialized with Serdes PASSED

org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize with a store name should create a Materialized with Serdes and a store name STARTED

org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize with a store name should create a Materialized with Serdes and a store name PASSED

org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize with a window store supplier should create a Materialized with Serdes and a store supplier STARTED

org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize with a window store supplier should create a Materialized with Serdes and a store supplier PASSED

org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize with a key value store supplier should create a Materialized with Serdes and a store supplier STARTED

org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize with a key value store supplier should create a Materialized with Serdes and a store supplier PASSED

org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize with a session store supplier should create a Materialized with Serdes and a store supplier STARTED

org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize with a session store supplier should create a Materialized with Serdes and a store supplier PASSED

org.apache.kafka.streams.scala.kstream.KTableTest > filter a KTable should filter records satisfying the predicate STARTED

org.apache.kafka.streams.scala.kstream.KTableTest > filter a KTable should filter records satisfying the predicate PASSED

org.apache.kafka.streams.scala.kstream.KTableTest > filterNot a KTable should filter records not satisfying the predicate STARTED

org.apache.kafka.streams.scala.kstream.KTableTest > filterNot a KTable should filter records not satisfying the predicate PASSED

org.apache.kafka.streams.scala.kstream.KTableTest > join 2 KTables should join correctly records STARTED

org.apache.kafka.streams.scala.kstream.KTableTest > join 2 KTables should join correctly records PASSED

org.apache.kafka.streams.scala.kstream.KTableTest > join 2 KTables with a Materialized should join correctly records and state store STARTED

org.apache.kafka.streams.scala.kstream.KTableTest > join 2 KTables with a Materialized should join correctly records and state store PASSED

org.apache.kafka.streams.scala.kstream.KTableTest > windowed KTable#suppress should correctly suppress results using Suppressed.untilTimeLimit STARTED

org.apache.kafka.streams.scala.kstream.KTableTest > windowed KTable#suppress should correctly suppress results using Suppressed.untilTimeLimit PASSED

org.apache.kafka.streams.scala.kstream.KTableTest > windowed KTable#suppress should correctly suppress results using Suppressed.untilWindowCloses STARTED

org.apache.kafka.streams.scala.kstream.KTableTest > windowed KTable#suppress should correctly suppress results using Suppressed.untilWindowCloses PASSED

org.apache.kafka.streams.scala.kstream.KTableTest > session windowed KTable#suppress should correctly suppress results using Suppressed.untilWindowCloses STARTED

org.apache.kafka.streams.scala.kstream.KTableTest > session windowed KTable#suppress should correctly suppress results using Suppressed.untilWindowCloses PASSED

org.apache.kafka.streams.scala.kstream.KTableTest > non-windowed KTable#suppress should correctly suppress results using Suppressed.untilTimeLimit STARTED

org.apache.kafka.streams.scala.kstream.KTableTest > non-windowed KTable#suppress should correctly suppress results using Suppressed.untilTimeLimit PASSED

org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaJoin STARTED

org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaJoin PASSED

org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaSimple STARTED

org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaSimple PASSED

org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaAggregate STARTED

org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaAggregate PASSED

org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaProperties STARTED

org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaProperties PASSED

org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaTransform STARTED

org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaTransform PASSED

org.apache.kafka.streams.scala.StreamToTableJoinScalaIntegrationTestImplicitSerdes > testShouldCountClicksPerRegionWithNamedRepartitionTopic STARTED

org.apache.kafka.streams.scala.StreamToTableJoinScalaIntegrationTestImplicitSerdes > testShouldCountClicksPerRegionWithNamedRepartitionTopic PASSED

org.apache.kafka.streams.scala.StreamToTableJoinScalaIntegrationTestImplicitSerdes > testShouldCountClicksPerRegionJava STARTED

org.apache.kafka.streams.scala.StreamToTableJoinScalaIntegrationTestImplicitSerdes > testShouldCountClicksPerRegionJava PASSED

org.apache.kafka.streams.scala.StreamToTableJoinScalaIntegrationTestImplicitSerdes > testShouldCountClicksPerRegion STARTED

org.apache.kafka.streams.scala.StreamToTableJoinScalaIntegrationTestImplicitSerdes > testShouldCountClicksPerRegion PASSED

org.apache.kafka.streams.scala.WordCountTest > testShouldCountWordsMaterialized STARTED

org.apache.kafka.streams.scala.WordCountTest > testShouldCountWordsMaterialized PASSED

org.apache.kafka.streams.scala.WordCountTest > testShouldCountWordsJava STARTED

org.apache.kafka.streams.scala.WordCountTest > testShouldCountWordsJava PASSED

org.apache.kafka.streams.scala.WordCountTest > testShouldCountWords STARTED

org.apache.kafka.streams.scala.WordCountTest > testShouldCountWords PASSED

> Task :log4j-appender:test

org.apache.kafka.log4jappender.KafkaLog4jAppenderTest > testLog4jAppendsWithSyncSendWithoutIgnoringExceptionsShouldNotThrowException STARTED

org.apache.kafka.log4jappender.KafkaLog4jAppenderTest > testLog4jAppendsWithSyncSendWithoutIgnoringExceptionsShouldNotThrowException PASSED

org.apache.kafka.log4jappender.KafkaLog4jAppenderTest > testLog4jAppendsWithRealProducerConfigWithSyncSendAndNotIgnoringExceptionsShouldThrowException STARTED

org.apache.kafka.log4jappender.KafkaLog4jAppenderTest > testLog4jAppendsWithRealProducerConfigWithSyncSendAndNotIgnoringExceptionsShouldThrowException PASSED

org.apache.kafka.log4jappender.KafkaLog4jAppenderTest > testLog4jAppends STARTED

org.apache.kafka.log4jappender.KafkaLog4jAppenderTest > testLog4jAppends PASSED

org.apache.kafka.log4jappender.KafkaLog4jAppenderTest > testSetSaslMechanism STARTED

org.apache.kafka.log4jappender.KafkaLog4jAppenderTest > testSetSaslMechanism PASSED

org.apache.kafka.log4jappender.KafkaLog4jAppenderTest > testJaasConfigNotSet STARTED

org.apache.kafka.log4jappender.KafkaLog4jAppenderTest > testJaasConfigNotSet PASSED

org.apache.kafka.log4jappender.KafkaLog4jAppenderTest > testKafkaLog4jConfigs STARTED

org.apache.kafka.log4jappender.KafkaLog4jAppenderTest > testKafkaLog4jConfigs PASSED

org.apache.kafka.log4jappender.KafkaLog4jAppenderTest > testLog4jAppendsWithSyncSendAndSimulateProducerFailShouldThrowException STARTED

org.apache.kafka.log4jappender.KafkaLog4jAppenderTest > testLog4jAppendsWithSyncSendAndSimulateProducerFailShouldThrowException PASSED

org.apache.kafka.log4jappender.KafkaLog4jAppenderTest > testLog4jAppendsWithRealProducerConfigWithSyncSendShouldNotThrowException STARTED

org.apache.kafka.log4jappender.KafkaLog4jAppenderTest > testLog4jAppendsWithRealProducerConfigWithSyncSendShouldNotThrowException PASSED

org.apache.kafka.log4jappender.KafkaLog4jAppenderTest > testSaslMechanismNotSet STARTED

org.apache.kafka.log4jappender.KafkaLog4jAppenderTest > testSaslMechanismNotSet PASSED

org.apache.kafka.log4jappender.KafkaLog4jAppenderTest > testSetJaasConfig STARTED

org.apache.kafka.log4jappender.KafkaLog4jAppenderTest > testSetJaasConfig PASSED

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':core:test'.
> There were failing tests. See the report at: file://<https://builds.apache.org/job/kafka-trunk-jdk11/ws/core/build/reports/tests/test/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 6.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/5.4.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 3h 10m 13s
178 actionable tasks: 175 executed, 3 up-to-date
Build step 'Execute shell' marked build as failure
Recording test results
ERROR: No tool found matching GRADLE_4_10_2_HOME
Setting GRADLE_4_10_3_HOME=/home/jenkins/tools/gradle/4.10.3
ERROR: No tool found matching GRADLE_4_10_2_HOME
Setting GRADLE_4_10_3_HOME=/home/jenkins/tools/gradle/4.10.3
Not sending mail to unregistered user ismael@juma.me.uk
Not sending mail to unregistered user vahid.hashemian@gmail.com
Not sending mail to unregistered user noreply@github.com

Jenkins build is back to normal : kafka-trunk-jdk11 #754

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/kafka-trunk-jdk11/754/display/redirect?page=changes>


Build failed in Jenkins: kafka-trunk-jdk11 #753

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/kafka-trunk-jdk11/753/display/redirect?page=changes>

Changes:

[bbejeck] MINOR: Correct typo in test name `TimetampedSegmentsTest` (#7210)

------------------------------------------
[...truncated 2.59 MB...]

org.apache.kafka.connect.runtime.standalone.StandaloneHerderTest > testPutTaskConfigs STARTED

org.apache.kafka.connect.runtime.standalone.StandaloneHerderTest > testPutTaskConfigs PASSED

org.apache.kafka.connect.runtime.standalone.StandaloneHerderTest > testCreateConnectorFailedBasicValidation STARTED

org.apache.kafka.connect.runtime.standalone.StandaloneHerderTest > testCreateConnectorFailedBasicValidation PASSED

org.apache.kafka.connect.runtime.standalone.StandaloneHerderTest > testCreateConnectorAlreadyExists STARTED

org.apache.kafka.connect.runtime.standalone.StandaloneHerderTest > testCreateConnectorAlreadyExists PASSED

org.apache.kafka.connect.connector.policy.PrincipalConnectorClientConfigOverridePolicyTest > testPrincipalPlusOtherConfigs STARTED

org.apache.kafka.connect.connector.policy.PrincipalConnectorClientConfigOverridePolicyTest > testPrincipalPlusOtherConfigs PASSED

org.apache.kafka.connect.connector.policy.PrincipalConnectorClientConfigOverridePolicyTest > testPrincipalOnly STARTED

org.apache.kafka.connect.connector.policy.PrincipalConnectorClientConfigOverridePolicyTest > testPrincipalOnly PASSED

org.apache.kafka.connect.connector.policy.NoneConnectorClientConfigOverridePolicyTest > testNoOverrides STARTED

org.apache.kafka.connect.connector.policy.NoneConnectorClientConfigOverridePolicyTest > testNoOverrides PASSED

org.apache.kafka.connect.connector.policy.NoneConnectorClientConfigOverridePolicyTest > testWithOverrides STARTED

org.apache.kafka.connect.connector.policy.NoneConnectorClientConfigOverridePolicyTest > testWithOverrides PASSED

> Task :streams:examples:test

org.apache.kafka.streams.examples.wordcount.WordCountProcessorTest > test STARTED

org.apache.kafka.streams.examples.wordcount.WordCountProcessorTest > test PASSED

> Task :streams:streams-scala:test

org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed should create a Consumed with Serdes STARTED

org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed should create a Consumed with Serdes PASSED

org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed with timestampExtractor and resetPolicy should create a Consumed with Serdes, timestampExtractor and resetPolicy STARTED

org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed with timestampExtractor and resetPolicy should create a Consumed with Serdes, timestampExtractor and resetPolicy PASSED

org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed with timestampExtractor should create a Consumed with Serdes and timestampExtractor STARTED

org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed with timestampExtractor should create a Consumed with Serdes and timestampExtractor PASSED

org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed with resetPolicy should create a Consumed with Serdes and resetPolicy STARTED

org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed with resetPolicy should create a Consumed with Serdes and resetPolicy PASSED

org.apache.kafka.streams.scala.kstream.KTableTest > filter a KTable should filter records satisfying the predicate STARTED

org.apache.kafka.streams.scala.kstream.KTableTest > filter a KTable should filter records satisfying the predicate PASSED

org.apache.kafka.streams.scala.kstream.KTableTest > filterNot a KTable should filter records not satisfying the predicate STARTED

org.apache.kafka.streams.scala.kstream.KTableTest > filterNot a KTable should filter records not satisfying the predicate PASSED

org.apache.kafka.streams.scala.kstream.KTableTest > join 2 KTables should join correctly records STARTED

org.apache.kafka.streams.scala.kstream.KTableTest > join 2 KTables should join correctly records PASSED

org.apache.kafka.streams.scala.kstream.KTableTest > join 2 KTables with a Materialized should join correctly records and state store STARTED

org.apache.kafka.streams.scala.kstream.KTableTest > join 2 KTables with a Materialized should join correctly records and state store PASSED

org.apache.kafka.streams.scala.kstream.KTableTest > windowed KTable#suppress should correctly suppress results using Suppressed.untilTimeLimit STARTED

org.apache.kafka.streams.scala.kstream.KTableTest > windowed KTable#suppress should correctly suppress results using Suppressed.untilTimeLimit PASSED

org.apache.kafka.streams.scala.kstream.KTableTest > windowed KTable#suppress should correctly suppress results using Suppressed.untilWindowCloses STARTED

org.apache.kafka.streams.scala.kstream.KTableTest > windowed KTable#suppress should correctly suppress results using Suppressed.untilWindowCloses PASSED

org.apache.kafka.streams.scala.kstream.KTableTest > session windowed KTable#suppress should correctly suppress results using Suppressed.untilWindowCloses STARTED

org.apache.kafka.streams.scala.kstream.KTableTest > session windowed KTable#suppress should correctly suppress results using Suppressed.untilWindowCloses PASSED

org.apache.kafka.streams.scala.kstream.KTableTest > non-windowed KTable#suppress should correctly suppress results using Suppressed.untilTimeLimit STARTED

org.apache.kafka.streams.scala.kstream.KTableTest > non-windowed KTable#suppress should correctly suppress results using Suppressed.untilTimeLimit PASSED

org.apache.kafka.streams.scala.kstream.ProducedTest > Create a Produced should create a Produced with Serdes STARTED

org.apache.kafka.streams.scala.kstream.ProducedTest > Create a Produced should create a Produced with Serdes PASSED

org.apache.kafka.streams.scala.kstream.ProducedTest > Create a Produced with timestampExtractor and resetPolicy should create a Consumed with Serdes, timestampExtractor and resetPolicy STARTED

org.apache.kafka.streams.scala.kstream.ProducedTest > Create a Produced with timestampExtractor and resetPolicy should create a Consumed with Serdes, timestampExtractor and resetPolicy PASSED

org.apache.kafka.streams.scala.kstream.SuppressedTest > Suppressed.untilWindowCloses should produce the correct suppression STARTED

org.apache.kafka.streams.scala.kstream.SuppressedTest > Suppressed.untilWindowCloses should produce the correct suppression PASSED

org.apache.kafka.streams.scala.kstream.SuppressedTest > Suppressed.untilTimeLimit should produce the correct suppression STARTED

org.apache.kafka.streams.scala.kstream.SuppressedTest > Suppressed.untilTimeLimit should produce the correct suppression PASSED

org.apache.kafka.streams.scala.kstream.SuppressedTest > BufferConfig.maxRecords should produce the correct buffer config STARTED

org.apache.kafka.streams.scala.kstream.SuppressedTest > BufferConfig.maxRecords should produce the correct buffer config PASSED

org.apache.kafka.streams.scala.kstream.SuppressedTest > BufferConfig.maxBytes should produce the correct buffer config STARTED

org.apache.kafka.streams.scala.kstream.SuppressedTest > BufferConfig.maxBytes should produce the correct buffer config PASSED

org.apache.kafka.streams.scala.kstream.SuppressedTest > BufferConfig.unbounded should produce the correct buffer config STARTED

org.apache.kafka.streams.scala.kstream.SuppressedTest > BufferConfig.unbounded should produce the correct buffer config PASSED

org.apache.kafka.streams.scala.kstream.SuppressedTest > BufferConfig should support very long chains of factory methods STARTED

org.apache.kafka.streams.scala.kstream.SuppressedTest > BufferConfig should support very long chains of factory methods PASSED

org.apache.kafka.streams.scala.kstream.GroupedTest > Create a Grouped should create a Grouped with Serdes STARTED

org.apache.kafka.streams.scala.kstream.GroupedTest > Create a Grouped should create a Grouped with Serdes PASSED

org.apache.kafka.streams.scala.kstream.GroupedTest > Create a Grouped with repartition topic name should create a Grouped with Serdes, and repartition topic name STARTED

org.apache.kafka.streams.scala.kstream.GroupedTest > Create a Grouped with repartition topic name should create a Grouped with Serdes, and repartition topic name PASSED

org.apache.kafka.streams.scala.kstream.JoinedTest > Create a Joined should create a Joined with Serdes STARTED

org.apache.kafka.streams.scala.kstream.JoinedTest > Create a Joined should create a Joined with Serdes PASSED

org.apache.kafka.streams.scala.kstream.JoinedTest > Create a Joined should create a Joined with Serdes and repartition topic name STARTED

org.apache.kafka.streams.scala.kstream.JoinedTest > Create a Joined should create a Joined with Serdes and repartition topic name PASSED

org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialized should create a Materialized with Serdes STARTED

org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialized should create a Materialized with Serdes PASSED

org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize with a store name should create a Materialized with Serdes and a store name STARTED

org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize with a store name should create a Materialized with Serdes and a store name PASSED

org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize with a window store supplier should create a Materialized with Serdes and a store supplier STARTED

org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize with a window store supplier should create a Materialized with Serdes and a store supplier PASSED

org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize with a key value store supplier should create a Materialized with Serdes and a store supplier STARTED

org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize with a key value store supplier should create a Materialized with Serdes and a store supplier PASSED

org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize with a session store supplier should create a Materialized with Serdes and a store supplier STARTED

org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize with a session store supplier should create a Materialized with Serdes and a store supplier PASSED

org.apache.kafka.streams.scala.kstream.KStreamTest > filter a KStream should filter records satisfying the predicate STARTED

org.apache.kafka.streams.scala.kstream.KStreamTest > filter a KStream should filter records satisfying the predicate PASSED

org.apache.kafka.streams.scala.kstream.KStreamTest > filterNot a KStream should filter records not satisfying the predicate STARTED

org.apache.kafka.streams.scala.kstream.KStreamTest > filterNot a KStream should filter records not satisfying the predicate PASSED

org.apache.kafka.streams.scala.kstream.KStreamTest > foreach a KStream should run foreach actions on records STARTED

org.apache.kafka.streams.scala.kstream.KStreamTest > foreach a KStream should run foreach actions on records PASSED

org.apache.kafka.streams.scala.kstream.KStreamTest > peek a KStream should run peek actions on records STARTED

org.apache.kafka.streams.scala.kstream.KStreamTest > peek a KStream should run peek actions on records PASSED

org.apache.kafka.streams.scala.kstream.KStreamTest > selectKey a KStream should select a new key STARTED

org.apache.kafka.streams.scala.kstream.KStreamTest > selectKey a KStream should select a new key PASSED

org.apache.kafka.streams.scala.kstream.KStreamTest > join 2 KStreams should join correctly records STARTED

org.apache.kafka.streams.scala.kstream.KStreamTest > join 2 KStreams should join correctly records PASSED

org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaJoin STARTED

org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaJoin PASSED

org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaSimple STARTED

org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaSimple PASSED

org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaAggregate STARTED

org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaAggregate PASSED

org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaProperties STARTED

org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaProperties PASSED

org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaTransform STARTED

org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaTransform PASSED

org.apache.kafka.streams.scala.StreamToTableJoinScalaIntegrationTestImplicitSerdes > testShouldCountClicksPerRegionWithNamedRepartitionTopic STARTED

org.apache.kafka.streams.scala.StreamToTableJoinScalaIntegrationTestImplicitSerdes > testShouldCountClicksPerRegionWithNamedRepartitionTopic PASSED

org.apache.kafka.streams.scala.StreamToTableJoinScalaIntegrationTestImplicitSerdes > testShouldCountClicksPerRegionJava STARTED

org.apache.kafka.streams.scala.StreamToTableJoinScalaIntegrationTestImplicitSerdes > testShouldCountClicksPerRegionJava PASSED

org.apache.kafka.streams.scala.StreamToTableJoinScalaIntegrationTestImplicitSerdes > testShouldCountClicksPerRegion STARTED

org.apache.kafka.streams.scala.StreamToTableJoinScalaIntegrationTestImplicitSerdes > testShouldCountClicksPerRegion PASSED

org.apache.kafka.streams.scala.WordCountTest > testShouldCountWordsMaterialized STARTED

org.apache.kafka.streams.scala.WordCountTest > testShouldCountWordsMaterialized PASSED

org.apache.kafka.streams.scala.WordCountTest > testShouldCountWordsJava STARTED

org.apache.kafka.streams.scala.WordCountTest > testShouldCountWordsJava PASSED

org.apache.kafka.streams.scala.WordCountTest > testShouldCountWords STARTED

org.apache.kafka.streams.scala.WordCountTest > testShouldCountWords PASSED

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':core:test'.
> There were failing tests. See the report at: file://<https://builds.apache.org/job/kafka-trunk-jdk11/ws/core/build/reports/tests/test/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 6.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/5.4.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 3h 49m 4s
178 actionable tasks: 152 executed, 26 up-to-date
Build step 'Execute shell' marked build as failure
Recording test results
ERROR: No tool found matching GRADLE_4_10_2_HOME
Setting GRADLE_4_10_3_HOME=/home/jenkins/tools/gradle/4.10.3
ERROR: No tool found matching GRADLE_4_10_2_HOME
Setting GRADLE_4_10_3_HOME=/home/jenkins/tools/gradle/4.10.3
Not sending mail to unregistered user ismael@juma.me.uk
Not sending mail to unregistered user vahid.hashemian@gmail.com
Not sending mail to unregistered user noreply@github.com
Not sending mail to unregistered user wangguoz@gmail.com
Not sending mail to unregistered user gwen@confluent.io

Build failed in Jenkins: kafka-trunk-jdk11 #752

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/kafka-trunk-jdk11/752/display/redirect?page=changes>

Changes:

[wangguoz] MINOR: remove unnecessary #remove overrides (#7178)

[github] KAFKA-8791: RocksDBTimestampedStore should open in regular mode by

[ismael] MINOR: Avoid unnecessary leaderFor calls when ProducerBatch queue empty

[github] KAFKA-8765: Remove interface annotations in Streams API (#7174)

[gwen] KAFKA-8792; Default ZK configuration to disable AdminServer

------------------------------------------
[...truncated 2.63 MB...]

org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperatorTest > testThrowExceptionInKafkaConsume STARTED

org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperatorTest > testThrowExceptionInKafkaConsume PASSED

org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperatorTest > testThrowExceptionInKafkaProduce STARTED

org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperatorTest > testThrowExceptionInKafkaProduce PASSED

org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperatorTest > testExecAndHandleRetriableErrorOnce STARTED

org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperatorTest > testExecAndHandleRetriableErrorOnce PASSED

org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperatorTest > testExecAndHandleRetriableErrorThrice STARTED

org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperatorTest > testExecAndHandleRetriableErrorThrice PASSED

org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperatorTest > testExecAndHandleNonRetriableErrorOnce STARTED

org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperatorTest > testExecAndHandleNonRetriableErrorOnce PASSED

org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperatorTest > testExecAndHandleNonRetriableErrorThrice STARTED

org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperatorTest > testExecAndHandleNonRetriableErrorThrice PASSED

org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperatorTest > testCheckRetryLimit STARTED

org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperatorTest > testCheckRetryLimit PASSED

org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperatorTest > testBackoffLimit STARTED

org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperatorTest > testBackoffLimit PASSED

org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperatorTest > testToleranceLimit STARTED

org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperatorTest > testToleranceLimit PASSED

org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperatorTest > testDefaultConfigs STARTED

org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperatorTest > testDefaultConfigs PASSED

org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperatorTest > testSetConfigs STARTED

org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperatorTest > testSetConfigs PASSED

org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperatorTest > testHandleExceptionInTransformations STARTED

org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperatorTest > testHandleExceptionInTransformations PASSED

org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperatorTest > testHandleExceptionInHeaderConverter STARTED

org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperatorTest > testHandleExceptionInHeaderConverter PASSED

org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperatorTest > testHandleExceptionInValueConverter STARTED

org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperatorTest > testHandleExceptionInValueConverter PASSED

org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperatorTest > testHandleExceptionInTaskPut STARTED

org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperatorTest > testHandleExceptionInTaskPut PASSED

org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperatorTest > testHandleExceptionInTaskPoll STARTED

org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperatorTest > testHandleExceptionInTaskPoll PASSED

org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperatorTest > testThrowExceptionInTaskPut STARTED

org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperatorTest > testThrowExceptionInTaskPut PASSED

org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperatorTest > testThrowExceptionInTaskPoll STARTED

org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperatorTest > testThrowExceptionInTaskPoll PASSED

org.apache.kafka.connect.runtime.errors.ErrorReporterTest > initializeDLQWithNullMetrics STARTED

org.apache.kafka.connect.runtime.errors.ErrorReporterTest > initializeDLQWithNullMetrics PASSED

org.apache.kafka.connect.runtime.errors.ErrorReporterTest > testDLQConfigWithEmptyTopicName STARTED

org.apache.kafka.connect.runtime.errors.ErrorReporterTest > testDLQConfigWithEmptyTopicName PASSED

org.apache.kafka.connect.runtime.errors.ErrorReporterTest > testDLQConfigWithValidTopicName STARTED

org.apache.kafka.connect.runtime.errors.ErrorReporterTest > testDLQConfigWithValidTopicName PASSED

org.apache.kafka.connect.runtime.errors.ErrorReporterTest > testReportDLQTwice STARTED

org.apache.kafka.connect.runtime.errors.ErrorReporterTest > testReportDLQTwice PASSED

org.apache.kafka.connect.runtime.errors.ErrorReporterTest > testLogOnDisabledLogReporter STARTED

org.apache.kafka.connect.runtime.errors.ErrorReporterTest > testLogOnDisabledLogReporter PASSED

org.apache.kafka.connect.runtime.errors.ErrorReporterTest > testLogOnEnabledLogReporter STARTED

org.apache.kafka.connect.runtime.errors.ErrorReporterTest > testLogOnEnabledLogReporter PASSED

org.apache.kafka.connect.runtime.errors.ErrorReporterTest > testLogMessageWithNoRecords STARTED

org.apache.kafka.connect.runtime.errors.ErrorReporterTest > testLogMessageWithNoRecords PASSED

org.apache.kafka.connect.runtime.errors.ErrorReporterTest > testLogMessageWithSinkRecords STARTED

org.apache.kafka.connect.runtime.errors.ErrorReporterTest > testLogMessageWithSinkRecords PASSED

org.apache.kafka.connect.runtime.errors.ErrorReporterTest > testSetDLQConfigs STARTED

org.apache.kafka.connect.runtime.errors.ErrorReporterTest > testSetDLQConfigs PASSED

org.apache.kafka.connect.runtime.errors.ErrorReporterTest > testDlqHeaderConsumerRecord STARTED

org.apache.kafka.connect.runtime.errors.ErrorReporterTest > testDlqHeaderConsumerRecord PASSED

org.apache.kafka.connect.runtime.errors.ErrorReporterTest > testDlqHeaderOnNullExceptionMessage STARTED

org.apache.kafka.connect.runtime.errors.ErrorReporterTest > testDlqHeaderOnNullExceptionMessage PASSED

org.apache.kafka.connect.runtime.errors.ErrorReporterTest > testDlqHeaderIsAppended STARTED

org.apache.kafka.connect.runtime.errors.ErrorReporterTest > testDlqHeaderIsAppended PASSED

org.apache.kafka.connect.runtime.WorkerTaskTest > standardStartup STARTED

org.apache.kafka.connect.runtime.WorkerTaskTest > standardStartup PASSED

org.apache.kafka.connect.runtime.WorkerTaskTest > stopBeforeStarting STARTED

org.apache.kafka.connect.runtime.WorkerTaskTest > stopBeforeStarting PASSED

org.apache.kafka.connect.runtime.WorkerTaskTest > cancelBeforeStopping STARTED

org.apache.kafka.connect.runtime.WorkerTaskTest > cancelBeforeStopping PASSED

org.apache.kafka.connect.runtime.WorkerTaskTest > updateMetricsOnListenerEventsForStartupPauseResumeAndShutdown STARTED

org.apache.kafka.connect.runtime.WorkerTaskTest > updateMetricsOnListenerEventsForStartupPauseResumeAndShutdown PASSED

org.apache.kafka.connect.runtime.WorkerTaskTest > updateMetricsOnListenerEventsForStartupPauseResumeAndFailure STARTED

org.apache.kafka.connect.runtime.WorkerTaskTest > updateMetricsOnListenerEventsForStartupPauseResumeAndFailure PASSED

org.apache.kafka.connect.runtime.WorkerConfigTransformerTest > testReplaceVariable STARTED

org.apache.kafka.connect.runtime.WorkerConfigTransformerTest > testReplaceVariable PASSED

org.apache.kafka.connect.runtime.WorkerConfigTransformerTest > testReplaceVariableWithTTL STARTED

org.apache.kafka.connect.runtime.WorkerConfigTransformerTest > testReplaceVariableWithTTL PASSED

org.apache.kafka.connect.runtime.WorkerConfigTransformerTest > testReplaceVariableWithTTLAndScheduleRestart STARTED

org.apache.kafka.connect.runtime.WorkerConfigTransformerTest > testReplaceVariableWithTTLAndScheduleRestart PASSED

org.apache.kafka.connect.runtime.WorkerConfigTransformerTest > testReplaceVariableWithTTLFirstCancelThenScheduleRestart STARTED

org.apache.kafka.connect.runtime.WorkerConfigTransformerTest > testReplaceVariableWithTTLFirstCancelThenScheduleRestart PASSED

org.apache.kafka.connect.runtime.WorkerConfigTransformerTest > testTransformNullConfiguration STARTED

org.apache.kafka.connect.runtime.WorkerConfigTransformerTest > testTransformNullConfiguration PASSED

org.apache.kafka.connect.runtime.WorkerSourceTaskTest > testStartPaused STARTED

org.apache.kafka.connect.runtime.WorkerSourceTaskTest > testStartPaused PASSED

org.apache.kafka.connect.runtime.WorkerSourceTaskTest > testPause STARTED

org.apache.kafka.connect.runtime.WorkerSourceTaskTest > testPause PASSED

org.apache.kafka.connect.runtime.WorkerSourceTaskTest > testPollsInBackground STARTED

org.apache.kafka.connect.runtime.WorkerSourceTaskTest > testPollsInBackground PASSED

org.apache.kafka.connect.runtime.WorkerSourceTaskTest > testCommit STARTED

org.apache.kafka.connect.runtime.WorkerSourceTaskTest > testCommit PASSED

org.apache.kafka.connect.runtime.WorkerSourceTaskTest > testCommitFailure STARTED

org.apache.kafka.connect.runtime.WorkerSourceTaskTest > testCommitFailure PASSED

org.apache.kafka.connect.runtime.WorkerSourceTaskTest > testMetricsGroup STARTED

org.apache.kafka.connect.runtime.WorkerSourceTaskTest > testMetricsGroup PASSED

org.apache.kafka.connect.runtime.WorkerSourceTaskTest > testSendRecordsTaskCommitRecordFail STARTED

org.apache.kafka.connect.runtime.WorkerSourceTaskTest > testSendRecordsTaskCommitRecordFail PASSED

org.apache.kafka.connect.runtime.WorkerSourceTaskTest > testSlowTaskStart STARTED

org.apache.kafka.connect.runtime.WorkerSourceTaskTest > testSlowTaskStart PASSED

org.apache.kafka.connect.runtime.WorkerSourceTaskTest > testSendRecordsPropagatesTimestamp STARTED

org.apache.kafka.connect.runtime.WorkerSourceTaskTest > testSendRecordsPropagatesTimestamp PASSED

org.apache.kafka.connect.runtime.WorkerSourceTaskTest > testSendRecordsCorruptTimestamp STARTED

org.apache.kafka.connect.runtime.WorkerSourceTaskTest > testSendRecordsCorruptTimestamp PASSED

org.apache.kafka.connect.runtime.WorkerSourceTaskTest > testSendRecordsNoTimestamp STARTED

org.apache.kafka.connect.runtime.WorkerSourceTaskTest > testSendRecordsNoTimestamp PASSED

org.apache.kafka.connect.runtime.WorkerSourceTaskTest > testSendRecordsRetries STARTED

org.apache.kafka.connect.runtime.WorkerSourceTaskTest > testSendRecordsRetries PASSED

org.apache.kafka.connect.runtime.WorkerSourceTaskTest > testFailureInPoll STARTED

org.apache.kafka.connect.runtime.WorkerSourceTaskTest > testFailureInPoll PASSED

org.apache.kafka.connect.runtime.WorkerSourceTaskTest > testSendRecordsConvertsData STARTED

org.apache.kafka.connect.runtime.WorkerSourceTaskTest > testSendRecordsConvertsData PASSED

481 tests completed, 1 failed

> Task :connect:runtime:test FAILED

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':connect:runtime:test'.
> There were failing tests. See the report at: file://<https://builds.apache.org/job/kafka-trunk-jdk11/ws/connect/runtime/build/reports/tests/test/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 6.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/5.4.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 3h 42m 26s
178 actionable tasks: 152 executed, 26 up-to-date
Build step 'Execute shell' marked build as failure
Recording test results
ERROR: No tool found matching GRADLE_4_10_2_HOME
Setting GRADLE_4_10_3_HOME=/home/jenkins/tools/gradle/4.10.3
ERROR: No tool found matching GRADLE_4_10_2_HOME
Setting GRADLE_4_10_3_HOME=/home/jenkins/tools/gradle/4.10.3
Not sending mail to unregistered user ismael@juma.me.uk
Not sending mail to unregistered user vahid.hashemian@gmail.com
Not sending mail to unregistered user noreply@github.com
Not sending mail to unregistered user wangguoz@gmail.com
Not sending mail to unregistered user gwen@confluent.io

Build failed in Jenkins: kafka-trunk-jdk11 #751

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/kafka-trunk-jdk11/751/display/redirect?page=changes>

Changes:

[gwen] KAFKA-8391; Improved the Connect integration tests to make them less

[jason] MINOR: Add fetch from follower system test (#7166)

[bbejeck] KAFKA-8736: Track size in InMemoryKeyValueStore (#7177)

------------------------------------------
[...truncated 6.52 MB...]
org.apache.kafka.streams.TopologyTestDriverTest > shouldPopulateGlobalStore[Eos enabled = true] PASSED

org.apache.kafka.streams.TopologyTestDriverTest > shouldThrowIfPersistentBuiltInStoreIsAccessedWithUntypedMethod[Eos enabled = true] STARTED

org.apache.kafka.streams.TopologyTestDriverTest > shouldThrowIfPersistentBuiltInStoreIsAccessedWithUntypedMethod[Eos enabled = true] PASSED

org.apache.kafka.streams.TopologyTestDriverTest > shouldAllowPrePopulatingStatesStoresWithCachingEnabled[Eos enabled = true] STARTED

org.apache.kafka.streams.TopologyTestDriverTest > shouldAllowPrePopulatingStatesStoresWithCachingEnabled[Eos enabled = true] PASSED

org.apache.kafka.streams.TopologyTestDriverTest > shouldReturnCorrectPersistentStoreTypeOnly[Eos enabled = true] STARTED

org.apache.kafka.streams.TopologyTestDriverTest > shouldReturnCorrectPersistentStoreTypeOnly[Eos enabled = true] PASSED

org.apache.kafka.streams.TopologyTestDriverTest > shouldUseSourceSpecificDeserializers[Eos enabled = true] STARTED

org.apache.kafka.streams.TopologyTestDriverTest > shouldUseSourceSpecificDeserializers[Eos enabled = true] PASSED

org.apache.kafka.streams.TopologyTestDriverTest > shouldReturnAllStores[Eos enabled = true] STARTED

org.apache.kafka.streams.TopologyTestDriverTest > shouldReturnAllStores[Eos enabled = true] PASSED

org.apache.kafka.streams.TopologyTestDriverTest > shouldSendRecordViaCorrectSourceTopic[Eos enabled = true] STARTED

org.apache.kafka.streams.TopologyTestDriverTest > shouldSendRecordViaCorrectSourceTopic[Eos enabled = true] PASSED

org.apache.kafka.streams.TopologyTestDriverTest > shouldNotCreateStateDirectoryForStatelessTopology[Eos enabled = true] STARTED

org.apache.kafka.streams.TopologyTestDriverTest > shouldNotCreateStateDirectoryForStatelessTopology[Eos enabled = true] PASSED

org.apache.kafka.streams.TopologyTestDriverTest > shouldReturnAllStoresNames[Eos enabled = true] STARTED

org.apache.kafka.streams.TopologyTestDriverTest > shouldReturnAllStoresNames[Eos enabled = true] PASSED

org.apache.kafka.streams.TopologyTestDriverTest > shouldProcessConsumerRecordList[Eos enabled = true] STARTED

org.apache.kafka.streams.TopologyTestDriverTest > shouldProcessConsumerRecordList[Eos enabled = true] PASSED

org.apache.kafka.streams.TopologyTestDriverTest > shouldUseSinkSpecificSerializers[Eos enabled = true] STARTED

org.apache.kafka.streams.TopologyTestDriverTest > shouldUseSinkSpecificSerializers[Eos enabled = true] PASSED

org.apache.kafka.streams.TopologyTestDriverTest > shouldFlushStoreForFirstInput[Eos enabled = true] STARTED

org.apache.kafka.streams.TopologyTestDriverTest > shouldFlushStoreForFirstInput[Eos enabled = true] PASSED

org.apache.kafka.streams.TopologyTestDriverTest > shouldProcessFromSourceThatMatchPattern[Eos enabled = true] STARTED

org.apache.kafka.streams.TopologyTestDriverTest > shouldProcessFromSourceThatMatchPattern[Eos enabled = true] PASSED

org.apache.kafka.streams.TopologyTestDriverTest > shouldUpdateStoreForNewKey[Eos enabled = true] STARTED

org.apache.kafka.streams.TopologyTestDriverTest > shouldUpdateStoreForNewKey[Eos enabled = true] PASSED

org.apache.kafka.streams.TopologyTestDriverTest > shouldPunctuateOnWallClockTime[Eos enabled = true] STARTED

org.apache.kafka.streams.TopologyTestDriverTest > shouldPunctuateOnWallClockTime[Eos enabled = true] PASSED

org.apache.kafka.streams.TopologyTestDriverTest > shouldSetRecordMetadata[Eos enabled = true] STARTED

org.apache.kafka.streams.TopologyTestDriverTest > shouldSetRecordMetadata[Eos enabled = true] PASSED

org.apache.kafka.streams.TopologyTestDriverTest > shouldNotUpdateStoreForLargerValue[Eos enabled = true] STARTED

org.apache.kafka.streams.TopologyTestDriverTest > shouldNotUpdateStoreForLargerValue[Eos enabled = true] PASSED

org.apache.kafka.streams.TopologyTestDriverTest > shouldReturnCorrectInMemoryStoreTypeOnly[Eos enabled = true] STARTED

org.apache.kafka.streams.TopologyTestDriverTest > shouldReturnCorrectInMemoryStoreTypeOnly[Eos enabled = true] PASSED

org.apache.kafka.streams.TopologyTestDriverTest > shouldProcessRecordForTopic[Eos enabled = true] STARTED

org.apache.kafka.streams.TopologyTestDriverTest > shouldProcessRecordForTopic[Eos enabled = true] PASSED

org.apache.kafka.streams.TopologyTestDriverTest > shouldForwardRecordsFromSubtopologyToSubtopology[Eos enabled = true] STARTED

org.apache.kafka.streams.TopologyTestDriverTest > shouldForwardRecordsFromSubtopologyToSubtopology[Eos enabled = true] PASSED

org.apache.kafka.streams.TopologyTestDriverTest > shouldNotUpdateStoreForSmallerValue[Eos enabled = true] STARTED

org.apache.kafka.streams.TopologyTestDriverTest > shouldNotUpdateStoreForSmallerValue[Eos enabled = true] PASSED

org.apache.kafka.streams.TopologyTestDriverTest > shouldCreateStateDirectoryForStatefulTopology[Eos enabled = true] STARTED

org.apache.kafka.streams.TopologyTestDriverTest > shouldCreateStateDirectoryForStatefulTopology[Eos enabled = true] PASSED

org.apache.kafka.streams.TopologyTestDriverTest > shouldPunctuateIfWallClockTimeAdvances[Eos enabled = true] STARTED

org.apache.kafka.streams.TopologyTestDriverTest > shouldPunctuateIfWallClockTimeAdvances[Eos enabled = true] PASSED

org.apache.kafka.streams.TopologyTestDriverTest > shouldCloseProcessor[Eos enabled = false] STARTED

org.apache.kafka.streams.TopologyTestDriverTest > shouldCloseProcessor[Eos enabled = false] PASSED

org.apache.kafka.streams.TopologyTestDriverTest > shouldFeedStoreFromGlobalKTable[Eos enabled = false] STARTED

org.apache.kafka.streams.TopologyTestDriverTest > shouldFeedStoreFromGlobalKTable[Eos enabled = false] PASSED

org.apache.kafka.streams.TopologyTestDriverTest > shouldCleanUpPersistentStateStoresOnClose[Eos enabled = false] STARTED

org.apache.kafka.streams.TopologyTestDriverTest > shouldCleanUpPersistentStateStoresOnClose[Eos enabled = false] PASSED

org.apache.kafka.streams.TopologyTestDriverTest > shouldThrowPatternNotValidForTopicNameException[Eos enabled = false] STARTED

org.apache.kafka.streams.TopologyTestDriverTest > shouldThrowPatternNotValidForTopicNameException[Eos enabled = false] PASSED

org.apache.kafka.streams.TopologyTestDriverTest > shouldPunctuateIfEvenTimeAdvances[Eos enabled = false] STARTED

org.apache.kafka.streams.TopologyTestDriverTest > shouldPunctuateIfEvenTimeAdvances[Eos enabled = false] PASSED

org.apache.kafka.streams.TopologyTestDriverTest > shouldInitProcessor[Eos enabled = false] STARTED

org.apache.kafka.streams.TopologyTestDriverTest > shouldInitProcessor[Eos enabled = false] PASSED

org.apache.kafka.streams.TopologyTestDriverTest > shouldThrowForUnknownTopic[Eos enabled = false] STARTED

org.apache.kafka.streams.TopologyTestDriverTest > shouldThrowForUnknownTopic[Eos enabled = false] PASSED

org.apache.kafka.streams.TopologyTestDriverTest > shouldPunctuateOnStreamsTime[Eos enabled = false] STARTED

org.apache.kafka.streams.TopologyTestDriverTest > shouldPunctuateOnStreamsTime[Eos enabled = false] PASSED

org.apache.kafka.streams.TopologyTestDriverTest > shouldThrowIfInMemoryBuiltInStoreIsAccessedWithUntypedMethod[Eos enabled = false] STARTED

org.apache.kafka.streams.TopologyTestDriverTest > shouldThrowIfInMemoryBuiltInStoreIsAccessedWithUntypedMethod[Eos enabled = false] PASSED

org.apache.kafka.streams.TopologyTestDriverTest > shouldProcessFromSourcesThatMatchMultiplePattern[Eos enabled = false] STARTED

org.apache.kafka.streams.TopologyTestDriverTest > shouldProcessFromSourcesThatMatchMultiplePattern[Eos enabled = false] PASSED

org.apache.kafka.streams.TopologyTestDriverTest > shouldPopulateGlobalStore[Eos enabled = false] STARTED

org.apache.kafka.streams.TopologyTestDriverTest > shouldPopulateGlobalStore[Eos enabled = false] PASSED

org.apache.kafka.streams.TopologyTestDriverTest > shouldThrowIfPersistentBuiltInStoreIsAccessedWithUntypedMethod[Eos enabled = false] STARTED

org.apache.kafka.streams.TopologyTestDriverTest > shouldThrowIfPersistentBuiltInStoreIsAccessedWithUntypedMethod[Eos enabled = false] PASSED

org.apache.kafka.streams.TopologyTestDriverTest > shouldAllowPrePopulatingStatesStoresWithCachingEnabled[Eos enabled = false] STARTED

org.apache.kafka.streams.TopologyTestDriverTest > shouldAllowPrePopulatingStatesStoresWithCachingEnabled[Eos enabled = false] PASSED

org.apache.kafka.streams.TopologyTestDriverTest > shouldReturnCorrectPersistentStoreTypeOnly[Eos enabled = false] STARTED

org.apache.kafka.streams.TopologyTestDriverTest > shouldReturnCorrectPersistentStoreTypeOnly[Eos enabled = false] PASSED

org.apache.kafka.streams.TopologyTestDriverTest > shouldUseSourceSpecificDeserializers[Eos enabled = false] STARTED

org.apache.kafka.streams.TopologyTestDriverTest > shouldUseSourceSpecificDeserializers[Eos enabled = false] PASSED

org.apache.kafka.streams.TopologyTestDriverTest > shouldReturnAllStores[Eos enabled = false] STARTED

org.apache.kafka.streams.TopologyTestDriverTest > shouldReturnAllStores[Eos enabled = false] PASSED

org.apache.kafka.streams.TopologyTestDriverTest > shouldSendRecordViaCorrectSourceTopic[Eos enabled = false] STARTED

org.apache.kafka.streams.TopologyTestDriverTest > shouldSendRecordViaCorrectSourceTopic[Eos enabled = false] PASSED

org.apache.kafka.streams.TopologyTestDriverTest > shouldNotCreateStateDirectoryForStatelessTopology[Eos enabled = false] STARTED

org.apache.kafka.streams.TopologyTestDriverTest > shouldNotCreateStateDirectoryForStatelessTopology[Eos enabled = false] PASSED

org.apache.kafka.streams.TopologyTestDriverTest > shouldReturnAllStoresNames[Eos enabled = false] STARTED

org.apache.kafka.streams.TopologyTestDriverTest > shouldReturnAllStoresNames[Eos enabled = false] PASSED

org.apache.kafka.streams.TopologyTestDriverTest > shouldProcessConsumerRecordList[Eos enabled = false] STARTED

org.apache.kafka.streams.TopologyTestDriverTest > shouldProcessConsumerRecordList[Eos enabled = false] PASSED

org.apache.kafka.streams.TopologyTestDriverTest > shouldUseSinkSpecificSerializers[Eos enabled = false] STARTED

org.apache.kafka.streams.TopologyTestDriverTest > shouldUseSinkSpecificSerializers[Eos enabled = false] PASSED

org.apache.kafka.streams.TopologyTestDriverTest > shouldFlushStoreForFirstInput[Eos enabled = false] STARTED

org.apache.kafka.streams.TopologyTestDriverTest > shouldFlushStoreForFirstInput[Eos enabled = false] PASSED

org.apache.kafka.streams.TopologyTestDriverTest > shouldProcessFromSourceThatMatchPattern[Eos enabled = false] STARTED

org.apache.kafka.streams.TopologyTestDriverTest > shouldProcessFromSourceThatMatchPattern[Eos enabled = false] PASSED

org.apache.kafka.streams.TopologyTestDriverTest > shouldUpdateStoreForNewKey[Eos enabled = false] STARTED

org.apache.kafka.streams.TopologyTestDriverTest > shouldUpdateStoreForNewKey[Eos enabled = false] PASSED

org.apache.kafka.streams.TopologyTestDriverTest > shouldPunctuateOnWallClockTime[Eos enabled = false] STARTED

org.apache.kafka.streams.TopologyTestDriverTest > shouldPunctuateOnWallClockTime[Eos enabled = false] PASSED

org.apache.kafka.streams.TopologyTestDriverTest > shouldSetRecordMetadata[Eos enabled = false] STARTED

org.apache.kafka.streams.TopologyTestDriverTest > shouldSetRecordMetadata[Eos enabled = false] PASSED

org.apache.kafka.streams.TopologyTestDriverTest > shouldNotUpdateStoreForLargerValue[Eos enabled = false] STARTED

org.apache.kafka.streams.TopologyTestDriverTest > shouldNotUpdateStoreForLargerValue[Eos enabled = false] PASSED

org.apache.kafka.streams.TopologyTestDriverTest > shouldReturnCorrectInMemoryStoreTypeOnly[Eos enabled = false] STARTED

org.apache.kafka.streams.TopologyTestDriverTest > shouldReturnCorrectInMemoryStoreTypeOnly[Eos enabled = false] PASSED

org.apache.kafka.streams.TopologyTestDriverTest > shouldProcessRecordForTopic[Eos enabled = false] STARTED

org.apache.kafka.streams.TopologyTestDriverTest > shouldProcessRecordForTopic[Eos enabled = false] PASSED

org.apache.kafka.streams.TopologyTestDriverTest > shouldForwardRecordsFromSubtopologyToSubtopology[Eos enabled = false] STARTED

org.apache.kafka.streams.TopologyTestDriverTest > shouldForwardRecordsFromSubtopologyToSubtopology[Eos enabled = false] PASSED

org.apache.kafka.streams.TopologyTestDriverTest > shouldNotUpdateStoreForSmallerValue[Eos enabled = false] STARTED

org.apache.kafka.streams.TopologyTestDriverTest > shouldNotUpdateStoreForSmallerValue[Eos enabled = false] PASSED

org.apache.kafka.streams.TopologyTestDriverTest > shouldCreateStateDirectoryForStatefulTopology[Eos enabled = false] STARTED

org.apache.kafka.streams.TopologyTestDriverTest > shouldCreateStateDirectoryForStatefulTopology[Eos enabled = false] PASSED

org.apache.kafka.streams.TopologyTestDriverTest > shouldPunctuateIfWallClockTimeAdvances[Eos enabled = false] STARTED

org.apache.kafka.streams.TopologyTestDriverTest > shouldPunctuateIfWallClockTimeAdvances[Eos enabled = false] PASSED

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':core:test'.
> There were failing tests. See the report at: file://<https://builds.apache.org/job/kafka-trunk-jdk11/ws/core/build/reports/tests/test/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 6.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/5.4.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 4h 12m 37s
178 actionable tasks: 175 executed, 3 up-to-date
Build step 'Execute shell' marked build as failure
Recording test results
ERROR: No tool found matching GRADLE_4_10_2_HOME
Setting GRADLE_4_10_3_HOME=/home/jenkins/tools/gradle/4.10.3
ERROR: No tool found matching GRADLE_4_10_2_HOME
Setting GRADLE_4_10_3_HOME=/home/jenkins/tools/gradle/4.10.3
Not sending mail to unregistered user ismael@juma.me.uk
Not sending mail to unregistered user vahid.hashemian@gmail.com
Not sending mail to unregistered user noreply@github.com
Not sending mail to unregistered user gwen@confluent.io

Build failed in Jenkins: kafka-trunk-jdk11 #750

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/kafka-trunk-jdk11/750/display/redirect?page=changes>

Changes:

[rhauch] KAFKA-8774: Regex can be found anywhere in config value (#7197)

[rhauch] KAFKA-7941: Catch TimeoutException in KafkaBasedLog worker thread

------------------------------------------
[...truncated 6.51 MB...]

org.apache.kafka.connect.transforms.TimestampConverterTest > testWithSchemaNullFieldToTime PASSED

org.apache.kafka.connect.transforms.TimestampConverterTest > testWithSchemaNullFieldToUnix STARTED

org.apache.kafka.connect.transforms.TimestampConverterTest > testWithSchemaNullFieldToUnix PASSED

org.apache.kafka.connect.transforms.TimestampConverterTest > testConfigNoTargetType STARTED

org.apache.kafka.connect.transforms.TimestampConverterTest > testConfigNoTargetType PASSED

org.apache.kafka.connect.transforms.TimestampConverterTest > testWithSchemaStringToTimestamp STARTED

org.apache.kafka.connect.transforms.TimestampConverterTest > testWithSchemaStringToTimestamp PASSED

org.apache.kafka.connect.transforms.TimestampConverterTest > testWithSchemaTimeToTimestamp STARTED

org.apache.kafka.connect.transforms.TimestampConverterTest > testWithSchemaTimeToTimestamp PASSED

org.apache.kafka.connect.transforms.TimestampConverterTest > testSchemalessUnixToTimestamp STARTED

org.apache.kafka.connect.transforms.TimestampConverterTest > testSchemalessUnixToTimestamp PASSED

org.apache.kafka.connect.transforms.TimestampConverterTest > testWithSchemaTimestampToString STARTED

org.apache.kafka.connect.transforms.TimestampConverterTest > testWithSchemaTimestampToString PASSED

org.apache.kafka.connect.transforms.TimestampConverterTest > testConfigInvalidFormat STARTED

org.apache.kafka.connect.transforms.TimestampConverterTest > testConfigInvalidFormat PASSED

> Task :streams:examples:test

org.apache.kafka.streams.examples.wordcount.WordCountProcessorTest > test STARTED

org.apache.kafka.streams.examples.wordcount.WordCountProcessorTest > test PASSED

> Task :streams:streams-scala:test

org.apache.kafka.streams.scala.kstream.KTableTest > filter a KTable should filter records satisfying the predicate STARTED

org.apache.kafka.streams.scala.kstream.KTableTest > filter a KTable should filter records satisfying the predicate PASSED

org.apache.kafka.streams.scala.kstream.KTableTest > filterNot a KTable should filter records not satisfying the predicate STARTED

org.apache.kafka.streams.scala.kstream.KTableTest > filterNot a KTable should filter records not satisfying the predicate PASSED

org.apache.kafka.streams.scala.kstream.KTableTest > join 2 KTables should join correctly records STARTED

org.apache.kafka.streams.scala.kstream.KTableTest > join 2 KTables should join correctly records PASSED

org.apache.kafka.streams.scala.kstream.KTableTest > join 2 KTables with a Materialized should join correctly records and state store STARTED

org.apache.kafka.streams.scala.kstream.KTableTest > join 2 KTables with a Materialized should join correctly records and state store PASSED

org.apache.kafka.streams.scala.kstream.KTableTest > windowed KTable#suppress should correctly suppress results using Suppressed.untilTimeLimit STARTED

org.apache.kafka.streams.scala.kstream.KTableTest > windowed KTable#suppress should correctly suppress results using Suppressed.untilTimeLimit PASSED

org.apache.kafka.streams.scala.kstream.KTableTest > windowed KTable#suppress should correctly suppress results using Suppressed.untilWindowCloses STARTED

org.apache.kafka.streams.scala.kstream.KTableTest > windowed KTable#suppress should correctly suppress results using Suppressed.untilWindowCloses PASSED

org.apache.kafka.streams.scala.kstream.KTableTest > session windowed KTable#suppress should correctly suppress results using Suppressed.untilWindowCloses STARTED

org.apache.kafka.streams.scala.kstream.KTableTest > session windowed KTable#suppress should correctly suppress results using Suppressed.untilWindowCloses PASSED

org.apache.kafka.streams.scala.kstream.KTableTest > non-windowed KTable#suppress should correctly suppress results using Suppressed.untilTimeLimit STARTED

org.apache.kafka.streams.scala.kstream.KTableTest > non-windowed KTable#suppress should correctly suppress results using Suppressed.untilTimeLimit PASSED

org.apache.kafka.streams.scala.kstream.ProducedTest > Create a Produced should create a Produced with Serdes STARTED

org.apache.kafka.streams.scala.kstream.ProducedTest > Create a Produced should create a Produced with Serdes PASSED

org.apache.kafka.streams.scala.kstream.ProducedTest > Create a Produced with timestampExtractor and resetPolicy should create a Consumed with Serdes, timestampExtractor and resetPolicy STARTED

org.apache.kafka.streams.scala.kstream.ProducedTest > Create a Produced with timestampExtractor and resetPolicy should create a Consumed with Serdes, timestampExtractor and resetPolicy PASSED

org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed should create a Consumed with Serdes STARTED

org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed should create a Consumed with Serdes PASSED

org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed with timestampExtractor and resetPolicy should create a Consumed with Serdes, timestampExtractor and resetPolicy STARTED

org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed with timestampExtractor and resetPolicy should create a Consumed with Serdes, timestampExtractor and resetPolicy PASSED

org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed with timestampExtractor should create a Consumed with Serdes and timestampExtractor STARTED

org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed with timestampExtractor should create a Consumed with Serdes and timestampExtractor PASSED

org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed with resetPolicy should create a Consumed with Serdes and resetPolicy STARTED

org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed with resetPolicy should create a Consumed with Serdes and resetPolicy PASSED

org.apache.kafka.streams.scala.kstream.JoinedTest > Create a Joined should create a Joined with Serdes STARTED

org.apache.kafka.streams.scala.kstream.JoinedTest > Create a Joined should create a Joined with Serdes PASSED

org.apache.kafka.streams.scala.kstream.JoinedTest > Create a Joined should create a Joined with Serdes and repartition topic name STARTED

org.apache.kafka.streams.scala.kstream.JoinedTest > Create a Joined should create a Joined with Serdes and repartition topic name PASSED

org.apache.kafka.streams.scala.kstream.KStreamTest > filter a KStream should filter records satisfying the predicate STARTED

org.apache.kafka.streams.scala.kstream.KStreamTest > filter a KStream should filter records satisfying the predicate PASSED

org.apache.kafka.streams.scala.kstream.KStreamTest > filterNot a KStream should filter records not satisfying the predicate STARTED

org.apache.kafka.streams.scala.kstream.KStreamTest > filterNot a KStream should filter records not satisfying the predicate PASSED

org.apache.kafka.streams.scala.kstream.KStreamTest > foreach a KStream should run foreach actions on records STARTED

org.apache.kafka.streams.scala.kstream.KStreamTest > foreach a KStream should run foreach actions on records PASSED

org.apache.kafka.streams.scala.kstream.KStreamTest > peek a KStream should run peek actions on records STARTED

org.apache.kafka.streams.scala.kstream.KStreamTest > peek a KStream should run peek actions on records PASSED

org.apache.kafka.streams.scala.kstream.KStreamTest > selectKey a KStream should select a new key STARTED

org.apache.kafka.streams.scala.kstream.KStreamTest > selectKey a KStream should select a new key PASSED

org.apache.kafka.streams.scala.kstream.KStreamTest > join 2 KStreams should join correctly records STARTED

org.apache.kafka.streams.scala.kstream.KStreamTest > join 2 KStreams should join correctly records PASSED

org.apache.kafka.streams.scala.kstream.SuppressedTest > Suppressed.untilWindowCloses should produce the correct suppression STARTED

org.apache.kafka.streams.scala.kstream.SuppressedTest > Suppressed.untilWindowCloses should produce the correct suppression PASSED

org.apache.kafka.streams.scala.kstream.SuppressedTest > Suppressed.untilTimeLimit should produce the correct suppression STARTED

org.apache.kafka.streams.scala.kstream.SuppressedTest > Suppressed.untilTimeLimit should produce the correct suppression PASSED

org.apache.kafka.streams.scala.kstream.SuppressedTest > BufferConfig.maxRecords should produce the correct buffer config STARTED

org.apache.kafka.streams.scala.kstream.SuppressedTest > BufferConfig.maxRecords should produce the correct buffer config PASSED

org.apache.kafka.streams.scala.kstream.SuppressedTest > BufferConfig.maxBytes should produce the correct buffer config STARTED

org.apache.kafka.streams.scala.kstream.SuppressedTest > BufferConfig.maxBytes should produce the correct buffer config PASSED

org.apache.kafka.streams.scala.kstream.SuppressedTest > BufferConfig.unbounded should produce the correct buffer config STARTED

org.apache.kafka.streams.scala.kstream.SuppressedTest > BufferConfig.unbounded should produce the correct buffer config PASSED

org.apache.kafka.streams.scala.kstream.SuppressedTest > BufferConfig should support very long chains of factory methods STARTED

org.apache.kafka.streams.scala.kstream.SuppressedTest > BufferConfig should support very long chains of factory methods PASSED

org.apache.kafka.streams.scala.kstream.GroupedTest > Create a Grouped should create a Grouped with Serdes STARTED

org.apache.kafka.streams.scala.kstream.GroupedTest > Create a Grouped should create a Grouped with Serdes PASSED

org.apache.kafka.streams.scala.kstream.GroupedTest > Create a Grouped with repartition topic name should create a Grouped with Serdes, and repartition topic name STARTED

org.apache.kafka.streams.scala.kstream.GroupedTest > Create a Grouped with repartition topic name should create a Grouped with Serdes, and repartition topic name PASSED

org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialized should create a Materialized with Serdes STARTED

org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialized should create a Materialized with Serdes PASSED

org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize with a store name should create a Materialized with Serdes and a store name STARTED

org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize with a store name should create a Materialized with Serdes and a store name PASSED

org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize with a window store supplier should create a Materialized with Serdes and a store supplier STARTED

org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize with a window store supplier should create a Materialized with Serdes and a store supplier PASSED

org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize with a key value store supplier should create a Materialized with Serdes and a store supplier STARTED

org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize with a key value store supplier should create a Materialized with Serdes and a store supplier PASSED

org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize with a session store supplier should create a Materialized with Serdes and a store supplier STARTED

org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize with a session store supplier should create a Materialized with Serdes and a store supplier PASSED

org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaJoin STARTED

org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaJoin PASSED

org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaSimple STARTED

org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaSimple PASSED

org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaAggregate STARTED

org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaAggregate PASSED

org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaProperties STARTED

org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaProperties PASSED

org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaTransform STARTED

org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaTransform PASSED

org.apache.kafka.streams.scala.StreamToTableJoinScalaIntegrationTestImplicitSerdes > testShouldCountClicksPerRegionWithNamedRepartitionTopic STARTED

org.apache.kafka.streams.scala.StreamToTableJoinScalaIntegrationTestImplicitSerdes > testShouldCountClicksPerRegionWithNamedRepartitionTopic PASSED

org.apache.kafka.streams.scala.StreamToTableJoinScalaIntegrationTestImplicitSerdes > testShouldCountClicksPerRegionJava STARTED

org.apache.kafka.streams.scala.StreamToTableJoinScalaIntegrationTestImplicitSerdes > testShouldCountClicksPerRegionJava PASSED

org.apache.kafka.streams.scala.StreamToTableJoinScalaIntegrationTestImplicitSerdes > testShouldCountClicksPerRegion STARTED

org.apache.kafka.streams.scala.StreamToTableJoinScalaIntegrationTestImplicitSerdes > testShouldCountClicksPerRegion PASSED

org.apache.kafka.streams.scala.WordCountTest > testShouldCountWordsMaterialized STARTED

org.apache.kafka.streams.scala.WordCountTest > testShouldCountWordsMaterialized PASSED

org.apache.kafka.streams.scala.WordCountTest > testShouldCountWordsJava STARTED

org.apache.kafka.streams.scala.WordCountTest > testShouldCountWordsJava PASSED

org.apache.kafka.streams.scala.WordCountTest > testShouldCountWords STARTED

org.apache.kafka.streams.scala.WordCountTest > testShouldCountWords PASSED

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':connect:runtime:test'.
> There were failing tests. See the report at: file://<https://builds.apache.org/job/kafka-trunk-jdk11/ws/connect/runtime/build/reports/tests/test/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 6.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/5.4.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 3h 7m 25s
178 actionable tasks: 175 executed, 3 up-to-date
Build step 'Execute shell' marked build as failure
Recording test results
ERROR: No tool found matching GRADLE_4_10_2_HOME
Setting GRADLE_4_10_3_HOME=/home/jenkins/tools/gradle/4.10.3
ERROR: No tool found matching GRADLE_4_10_2_HOME
Setting GRADLE_4_10_3_HOME=/home/jenkins/tools/gradle/4.10.3
Not sending mail to unregistered user ismael@juma.me.uk
Not sending mail to unregistered user vahid.hashemian@gmail.com
Not sending mail to unregistered user noreply@github.com

Build failed in Jenkins: kafka-trunk-jdk11 #749

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/kafka-trunk-jdk11/749/display/redirect?page=changes>

Changes:

[cmccabe] MINOR: Clean up the sticky partitioner code a bit (#7151)

------------------------------------------
[...truncated 6.38 MB...]
org.apache.kafka.streams.TopologyTest > multipleSourcesWithSinksShouldHaveDistinctSubtopologies STARTED

org.apache.kafka.streams.TopologyTest > multipleSourcesWithSinksShouldHaveDistinctSubtopologies PASSED

org.apache.kafka.streams.TopologyTest > kGroupedStreamZeroArgCountShouldPreserveTopologyStructure STARTED

org.apache.kafka.streams.TopologyTest > kGroupedStreamZeroArgCountShouldPreserveTopologyStructure PASSED

org.apache.kafka.streams.TopologyTest > sessionWindowNamedMaterializedCountShouldPreserveTopologyStructure STARTED

org.apache.kafka.streams.TopologyTest > sessionWindowNamedMaterializedCountShouldPreserveTopologyStructure PASSED

org.apache.kafka.streams.internals.ApiUtilsTest > shouldReturnMillisecondsOnValidDuration STARTED

org.apache.kafka.streams.internals.ApiUtilsTest > shouldReturnMillisecondsOnValidDuration PASSED

org.apache.kafka.streams.internals.ApiUtilsTest > shouldThrowArithmeticExceptionForMaxInstant STARTED

org.apache.kafka.streams.internals.ApiUtilsTest > shouldThrowArithmeticExceptionForMaxInstant PASSED

org.apache.kafka.streams.internals.ApiUtilsTest > shouldThrowArithmeticExceptionForMaxDuration STARTED

org.apache.kafka.streams.internals.ApiUtilsTest > shouldThrowArithmeticExceptionForMaxDuration PASSED

org.apache.kafka.streams.internals.ApiUtilsTest > shouldReturnMillisecondsOnValidInstant STARTED

org.apache.kafka.streams.internals.ApiUtilsTest > shouldReturnMillisecondsOnValidInstant PASSED

org.apache.kafka.streams.internals.ApiUtilsTest > shouldThrowNullPointerExceptionForNullInstant STARTED

org.apache.kafka.streams.internals.ApiUtilsTest > shouldThrowNullPointerExceptionForNullInstant PASSED

org.apache.kafka.streams.internals.ApiUtilsTest > shouldContainsNameAndValueInFailMsgPrefix STARTED

org.apache.kafka.streams.internals.ApiUtilsTest > shouldContainsNameAndValueInFailMsgPrefix PASSED

org.apache.kafka.streams.internals.ApiUtilsTest > shouldThrowNullPointerExceptionForNullDuration STARTED

org.apache.kafka.streams.internals.ApiUtilsTest > shouldThrowNullPointerExceptionForNullDuration PASSED

org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testLeftInner[caching enabled = true] STARTED

org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testLeftInner[caching enabled = true] PASSED

org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testLeftOuter[caching enabled = true] STARTED

org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testLeftOuter[caching enabled = true] PASSED

org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testLeftLeft[caching enabled = true] STARTED

org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testLeftLeft[caching enabled = true] PASSED

org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testOuterLeft[caching enabled = true] STARTED

org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testOuterLeft[caching enabled = true] PASSED

org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testInner[caching enabled = true] STARTED

org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testInner[caching enabled = true] PASSED

org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testOuter[caching enabled = true] STARTED

org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testOuter[caching enabled = true] PASSED

org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testLeft[caching enabled = true] STARTED

org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testLeft[caching enabled = true] PASSED

org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testInnerInner[caching enabled = true] STARTED

org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testInnerInner[caching enabled = true] PASSED

org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testInnerOuter[caching enabled = true] STARTED

org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testInnerOuter[caching enabled = true] PASSED

org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testInnerLeft[caching enabled = true] STARTED

org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testInnerLeft[caching enabled = true] PASSED

org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testOuterInner[caching enabled = true] STARTED

org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testOuterInner[caching enabled = true] PASSED

org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testOuterOuter[caching enabled = true] STARTED

org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testOuterOuter[caching enabled = true] PASSED

org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testLeftInner[caching enabled = false] STARTED

org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testLeftInner[caching enabled = false] PASSED

org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testLeftOuter[caching enabled = false] STARTED

org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testLeftOuter[caching enabled = false] PASSED

org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testLeftLeft[caching enabled = false] STARTED

org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testLeftLeft[caching enabled = false] PASSED

org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testOuterLeft[caching enabled = false] STARTED

org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testOuterLeft[caching enabled = false] PASSED

org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testInner[caching enabled = false] STARTED

org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testInner[caching enabled = false] PASSED

org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testOuter[caching enabled = false] STARTED

org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testOuter[caching enabled = false] PASSED

org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testLeft[caching enabled = false] STARTED

org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testLeft[caching enabled = false] PASSED

org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testInnerInner[caching enabled = false] STARTED

org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testInnerInner[caching enabled = false] PASSED

org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testInnerOuter[caching enabled = false] STARTED

org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testInnerOuter[caching enabled = false] PASSED

org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testInnerLeft[caching enabled = false] STARTED

org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testInnerLeft[caching enabled = false] PASSED

org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testOuterInner[caching enabled = false] STARTED

org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testOuterInner[caching enabled = false] PASSED

org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testOuterOuter[caching enabled = false] STARTED

org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testOuterOuter[caching enabled = false] PASSED

org.apache.kafka.streams.integration.InternalTopicIntegrationTest > shouldCompactTopicsForKeyValueStoreChangelogs STARTED

org.apache.kafka.streams.integration.InternalTopicIntegrationTest > shouldCompactTopicsForKeyValueStoreChangelogs PASSED

org.apache.kafka.streams.integration.InternalTopicIntegrationTest > shouldCompactAndDeleteTopicsForWindowStoreChangelogs STARTED

org.apache.kafka.streams.integration.InternalTopicIntegrationTest > shouldCompactAndDeleteTopicsForWindowStoreChangelogs PASSED

org.apache.kafka.streams.integration.KStreamTransformIntegrationTest > shouldFlatTransform STARTED

org.apache.kafka.streams.integration.KStreamTransformIntegrationTest > shouldFlatTransform PASSED

org.apache.kafka.streams.integration.KStreamTransformIntegrationTest > shouldFlatTransformValuesWithValueTransformerWithoutKey STARTED

org.apache.kafka.streams.integration.KStreamTransformIntegrationTest > shouldFlatTransformValuesWithValueTransformerWithoutKey PASSED

org.apache.kafka.streams.integration.KStreamTransformIntegrationTest > shouldFlatTransformValuesWithKey STARTED

org.apache.kafka.streams.integration.KStreamTransformIntegrationTest > shouldFlatTransformValuesWithKey PASSED

org.apache.kafka.streams.integration.KStreamTransformIntegrationTest > shouldTransform STARTED

org.apache.kafka.streams.integration.KStreamTransformIntegrationTest > shouldTransform PASSED

org.apache.kafka.streams.integration.KStreamTransformIntegrationTest > shouldTransformValuesWithValueTransformerWithKey STARTED

org.apache.kafka.streams.integration.KStreamTransformIntegrationTest > shouldTransformValuesWithValueTransformerWithKey PASSED

org.apache.kafka.streams.integration.KStreamTransformIntegrationTest > shouldTransformValuesWithValueTransformerWithoutKey STARTED

org.apache.kafka.streams.integration.KStreamTransformIntegrationTest > shouldTransformValuesWithValueTransformerWithoutKey PASSED

org.apache.kafka.streams.integration.QueryableStateIntegrationTest > shouldBeAbleToQueryMapValuesState STARTED

org.apache.kafka.streams.integration.QueryableStateIntegrationTest > shouldBeAbleToQueryMapValuesState PASSED

org.apache.kafka.streams.integration.QueryableStateIntegrationTest > shouldNotMakeStoreAvailableUntilAllStoresAvailable STARTED

org.apache.kafka.streams.integration.QueryableStateIntegrationTest > shouldNotMakeStoreAvailableUntilAllStoresAvailable PASSED

org.apache.kafka.streams.integration.QueryableStateIntegrationTest > queryOnRebalance STARTED

org.apache.kafka.streams.integration.QueryableStateIntegrationTest > queryOnRebalance PASSED

org.apache.kafka.streams.integration.QueryableStateIntegrationTest > concurrentAccesses STARTED

org.apache.kafka.streams.integration.QueryableStateIntegrationTest > concurrentAccesses PASSED

org.apache.kafka.streams.integration.QueryableStateIntegrationTest > shouldAllowToQueryAfterThreadDied STARTED

org.apache.kafka.streams.integration.QueryableStateIntegrationTest > shouldAllowToQueryAfterThreadDied PASSED

org.apache.kafka.streams.integration.QueryableStateIntegrationTest > shouldBeAbleToQueryStateWithZeroSizedCache STARTED

org.apache.kafka.streams.integration.QueryableStateIntegrationTest > shouldBeAbleToQueryStateWithZeroSizedCache PASSED

org.apache.kafka.streams.integration.QueryableStateIntegrationTest > shouldBeAbleToQueryMapValuesAfterFilterState STARTED

org.apache.kafka.streams.integration.QueryableStateIntegrationTest > shouldBeAbleToQueryMapValuesAfterFilterState PASSED

org.apache.kafka.streams.integration.QueryableStateIntegrationTest > shouldBeAbleToQueryFilterState STARTED

org.apache.kafka.streams.integration.QueryableStateIntegrationTest > shouldBeAbleToQueryFilterState PASSED

org.apache.kafka.streams.integration.QueryableStateIntegrationTest > shouldBeAbleToQueryStateWithNonZeroSizedCache STARTED

org.apache.kafka.streams.integration.QueryableStateIntegrationTest > shouldBeAbleToQueryStateWithNonZeroSizedCache PASSED

org.apache.kafka.streams.integration.KStreamAggregationDedupIntegrationTest > shouldReduce STARTED

org.apache.kafka.streams.integration.KStreamAggregationDedupIntegrationTest > shouldReduce PASSED

org.apache.kafka.streams.integration.KStreamAggregationDedupIntegrationTest > shouldGroupByKey STARTED

org.apache.kafka.streams.integration.KStreamAggregationDedupIntegrationTest > shouldGroupByKey PASSED

org.apache.kafka.streams.integration.KStreamAggregationDedupIntegrationTest > shouldReduceWindowed STARTED

org.apache.kafka.streams.integration.KStreamAggregationDedupIntegrationTest > shouldReduceWindowed PASSED

org.apache.kafka.streams.integration.PurgeRepartitionTopicIntegrationTest > shouldRestoreState STARTED

org.apache.kafka.streams.integration.PurgeRepartitionTopicIntegrationTest > shouldRestoreState PASSED

org.apache.kafka.streams.integration.RepartitionOptimizingIntegrationTest > shouldSendCorrectRecords_OPTIMIZED STARTED

org.apache.kafka.streams.integration.RepartitionOptimizingIntegrationTest > shouldSendCorrectRecords_OPTIMIZED PASSED

org.apache.kafka.streams.integration.RepartitionOptimizingIntegrationTest > shouldSendCorrectResults_NO_OPTIMIZATION STARTED

org.apache.kafka.streams.integration.RepartitionOptimizingIntegrationTest > shouldSendCorrectResults_NO_OPTIMIZATION PASSED

org.apache.kafka.streams.integration.MetricsIntegrationTest > shouldAddMetricsOnAllLevels STARTED

org.apache.kafka.streams.integration.MetricsIntegrationTest > shouldAddMetricsOnAllLevels PASSED

org.apache.kafka.streams.integration.MetricsIntegrationTest > shouldAddMetricsForWindowStore STARTED

org.apache.kafka.streams.integration.MetricsIntegrationTest > shouldAddMetricsForWindowStore PASSED

org.apache.kafka.streams.integration.MetricsIntegrationTest > shouldAddMetricsForSessionStore STARTED

org.apache.kafka.streams.integration.MetricsIntegrationTest > shouldAddMetricsForSessionStore PASSED
Build timed out (after 480 minutes). Marking the build as failed.
Build was aborted
Recording test results
ERROR: No tool found matching GRADLE_4_10_2_HOME
Setting GRADLE_4_10_3_HOME=/home/jenkins/tools/gradle/4.10.3
ERROR: No tool found matching GRADLE_4_10_2_HOME
Setting GRADLE_4_10_3_HOME=/home/jenkins/tools/gradle/4.10.3
Not sending mail to unregistered user ismael@juma.me.uk
Not sending mail to unregistered user vahid.hashemian@gmail.com
Not sending mail to unregistered user noreply@github.com