You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@kafka.apache.org by Apache Jenkins Server <je...@builds.apache.org> on 2019/08/13 02:59:24 UTC
Build failed in Jenkins: kafka-trunk-jdk11 #748
See <https://builds.apache.org/job/kafka-trunk-jdk11/748/display/redirect?page=changes>
Changes:
[jason] MINOR: Format AdminUtils::assignReplicasToBrokers java documentation
------------------------------------------
[...truncated 6.52 MB...]
org.apache.kafka.streams.scala.kstream.JoinedTest > Create a Joined should create a Joined with Serdes and repartition topic name STARTED
org.apache.kafka.streams.scala.kstream.JoinedTest > Create a Joined should create a Joined with Serdes and repartition topic name PASSED
org.apache.kafka.streams.scala.kstream.GroupedTest > Create a Grouped should create a Grouped with Serdes STARTED
org.apache.kafka.streams.scala.kstream.GroupedTest > Create a Grouped should create a Grouped with Serdes PASSED
org.apache.kafka.streams.scala.kstream.GroupedTest > Create a Grouped with repartition topic name should create a Grouped with Serdes, and repartition topic name STARTED
org.apache.kafka.streams.scala.kstream.GroupedTest > Create a Grouped with repartition topic name should create a Grouped with Serdes, and repartition topic name PASSED
org.apache.kafka.streams.scala.kstream.SuppressedTest > Suppressed.untilWindowCloses should produce the correct suppression STARTED
org.apache.kafka.streams.scala.kstream.SuppressedTest > Suppressed.untilWindowCloses should produce the correct suppression PASSED
org.apache.kafka.streams.scala.kstream.SuppressedTest > Suppressed.untilTimeLimit should produce the correct suppression STARTED
org.apache.kafka.streams.scala.kstream.SuppressedTest > Suppressed.untilTimeLimit should produce the correct suppression PASSED
org.apache.kafka.streams.scala.kstream.SuppressedTest > BufferConfig.maxRecords should produce the correct buffer config STARTED
org.apache.kafka.streams.scala.kstream.SuppressedTest > BufferConfig.maxRecords should produce the correct buffer config PASSED
org.apache.kafka.streams.scala.kstream.SuppressedTest > BufferConfig.maxBytes should produce the correct buffer config STARTED
org.apache.kafka.streams.scala.kstream.SuppressedTest > BufferConfig.maxBytes should produce the correct buffer config PASSED
org.apache.kafka.streams.scala.kstream.SuppressedTest > BufferConfig.unbounded should produce the correct buffer config STARTED
org.apache.kafka.streams.scala.kstream.SuppressedTest > BufferConfig.unbounded should produce the correct buffer config PASSED
org.apache.kafka.streams.scala.kstream.SuppressedTest > BufferConfig should support very long chains of factory methods STARTED
org.apache.kafka.streams.scala.kstream.SuppressedTest > BufferConfig should support very long chains of factory methods PASSED
org.apache.kafka.streams.scala.kstream.ProducedTest > Create a Produced should create a Produced with Serdes STARTED
org.apache.kafka.streams.scala.kstream.ProducedTest > Create a Produced should create a Produced with Serdes PASSED
org.apache.kafka.streams.scala.kstream.ProducedTest > Create a Produced with timestampExtractor and resetPolicy should create a Consumed with Serdes, timestampExtractor and resetPolicy STARTED
org.apache.kafka.streams.scala.kstream.ProducedTest > Create a Produced with timestampExtractor and resetPolicy should create a Consumed with Serdes, timestampExtractor and resetPolicy PASSED
org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed should create a Consumed with Serdes STARTED
org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed should create a Consumed with Serdes PASSED
org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed with timestampExtractor and resetPolicy should create a Consumed with Serdes, timestampExtractor and resetPolicy STARTED
org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed with timestampExtractor and resetPolicy should create a Consumed with Serdes, timestampExtractor and resetPolicy PASSED
org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed with timestampExtractor should create a Consumed with Serdes and timestampExtractor STARTED
org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed with timestampExtractor should create a Consumed with Serdes and timestampExtractor PASSED
org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed with resetPolicy should create a Consumed with Serdes and resetPolicy STARTED
org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed with resetPolicy should create a Consumed with Serdes and resetPolicy PASSED
org.apache.kafka.streams.scala.kstream.KStreamTest > filter a KStream should filter records satisfying the predicate STARTED
org.apache.kafka.streams.scala.kstream.KStreamTest > filter a KStream should filter records satisfying the predicate PASSED
org.apache.kafka.streams.scala.kstream.KStreamTest > filterNot a KStream should filter records not satisfying the predicate STARTED
org.apache.kafka.streams.scala.kstream.KStreamTest > filterNot a KStream should filter records not satisfying the predicate PASSED
org.apache.kafka.streams.scala.kstream.KStreamTest > foreach a KStream should run foreach actions on records STARTED
org.apache.kafka.streams.scala.kstream.KStreamTest > foreach a KStream should run foreach actions on records PASSED
org.apache.kafka.streams.scala.kstream.KStreamTest > peek a KStream should run peek actions on records STARTED
org.apache.kafka.streams.scala.kstream.KStreamTest > peek a KStream should run peek actions on records PASSED
org.apache.kafka.streams.scala.kstream.KStreamTest > selectKey a KStream should select a new key STARTED
org.apache.kafka.streams.scala.kstream.KStreamTest > selectKey a KStream should select a new key PASSED
org.apache.kafka.streams.scala.kstream.KStreamTest > join 2 KStreams should join correctly records STARTED
org.apache.kafka.streams.scala.kstream.KStreamTest > join 2 KStreams should join correctly records PASSED
org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialized should create a Materialized with Serdes STARTED
org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialized should create a Materialized with Serdes PASSED
org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize with a store name should create a Materialized with Serdes and a store name STARTED
org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize with a store name should create a Materialized with Serdes and a store name PASSED
org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize with a window store supplier should create a Materialized with Serdes and a store supplier STARTED
org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize with a window store supplier should create a Materialized with Serdes and a store supplier PASSED
org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize with a key value store supplier should create a Materialized with Serdes and a store supplier STARTED
org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize with a key value store supplier should create a Materialized with Serdes and a store supplier PASSED
org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize with a session store supplier should create a Materialized with Serdes and a store supplier STARTED
org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize with a session store supplier should create a Materialized with Serdes and a store supplier PASSED
org.apache.kafka.streams.scala.kstream.KTableTest > filter a KTable should filter records satisfying the predicate STARTED
org.apache.kafka.streams.scala.kstream.KTableTest > filter a KTable should filter records satisfying the predicate PASSED
org.apache.kafka.streams.scala.kstream.KTableTest > filterNot a KTable should filter records not satisfying the predicate STARTED
org.apache.kafka.streams.scala.kstream.KTableTest > filterNot a KTable should filter records not satisfying the predicate PASSED
org.apache.kafka.streams.scala.kstream.KTableTest > join 2 KTables should join correctly records STARTED
org.apache.kafka.streams.scala.kstream.KTableTest > join 2 KTables should join correctly records PASSED
org.apache.kafka.streams.scala.kstream.KTableTest > join 2 KTables with a Materialized should join correctly records and state store STARTED
org.apache.kafka.streams.scala.kstream.KTableTest > join 2 KTables with a Materialized should join correctly records and state store PASSED
org.apache.kafka.streams.scala.kstream.KTableTest > windowed KTable#suppress should correctly suppress results using Suppressed.untilTimeLimit STARTED
org.apache.kafka.streams.scala.kstream.KTableTest > windowed KTable#suppress should correctly suppress results using Suppressed.untilTimeLimit PASSED
org.apache.kafka.streams.scala.kstream.KTableTest > windowed KTable#suppress should correctly suppress results using Suppressed.untilWindowCloses STARTED
org.apache.kafka.streams.scala.kstream.KTableTest > windowed KTable#suppress should correctly suppress results using Suppressed.untilWindowCloses PASSED
org.apache.kafka.streams.scala.kstream.KTableTest > session windowed KTable#suppress should correctly suppress results using Suppressed.untilWindowCloses STARTED
org.apache.kafka.streams.scala.kstream.KTableTest > session windowed KTable#suppress should correctly suppress results using Suppressed.untilWindowCloses PASSED
org.apache.kafka.streams.scala.kstream.KTableTest > non-windowed KTable#suppress should correctly suppress results using Suppressed.untilTimeLimit STARTED
org.apache.kafka.streams.scala.kstream.KTableTest > non-windowed KTable#suppress should correctly suppress results using Suppressed.untilTimeLimit PASSED
org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaJoin STARTED
org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaJoin PASSED
org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaSimple STARTED
org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaSimple PASSED
org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaAggregate STARTED
org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaAggregate PASSED
org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaProperties STARTED
org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaProperties PASSED
org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaTransform STARTED
org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaTransform PASSED
org.apache.kafka.streams.scala.StreamToTableJoinScalaIntegrationTestImplicitSerdes > testShouldCountClicksPerRegionWithNamedRepartitionTopic STARTED
org.apache.kafka.streams.scala.StreamToTableJoinScalaIntegrationTestImplicitSerdes > testShouldCountClicksPerRegionWithNamedRepartitionTopic PASSED
org.apache.kafka.streams.scala.StreamToTableJoinScalaIntegrationTestImplicitSerdes > testShouldCountClicksPerRegionJava STARTED
org.apache.kafka.streams.scala.StreamToTableJoinScalaIntegrationTestImplicitSerdes > testShouldCountClicksPerRegionJava PASSED
org.apache.kafka.streams.scala.StreamToTableJoinScalaIntegrationTestImplicitSerdes > testShouldCountClicksPerRegion STARTED
org.apache.kafka.streams.scala.StreamToTableJoinScalaIntegrationTestImplicitSerdes > testShouldCountClicksPerRegion PASSED
org.apache.kafka.streams.scala.WordCountTest > testShouldCountWordsMaterialized STARTED
org.apache.kafka.streams.scala.WordCountTest > testShouldCountWordsMaterialized PASSED
org.apache.kafka.streams.scala.WordCountTest > testShouldCountWordsJava STARTED
org.apache.kafka.streams.scala.WordCountTest > testShouldCountWordsJava PASSED
org.apache.kafka.streams.scala.WordCountTest > testShouldCountWords STARTED
org.apache.kafka.streams.scala.WordCountTest > testShouldCountWords PASSED
> Task :log4j-appender:test
org.apache.kafka.log4jappender.KafkaLog4jAppenderTest > testLog4jAppendsWithSyncSendWithoutIgnoringExceptionsShouldNotThrowException STARTED
org.apache.kafka.log4jappender.KafkaLog4jAppenderTest > testLog4jAppendsWithSyncSendWithoutIgnoringExceptionsShouldNotThrowException PASSED
org.apache.kafka.log4jappender.KafkaLog4jAppenderTest > testLog4jAppendsWithRealProducerConfigWithSyncSendAndNotIgnoringExceptionsShouldThrowException STARTED
org.apache.kafka.log4jappender.KafkaLog4jAppenderTest > testLog4jAppendsWithRealProducerConfigWithSyncSendAndNotIgnoringExceptionsShouldThrowException PASSED
org.apache.kafka.log4jappender.KafkaLog4jAppenderTest > testLog4jAppends STARTED
org.apache.kafka.log4jappender.KafkaLog4jAppenderTest > testLog4jAppends PASSED
org.apache.kafka.log4jappender.KafkaLog4jAppenderTest > testSetSaslMechanism STARTED
org.apache.kafka.log4jappender.KafkaLog4jAppenderTest > testSetSaslMechanism PASSED
org.apache.kafka.log4jappender.KafkaLog4jAppenderTest > testJaasConfigNotSet STARTED
org.apache.kafka.log4jappender.KafkaLog4jAppenderTest > testJaasConfigNotSet PASSED
org.apache.kafka.log4jappender.KafkaLog4jAppenderTest > testKafkaLog4jConfigs STARTED
org.apache.kafka.log4jappender.KafkaLog4jAppenderTest > testKafkaLog4jConfigs PASSED
org.apache.kafka.log4jappender.KafkaLog4jAppenderTest > testLog4jAppendsWithSyncSendAndSimulateProducerFailShouldThrowException STARTED
org.apache.kafka.log4jappender.KafkaLog4jAppenderTest > testLog4jAppendsWithSyncSendAndSimulateProducerFailShouldThrowException PASSED
org.apache.kafka.log4jappender.KafkaLog4jAppenderTest > testLog4jAppendsWithRealProducerConfigWithSyncSendShouldNotThrowException STARTED
org.apache.kafka.log4jappender.KafkaLog4jAppenderTest > testLog4jAppendsWithRealProducerConfigWithSyncSendShouldNotThrowException PASSED
org.apache.kafka.log4jappender.KafkaLog4jAppenderTest > testSaslMechanismNotSet STARTED
org.apache.kafka.log4jappender.KafkaLog4jAppenderTest > testSaslMechanismNotSet PASSED
org.apache.kafka.log4jappender.KafkaLog4jAppenderTest > testSetJaasConfig STARTED
org.apache.kafka.log4jappender.KafkaLog4jAppenderTest > testSetJaasConfig PASSED
FAILURE: Build failed with an exception.
* What went wrong:
Execution failed for task ':core:test'.
> There were failing tests. See the report at: file://<https://builds.apache.org/job/kafka-trunk-jdk11/ws/core/build/reports/tests/test/index.html>
* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
* Get more help at https://help.gradle.org
Deprecated Gradle features were used in this build, making it incompatible with Gradle 6.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/5.4.1/userguide/command_line_interface.html#sec:command_line_warnings
BUILD FAILED in 3h 10m 13s
178 actionable tasks: 175 executed, 3 up-to-date
Build step 'Execute shell' marked build as failure
Recording test results
ERROR: No tool found matching GRADLE_4_10_2_HOME
Setting GRADLE_4_10_3_HOME=/home/jenkins/tools/gradle/4.10.3
ERROR: No tool found matching GRADLE_4_10_2_HOME
Setting GRADLE_4_10_3_HOME=/home/jenkins/tools/gradle/4.10.3
Not sending mail to unregistered user ismael@juma.me.uk
Not sending mail to unregistered user vahid.hashemian@gmail.com
Not sending mail to unregistered user noreply@github.com
Jenkins build is back to normal : kafka-trunk-jdk11 #754
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/kafka-trunk-jdk11/754/display/redirect?page=changes>
Build failed in Jenkins: kafka-trunk-jdk11 #753
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/kafka-trunk-jdk11/753/display/redirect?page=changes>
Changes:
[bbejeck] MINOR: Correct typo in test name `TimetampedSegmentsTest` (#7210)
------------------------------------------
[...truncated 2.59 MB...]
org.apache.kafka.connect.runtime.standalone.StandaloneHerderTest > testPutTaskConfigs STARTED
org.apache.kafka.connect.runtime.standalone.StandaloneHerderTest > testPutTaskConfigs PASSED
org.apache.kafka.connect.runtime.standalone.StandaloneHerderTest > testCreateConnectorFailedBasicValidation STARTED
org.apache.kafka.connect.runtime.standalone.StandaloneHerderTest > testCreateConnectorFailedBasicValidation PASSED
org.apache.kafka.connect.runtime.standalone.StandaloneHerderTest > testCreateConnectorAlreadyExists STARTED
org.apache.kafka.connect.runtime.standalone.StandaloneHerderTest > testCreateConnectorAlreadyExists PASSED
org.apache.kafka.connect.connector.policy.PrincipalConnectorClientConfigOverridePolicyTest > testPrincipalPlusOtherConfigs STARTED
org.apache.kafka.connect.connector.policy.PrincipalConnectorClientConfigOverridePolicyTest > testPrincipalPlusOtherConfigs PASSED
org.apache.kafka.connect.connector.policy.PrincipalConnectorClientConfigOverridePolicyTest > testPrincipalOnly STARTED
org.apache.kafka.connect.connector.policy.PrincipalConnectorClientConfigOverridePolicyTest > testPrincipalOnly PASSED
org.apache.kafka.connect.connector.policy.NoneConnectorClientConfigOverridePolicyTest > testNoOverrides STARTED
org.apache.kafka.connect.connector.policy.NoneConnectorClientConfigOverridePolicyTest > testNoOverrides PASSED
org.apache.kafka.connect.connector.policy.NoneConnectorClientConfigOverridePolicyTest > testWithOverrides STARTED
org.apache.kafka.connect.connector.policy.NoneConnectorClientConfigOverridePolicyTest > testWithOverrides PASSED
> Task :streams:examples:test
org.apache.kafka.streams.examples.wordcount.WordCountProcessorTest > test STARTED
org.apache.kafka.streams.examples.wordcount.WordCountProcessorTest > test PASSED
> Task :streams:streams-scala:test
org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed should create a Consumed with Serdes STARTED
org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed should create a Consumed with Serdes PASSED
org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed with timestampExtractor and resetPolicy should create a Consumed with Serdes, timestampExtractor and resetPolicy STARTED
org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed with timestampExtractor and resetPolicy should create a Consumed with Serdes, timestampExtractor and resetPolicy PASSED
org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed with timestampExtractor should create a Consumed with Serdes and timestampExtractor STARTED
org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed with timestampExtractor should create a Consumed with Serdes and timestampExtractor PASSED
org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed with resetPolicy should create a Consumed with Serdes and resetPolicy STARTED
org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed with resetPolicy should create a Consumed with Serdes and resetPolicy PASSED
org.apache.kafka.streams.scala.kstream.KTableTest > filter a KTable should filter records satisfying the predicate STARTED
org.apache.kafka.streams.scala.kstream.KTableTest > filter a KTable should filter records satisfying the predicate PASSED
org.apache.kafka.streams.scala.kstream.KTableTest > filterNot a KTable should filter records not satisfying the predicate STARTED
org.apache.kafka.streams.scala.kstream.KTableTest > filterNot a KTable should filter records not satisfying the predicate PASSED
org.apache.kafka.streams.scala.kstream.KTableTest > join 2 KTables should join correctly records STARTED
org.apache.kafka.streams.scala.kstream.KTableTest > join 2 KTables should join correctly records PASSED
org.apache.kafka.streams.scala.kstream.KTableTest > join 2 KTables with a Materialized should join correctly records and state store STARTED
org.apache.kafka.streams.scala.kstream.KTableTest > join 2 KTables with a Materialized should join correctly records and state store PASSED
org.apache.kafka.streams.scala.kstream.KTableTest > windowed KTable#suppress should correctly suppress results using Suppressed.untilTimeLimit STARTED
org.apache.kafka.streams.scala.kstream.KTableTest > windowed KTable#suppress should correctly suppress results using Suppressed.untilTimeLimit PASSED
org.apache.kafka.streams.scala.kstream.KTableTest > windowed KTable#suppress should correctly suppress results using Suppressed.untilWindowCloses STARTED
org.apache.kafka.streams.scala.kstream.KTableTest > windowed KTable#suppress should correctly suppress results using Suppressed.untilWindowCloses PASSED
org.apache.kafka.streams.scala.kstream.KTableTest > session windowed KTable#suppress should correctly suppress results using Suppressed.untilWindowCloses STARTED
org.apache.kafka.streams.scala.kstream.KTableTest > session windowed KTable#suppress should correctly suppress results using Suppressed.untilWindowCloses PASSED
org.apache.kafka.streams.scala.kstream.KTableTest > non-windowed KTable#suppress should correctly suppress results using Suppressed.untilTimeLimit STARTED
org.apache.kafka.streams.scala.kstream.KTableTest > non-windowed KTable#suppress should correctly suppress results using Suppressed.untilTimeLimit PASSED
org.apache.kafka.streams.scala.kstream.ProducedTest > Create a Produced should create a Produced with Serdes STARTED
org.apache.kafka.streams.scala.kstream.ProducedTest > Create a Produced should create a Produced with Serdes PASSED
org.apache.kafka.streams.scala.kstream.ProducedTest > Create a Produced with timestampExtractor and resetPolicy should create a Consumed with Serdes, timestampExtractor and resetPolicy STARTED
org.apache.kafka.streams.scala.kstream.ProducedTest > Create a Produced with timestampExtractor and resetPolicy should create a Consumed with Serdes, timestampExtractor and resetPolicy PASSED
org.apache.kafka.streams.scala.kstream.SuppressedTest > Suppressed.untilWindowCloses should produce the correct suppression STARTED
org.apache.kafka.streams.scala.kstream.SuppressedTest > Suppressed.untilWindowCloses should produce the correct suppression PASSED
org.apache.kafka.streams.scala.kstream.SuppressedTest > Suppressed.untilTimeLimit should produce the correct suppression STARTED
org.apache.kafka.streams.scala.kstream.SuppressedTest > Suppressed.untilTimeLimit should produce the correct suppression PASSED
org.apache.kafka.streams.scala.kstream.SuppressedTest > BufferConfig.maxRecords should produce the correct buffer config STARTED
org.apache.kafka.streams.scala.kstream.SuppressedTest > BufferConfig.maxRecords should produce the correct buffer config PASSED
org.apache.kafka.streams.scala.kstream.SuppressedTest > BufferConfig.maxBytes should produce the correct buffer config STARTED
org.apache.kafka.streams.scala.kstream.SuppressedTest > BufferConfig.maxBytes should produce the correct buffer config PASSED
org.apache.kafka.streams.scala.kstream.SuppressedTest > BufferConfig.unbounded should produce the correct buffer config STARTED
org.apache.kafka.streams.scala.kstream.SuppressedTest > BufferConfig.unbounded should produce the correct buffer config PASSED
org.apache.kafka.streams.scala.kstream.SuppressedTest > BufferConfig should support very long chains of factory methods STARTED
org.apache.kafka.streams.scala.kstream.SuppressedTest > BufferConfig should support very long chains of factory methods PASSED
org.apache.kafka.streams.scala.kstream.GroupedTest > Create a Grouped should create a Grouped with Serdes STARTED
org.apache.kafka.streams.scala.kstream.GroupedTest > Create a Grouped should create a Grouped with Serdes PASSED
org.apache.kafka.streams.scala.kstream.GroupedTest > Create a Grouped with repartition topic name should create a Grouped with Serdes, and repartition topic name STARTED
org.apache.kafka.streams.scala.kstream.GroupedTest > Create a Grouped with repartition topic name should create a Grouped with Serdes, and repartition topic name PASSED
org.apache.kafka.streams.scala.kstream.JoinedTest > Create a Joined should create a Joined with Serdes STARTED
org.apache.kafka.streams.scala.kstream.JoinedTest > Create a Joined should create a Joined with Serdes PASSED
org.apache.kafka.streams.scala.kstream.JoinedTest > Create a Joined should create a Joined with Serdes and repartition topic name STARTED
org.apache.kafka.streams.scala.kstream.JoinedTest > Create a Joined should create a Joined with Serdes and repartition topic name PASSED
org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialized should create a Materialized with Serdes STARTED
org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialized should create a Materialized with Serdes PASSED
org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize with a store name should create a Materialized with Serdes and a store name STARTED
org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize with a store name should create a Materialized with Serdes and a store name PASSED
org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize with a window store supplier should create a Materialized with Serdes and a store supplier STARTED
org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize with a window store supplier should create a Materialized with Serdes and a store supplier PASSED
org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize with a key value store supplier should create a Materialized with Serdes and a store supplier STARTED
org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize with a key value store supplier should create a Materialized with Serdes and a store supplier PASSED
org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize with a session store supplier should create a Materialized with Serdes and a store supplier STARTED
org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize with a session store supplier should create a Materialized with Serdes and a store supplier PASSED
org.apache.kafka.streams.scala.kstream.KStreamTest > filter a KStream should filter records satisfying the predicate STARTED
org.apache.kafka.streams.scala.kstream.KStreamTest > filter a KStream should filter records satisfying the predicate PASSED
org.apache.kafka.streams.scala.kstream.KStreamTest > filterNot a KStream should filter records not satisfying the predicate STARTED
org.apache.kafka.streams.scala.kstream.KStreamTest > filterNot a KStream should filter records not satisfying the predicate PASSED
org.apache.kafka.streams.scala.kstream.KStreamTest > foreach a KStream should run foreach actions on records STARTED
org.apache.kafka.streams.scala.kstream.KStreamTest > foreach a KStream should run foreach actions on records PASSED
org.apache.kafka.streams.scala.kstream.KStreamTest > peek a KStream should run peek actions on records STARTED
org.apache.kafka.streams.scala.kstream.KStreamTest > peek a KStream should run peek actions on records PASSED
org.apache.kafka.streams.scala.kstream.KStreamTest > selectKey a KStream should select a new key STARTED
org.apache.kafka.streams.scala.kstream.KStreamTest > selectKey a KStream should select a new key PASSED
org.apache.kafka.streams.scala.kstream.KStreamTest > join 2 KStreams should join correctly records STARTED
org.apache.kafka.streams.scala.kstream.KStreamTest > join 2 KStreams should join correctly records PASSED
org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaJoin STARTED
org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaJoin PASSED
org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaSimple STARTED
org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaSimple PASSED
org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaAggregate STARTED
org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaAggregate PASSED
org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaProperties STARTED
org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaProperties PASSED
org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaTransform STARTED
org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaTransform PASSED
org.apache.kafka.streams.scala.StreamToTableJoinScalaIntegrationTestImplicitSerdes > testShouldCountClicksPerRegionWithNamedRepartitionTopic STARTED
org.apache.kafka.streams.scala.StreamToTableJoinScalaIntegrationTestImplicitSerdes > testShouldCountClicksPerRegionWithNamedRepartitionTopic PASSED
org.apache.kafka.streams.scala.StreamToTableJoinScalaIntegrationTestImplicitSerdes > testShouldCountClicksPerRegionJava STARTED
org.apache.kafka.streams.scala.StreamToTableJoinScalaIntegrationTestImplicitSerdes > testShouldCountClicksPerRegionJava PASSED
org.apache.kafka.streams.scala.StreamToTableJoinScalaIntegrationTestImplicitSerdes > testShouldCountClicksPerRegion STARTED
org.apache.kafka.streams.scala.StreamToTableJoinScalaIntegrationTestImplicitSerdes > testShouldCountClicksPerRegion PASSED
org.apache.kafka.streams.scala.WordCountTest > testShouldCountWordsMaterialized STARTED
org.apache.kafka.streams.scala.WordCountTest > testShouldCountWordsMaterialized PASSED
org.apache.kafka.streams.scala.WordCountTest > testShouldCountWordsJava STARTED
org.apache.kafka.streams.scala.WordCountTest > testShouldCountWordsJava PASSED
org.apache.kafka.streams.scala.WordCountTest > testShouldCountWords STARTED
org.apache.kafka.streams.scala.WordCountTest > testShouldCountWords PASSED
FAILURE: Build failed with an exception.
* What went wrong:
Execution failed for task ':core:test'.
> There were failing tests. See the report at: file://<https://builds.apache.org/job/kafka-trunk-jdk11/ws/core/build/reports/tests/test/index.html>
* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
* Get more help at https://help.gradle.org
Deprecated Gradle features were used in this build, making it incompatible with Gradle 6.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/5.4.1/userguide/command_line_interface.html#sec:command_line_warnings
BUILD FAILED in 3h 49m 4s
178 actionable tasks: 152 executed, 26 up-to-date
Build step 'Execute shell' marked build as failure
Recording test results
ERROR: No tool found matching GRADLE_4_10_2_HOME
Setting GRADLE_4_10_3_HOME=/home/jenkins/tools/gradle/4.10.3
ERROR: No tool found matching GRADLE_4_10_2_HOME
Setting GRADLE_4_10_3_HOME=/home/jenkins/tools/gradle/4.10.3
Not sending mail to unregistered user ismael@juma.me.uk
Not sending mail to unregistered user vahid.hashemian@gmail.com
Not sending mail to unregistered user noreply@github.com
Not sending mail to unregistered user wangguoz@gmail.com
Not sending mail to unregistered user gwen@confluent.io
Build failed in Jenkins: kafka-trunk-jdk11 #752
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/kafka-trunk-jdk11/752/display/redirect?page=changes>
Changes:
[wangguoz] MINOR: remove unnecessary #remove overrides (#7178)
[github] KAFKA-8791: RocksDBTimestampedStore should open in regular mode by
[ismael] MINOR: Avoid unnecessary leaderFor calls when ProducerBatch queue empty
[github] KAFKA-8765: Remove interface annotations in Streams API (#7174)
[gwen] KAFKA-8792; Default ZK configuration to disable AdminServer
------------------------------------------
[...truncated 2.63 MB...]
org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperatorTest > testThrowExceptionInKafkaConsume STARTED
org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperatorTest > testThrowExceptionInKafkaConsume PASSED
org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperatorTest > testThrowExceptionInKafkaProduce STARTED
org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperatorTest > testThrowExceptionInKafkaProduce PASSED
org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperatorTest > testExecAndHandleRetriableErrorOnce STARTED
org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperatorTest > testExecAndHandleRetriableErrorOnce PASSED
org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperatorTest > testExecAndHandleRetriableErrorThrice STARTED
org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperatorTest > testExecAndHandleRetriableErrorThrice PASSED
org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperatorTest > testExecAndHandleNonRetriableErrorOnce STARTED
org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperatorTest > testExecAndHandleNonRetriableErrorOnce PASSED
org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperatorTest > testExecAndHandleNonRetriableErrorThrice STARTED
org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperatorTest > testExecAndHandleNonRetriableErrorThrice PASSED
org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperatorTest > testCheckRetryLimit STARTED
org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperatorTest > testCheckRetryLimit PASSED
org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperatorTest > testBackoffLimit STARTED
org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperatorTest > testBackoffLimit PASSED
org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperatorTest > testToleranceLimit STARTED
org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperatorTest > testToleranceLimit PASSED
org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperatorTest > testDefaultConfigs STARTED
org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperatorTest > testDefaultConfigs PASSED
org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperatorTest > testSetConfigs STARTED
org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperatorTest > testSetConfigs PASSED
org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperatorTest > testHandleExceptionInTransformations STARTED
org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperatorTest > testHandleExceptionInTransformations PASSED
org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperatorTest > testHandleExceptionInHeaderConverter STARTED
org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperatorTest > testHandleExceptionInHeaderConverter PASSED
org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperatorTest > testHandleExceptionInValueConverter STARTED
org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperatorTest > testHandleExceptionInValueConverter PASSED
org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperatorTest > testHandleExceptionInTaskPut STARTED
org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperatorTest > testHandleExceptionInTaskPut PASSED
org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperatorTest > testHandleExceptionInTaskPoll STARTED
org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperatorTest > testHandleExceptionInTaskPoll PASSED
org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperatorTest > testThrowExceptionInTaskPut STARTED
org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperatorTest > testThrowExceptionInTaskPut PASSED
org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperatorTest > testThrowExceptionInTaskPoll STARTED
org.apache.kafka.connect.runtime.errors.RetryWithToleranceOperatorTest > testThrowExceptionInTaskPoll PASSED
org.apache.kafka.connect.runtime.errors.ErrorReporterTest > initializeDLQWithNullMetrics STARTED
org.apache.kafka.connect.runtime.errors.ErrorReporterTest > initializeDLQWithNullMetrics PASSED
org.apache.kafka.connect.runtime.errors.ErrorReporterTest > testDLQConfigWithEmptyTopicName STARTED
org.apache.kafka.connect.runtime.errors.ErrorReporterTest > testDLQConfigWithEmptyTopicName PASSED
org.apache.kafka.connect.runtime.errors.ErrorReporterTest > testDLQConfigWithValidTopicName STARTED
org.apache.kafka.connect.runtime.errors.ErrorReporterTest > testDLQConfigWithValidTopicName PASSED
org.apache.kafka.connect.runtime.errors.ErrorReporterTest > testReportDLQTwice STARTED
org.apache.kafka.connect.runtime.errors.ErrorReporterTest > testReportDLQTwice PASSED
org.apache.kafka.connect.runtime.errors.ErrorReporterTest > testLogOnDisabledLogReporter STARTED
org.apache.kafka.connect.runtime.errors.ErrorReporterTest > testLogOnDisabledLogReporter PASSED
org.apache.kafka.connect.runtime.errors.ErrorReporterTest > testLogOnEnabledLogReporter STARTED
org.apache.kafka.connect.runtime.errors.ErrorReporterTest > testLogOnEnabledLogReporter PASSED
org.apache.kafka.connect.runtime.errors.ErrorReporterTest > testLogMessageWithNoRecords STARTED
org.apache.kafka.connect.runtime.errors.ErrorReporterTest > testLogMessageWithNoRecords PASSED
org.apache.kafka.connect.runtime.errors.ErrorReporterTest > testLogMessageWithSinkRecords STARTED
org.apache.kafka.connect.runtime.errors.ErrorReporterTest > testLogMessageWithSinkRecords PASSED
org.apache.kafka.connect.runtime.errors.ErrorReporterTest > testSetDLQConfigs STARTED
org.apache.kafka.connect.runtime.errors.ErrorReporterTest > testSetDLQConfigs PASSED
org.apache.kafka.connect.runtime.errors.ErrorReporterTest > testDlqHeaderConsumerRecord STARTED
org.apache.kafka.connect.runtime.errors.ErrorReporterTest > testDlqHeaderConsumerRecord PASSED
org.apache.kafka.connect.runtime.errors.ErrorReporterTest > testDlqHeaderOnNullExceptionMessage STARTED
org.apache.kafka.connect.runtime.errors.ErrorReporterTest > testDlqHeaderOnNullExceptionMessage PASSED
org.apache.kafka.connect.runtime.errors.ErrorReporterTest > testDlqHeaderIsAppended STARTED
org.apache.kafka.connect.runtime.errors.ErrorReporterTest > testDlqHeaderIsAppended PASSED
org.apache.kafka.connect.runtime.WorkerTaskTest > standardStartup STARTED
org.apache.kafka.connect.runtime.WorkerTaskTest > standardStartup PASSED
org.apache.kafka.connect.runtime.WorkerTaskTest > stopBeforeStarting STARTED
org.apache.kafka.connect.runtime.WorkerTaskTest > stopBeforeStarting PASSED
org.apache.kafka.connect.runtime.WorkerTaskTest > cancelBeforeStopping STARTED
org.apache.kafka.connect.runtime.WorkerTaskTest > cancelBeforeStopping PASSED
org.apache.kafka.connect.runtime.WorkerTaskTest > updateMetricsOnListenerEventsForStartupPauseResumeAndShutdown STARTED
org.apache.kafka.connect.runtime.WorkerTaskTest > updateMetricsOnListenerEventsForStartupPauseResumeAndShutdown PASSED
org.apache.kafka.connect.runtime.WorkerTaskTest > updateMetricsOnListenerEventsForStartupPauseResumeAndFailure STARTED
org.apache.kafka.connect.runtime.WorkerTaskTest > updateMetricsOnListenerEventsForStartupPauseResumeAndFailure PASSED
org.apache.kafka.connect.runtime.WorkerConfigTransformerTest > testReplaceVariable STARTED
org.apache.kafka.connect.runtime.WorkerConfigTransformerTest > testReplaceVariable PASSED
org.apache.kafka.connect.runtime.WorkerConfigTransformerTest > testReplaceVariableWithTTL STARTED
org.apache.kafka.connect.runtime.WorkerConfigTransformerTest > testReplaceVariableWithTTL PASSED
org.apache.kafka.connect.runtime.WorkerConfigTransformerTest > testReplaceVariableWithTTLAndScheduleRestart STARTED
org.apache.kafka.connect.runtime.WorkerConfigTransformerTest > testReplaceVariableWithTTLAndScheduleRestart PASSED
org.apache.kafka.connect.runtime.WorkerConfigTransformerTest > testReplaceVariableWithTTLFirstCancelThenScheduleRestart STARTED
org.apache.kafka.connect.runtime.WorkerConfigTransformerTest > testReplaceVariableWithTTLFirstCancelThenScheduleRestart PASSED
org.apache.kafka.connect.runtime.WorkerConfigTransformerTest > testTransformNullConfiguration STARTED
org.apache.kafka.connect.runtime.WorkerConfigTransformerTest > testTransformNullConfiguration PASSED
org.apache.kafka.connect.runtime.WorkerSourceTaskTest > testStartPaused STARTED
org.apache.kafka.connect.runtime.WorkerSourceTaskTest > testStartPaused PASSED
org.apache.kafka.connect.runtime.WorkerSourceTaskTest > testPause STARTED
org.apache.kafka.connect.runtime.WorkerSourceTaskTest > testPause PASSED
org.apache.kafka.connect.runtime.WorkerSourceTaskTest > testPollsInBackground STARTED
org.apache.kafka.connect.runtime.WorkerSourceTaskTest > testPollsInBackground PASSED
org.apache.kafka.connect.runtime.WorkerSourceTaskTest > testCommit STARTED
org.apache.kafka.connect.runtime.WorkerSourceTaskTest > testCommit PASSED
org.apache.kafka.connect.runtime.WorkerSourceTaskTest > testCommitFailure STARTED
org.apache.kafka.connect.runtime.WorkerSourceTaskTest > testCommitFailure PASSED
org.apache.kafka.connect.runtime.WorkerSourceTaskTest > testMetricsGroup STARTED
org.apache.kafka.connect.runtime.WorkerSourceTaskTest > testMetricsGroup PASSED
org.apache.kafka.connect.runtime.WorkerSourceTaskTest > testSendRecordsTaskCommitRecordFail STARTED
org.apache.kafka.connect.runtime.WorkerSourceTaskTest > testSendRecordsTaskCommitRecordFail PASSED
org.apache.kafka.connect.runtime.WorkerSourceTaskTest > testSlowTaskStart STARTED
org.apache.kafka.connect.runtime.WorkerSourceTaskTest > testSlowTaskStart PASSED
org.apache.kafka.connect.runtime.WorkerSourceTaskTest > testSendRecordsPropagatesTimestamp STARTED
org.apache.kafka.connect.runtime.WorkerSourceTaskTest > testSendRecordsPropagatesTimestamp PASSED
org.apache.kafka.connect.runtime.WorkerSourceTaskTest > testSendRecordsCorruptTimestamp STARTED
org.apache.kafka.connect.runtime.WorkerSourceTaskTest > testSendRecordsCorruptTimestamp PASSED
org.apache.kafka.connect.runtime.WorkerSourceTaskTest > testSendRecordsNoTimestamp STARTED
org.apache.kafka.connect.runtime.WorkerSourceTaskTest > testSendRecordsNoTimestamp PASSED
org.apache.kafka.connect.runtime.WorkerSourceTaskTest > testSendRecordsRetries STARTED
org.apache.kafka.connect.runtime.WorkerSourceTaskTest > testSendRecordsRetries PASSED
org.apache.kafka.connect.runtime.WorkerSourceTaskTest > testFailureInPoll STARTED
org.apache.kafka.connect.runtime.WorkerSourceTaskTest > testFailureInPoll PASSED
org.apache.kafka.connect.runtime.WorkerSourceTaskTest > testSendRecordsConvertsData STARTED
org.apache.kafka.connect.runtime.WorkerSourceTaskTest > testSendRecordsConvertsData PASSED
481 tests completed, 1 failed
> Task :connect:runtime:test FAILED
FAILURE: Build failed with an exception.
* What went wrong:
Execution failed for task ':connect:runtime:test'.
> There were failing tests. See the report at: file://<https://builds.apache.org/job/kafka-trunk-jdk11/ws/connect/runtime/build/reports/tests/test/index.html>
* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
* Get more help at https://help.gradle.org
Deprecated Gradle features were used in this build, making it incompatible with Gradle 6.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/5.4.1/userguide/command_line_interface.html#sec:command_line_warnings
BUILD FAILED in 3h 42m 26s
178 actionable tasks: 152 executed, 26 up-to-date
Build step 'Execute shell' marked build as failure
Recording test results
ERROR: No tool found matching GRADLE_4_10_2_HOME
Setting GRADLE_4_10_3_HOME=/home/jenkins/tools/gradle/4.10.3
ERROR: No tool found matching GRADLE_4_10_2_HOME
Setting GRADLE_4_10_3_HOME=/home/jenkins/tools/gradle/4.10.3
Not sending mail to unregistered user ismael@juma.me.uk
Not sending mail to unregistered user vahid.hashemian@gmail.com
Not sending mail to unregistered user noreply@github.com
Not sending mail to unregistered user wangguoz@gmail.com
Not sending mail to unregistered user gwen@confluent.io
Build failed in Jenkins: kafka-trunk-jdk11 #751
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/kafka-trunk-jdk11/751/display/redirect?page=changes>
Changes:
[gwen] KAFKA-8391; Improved the Connect integration tests to make them less
[jason] MINOR: Add fetch from follower system test (#7166)
[bbejeck] KAFKA-8736: Track size in InMemoryKeyValueStore (#7177)
------------------------------------------
[...truncated 6.52 MB...]
org.apache.kafka.streams.TopologyTestDriverTest > shouldPopulateGlobalStore[Eos enabled = true] PASSED
org.apache.kafka.streams.TopologyTestDriverTest > shouldThrowIfPersistentBuiltInStoreIsAccessedWithUntypedMethod[Eos enabled = true] STARTED
org.apache.kafka.streams.TopologyTestDriverTest > shouldThrowIfPersistentBuiltInStoreIsAccessedWithUntypedMethod[Eos enabled = true] PASSED
org.apache.kafka.streams.TopologyTestDriverTest > shouldAllowPrePopulatingStatesStoresWithCachingEnabled[Eos enabled = true] STARTED
org.apache.kafka.streams.TopologyTestDriverTest > shouldAllowPrePopulatingStatesStoresWithCachingEnabled[Eos enabled = true] PASSED
org.apache.kafka.streams.TopologyTestDriverTest > shouldReturnCorrectPersistentStoreTypeOnly[Eos enabled = true] STARTED
org.apache.kafka.streams.TopologyTestDriverTest > shouldReturnCorrectPersistentStoreTypeOnly[Eos enabled = true] PASSED
org.apache.kafka.streams.TopologyTestDriverTest > shouldUseSourceSpecificDeserializers[Eos enabled = true] STARTED
org.apache.kafka.streams.TopologyTestDriverTest > shouldUseSourceSpecificDeserializers[Eos enabled = true] PASSED
org.apache.kafka.streams.TopologyTestDriverTest > shouldReturnAllStores[Eos enabled = true] STARTED
org.apache.kafka.streams.TopologyTestDriverTest > shouldReturnAllStores[Eos enabled = true] PASSED
org.apache.kafka.streams.TopologyTestDriverTest > shouldSendRecordViaCorrectSourceTopic[Eos enabled = true] STARTED
org.apache.kafka.streams.TopologyTestDriverTest > shouldSendRecordViaCorrectSourceTopic[Eos enabled = true] PASSED
org.apache.kafka.streams.TopologyTestDriverTest > shouldNotCreateStateDirectoryForStatelessTopology[Eos enabled = true] STARTED
org.apache.kafka.streams.TopologyTestDriverTest > shouldNotCreateStateDirectoryForStatelessTopology[Eos enabled = true] PASSED
org.apache.kafka.streams.TopologyTestDriverTest > shouldReturnAllStoresNames[Eos enabled = true] STARTED
org.apache.kafka.streams.TopologyTestDriverTest > shouldReturnAllStoresNames[Eos enabled = true] PASSED
org.apache.kafka.streams.TopologyTestDriverTest > shouldProcessConsumerRecordList[Eos enabled = true] STARTED
org.apache.kafka.streams.TopologyTestDriverTest > shouldProcessConsumerRecordList[Eos enabled = true] PASSED
org.apache.kafka.streams.TopologyTestDriverTest > shouldUseSinkSpecificSerializers[Eos enabled = true] STARTED
org.apache.kafka.streams.TopologyTestDriverTest > shouldUseSinkSpecificSerializers[Eos enabled = true] PASSED
org.apache.kafka.streams.TopologyTestDriverTest > shouldFlushStoreForFirstInput[Eos enabled = true] STARTED
org.apache.kafka.streams.TopologyTestDriverTest > shouldFlushStoreForFirstInput[Eos enabled = true] PASSED
org.apache.kafka.streams.TopologyTestDriverTest > shouldProcessFromSourceThatMatchPattern[Eos enabled = true] STARTED
org.apache.kafka.streams.TopologyTestDriverTest > shouldProcessFromSourceThatMatchPattern[Eos enabled = true] PASSED
org.apache.kafka.streams.TopologyTestDriverTest > shouldUpdateStoreForNewKey[Eos enabled = true] STARTED
org.apache.kafka.streams.TopologyTestDriverTest > shouldUpdateStoreForNewKey[Eos enabled = true] PASSED
org.apache.kafka.streams.TopologyTestDriverTest > shouldPunctuateOnWallClockTime[Eos enabled = true] STARTED
org.apache.kafka.streams.TopologyTestDriverTest > shouldPunctuateOnWallClockTime[Eos enabled = true] PASSED
org.apache.kafka.streams.TopologyTestDriverTest > shouldSetRecordMetadata[Eos enabled = true] STARTED
org.apache.kafka.streams.TopologyTestDriverTest > shouldSetRecordMetadata[Eos enabled = true] PASSED
org.apache.kafka.streams.TopologyTestDriverTest > shouldNotUpdateStoreForLargerValue[Eos enabled = true] STARTED
org.apache.kafka.streams.TopologyTestDriverTest > shouldNotUpdateStoreForLargerValue[Eos enabled = true] PASSED
org.apache.kafka.streams.TopologyTestDriverTest > shouldReturnCorrectInMemoryStoreTypeOnly[Eos enabled = true] STARTED
org.apache.kafka.streams.TopologyTestDriverTest > shouldReturnCorrectInMemoryStoreTypeOnly[Eos enabled = true] PASSED
org.apache.kafka.streams.TopologyTestDriverTest > shouldProcessRecordForTopic[Eos enabled = true] STARTED
org.apache.kafka.streams.TopologyTestDriverTest > shouldProcessRecordForTopic[Eos enabled = true] PASSED
org.apache.kafka.streams.TopologyTestDriverTest > shouldForwardRecordsFromSubtopologyToSubtopology[Eos enabled = true] STARTED
org.apache.kafka.streams.TopologyTestDriverTest > shouldForwardRecordsFromSubtopologyToSubtopology[Eos enabled = true] PASSED
org.apache.kafka.streams.TopologyTestDriverTest > shouldNotUpdateStoreForSmallerValue[Eos enabled = true] STARTED
org.apache.kafka.streams.TopologyTestDriverTest > shouldNotUpdateStoreForSmallerValue[Eos enabled = true] PASSED
org.apache.kafka.streams.TopologyTestDriverTest > shouldCreateStateDirectoryForStatefulTopology[Eos enabled = true] STARTED
org.apache.kafka.streams.TopologyTestDriverTest > shouldCreateStateDirectoryForStatefulTopology[Eos enabled = true] PASSED
org.apache.kafka.streams.TopologyTestDriverTest > shouldPunctuateIfWallClockTimeAdvances[Eos enabled = true] STARTED
org.apache.kafka.streams.TopologyTestDriverTest > shouldPunctuateIfWallClockTimeAdvances[Eos enabled = true] PASSED
org.apache.kafka.streams.TopologyTestDriverTest > shouldCloseProcessor[Eos enabled = false] STARTED
org.apache.kafka.streams.TopologyTestDriverTest > shouldCloseProcessor[Eos enabled = false] PASSED
org.apache.kafka.streams.TopologyTestDriverTest > shouldFeedStoreFromGlobalKTable[Eos enabled = false] STARTED
org.apache.kafka.streams.TopologyTestDriverTest > shouldFeedStoreFromGlobalKTable[Eos enabled = false] PASSED
org.apache.kafka.streams.TopologyTestDriverTest > shouldCleanUpPersistentStateStoresOnClose[Eos enabled = false] STARTED
org.apache.kafka.streams.TopologyTestDriverTest > shouldCleanUpPersistentStateStoresOnClose[Eos enabled = false] PASSED
org.apache.kafka.streams.TopologyTestDriverTest > shouldThrowPatternNotValidForTopicNameException[Eos enabled = false] STARTED
org.apache.kafka.streams.TopologyTestDriverTest > shouldThrowPatternNotValidForTopicNameException[Eos enabled = false] PASSED
org.apache.kafka.streams.TopologyTestDriverTest > shouldPunctuateIfEvenTimeAdvances[Eos enabled = false] STARTED
org.apache.kafka.streams.TopologyTestDriverTest > shouldPunctuateIfEvenTimeAdvances[Eos enabled = false] PASSED
org.apache.kafka.streams.TopologyTestDriverTest > shouldInitProcessor[Eos enabled = false] STARTED
org.apache.kafka.streams.TopologyTestDriverTest > shouldInitProcessor[Eos enabled = false] PASSED
org.apache.kafka.streams.TopologyTestDriverTest > shouldThrowForUnknownTopic[Eos enabled = false] STARTED
org.apache.kafka.streams.TopologyTestDriverTest > shouldThrowForUnknownTopic[Eos enabled = false] PASSED
org.apache.kafka.streams.TopologyTestDriverTest > shouldPunctuateOnStreamsTime[Eos enabled = false] STARTED
org.apache.kafka.streams.TopologyTestDriverTest > shouldPunctuateOnStreamsTime[Eos enabled = false] PASSED
org.apache.kafka.streams.TopologyTestDriverTest > shouldThrowIfInMemoryBuiltInStoreIsAccessedWithUntypedMethod[Eos enabled = false] STARTED
org.apache.kafka.streams.TopologyTestDriverTest > shouldThrowIfInMemoryBuiltInStoreIsAccessedWithUntypedMethod[Eos enabled = false] PASSED
org.apache.kafka.streams.TopologyTestDriverTest > shouldProcessFromSourcesThatMatchMultiplePattern[Eos enabled = false] STARTED
org.apache.kafka.streams.TopologyTestDriverTest > shouldProcessFromSourcesThatMatchMultiplePattern[Eos enabled = false] PASSED
org.apache.kafka.streams.TopologyTestDriverTest > shouldPopulateGlobalStore[Eos enabled = false] STARTED
org.apache.kafka.streams.TopologyTestDriverTest > shouldPopulateGlobalStore[Eos enabled = false] PASSED
org.apache.kafka.streams.TopologyTestDriverTest > shouldThrowIfPersistentBuiltInStoreIsAccessedWithUntypedMethod[Eos enabled = false] STARTED
org.apache.kafka.streams.TopologyTestDriverTest > shouldThrowIfPersistentBuiltInStoreIsAccessedWithUntypedMethod[Eos enabled = false] PASSED
org.apache.kafka.streams.TopologyTestDriverTest > shouldAllowPrePopulatingStatesStoresWithCachingEnabled[Eos enabled = false] STARTED
org.apache.kafka.streams.TopologyTestDriverTest > shouldAllowPrePopulatingStatesStoresWithCachingEnabled[Eos enabled = false] PASSED
org.apache.kafka.streams.TopologyTestDriverTest > shouldReturnCorrectPersistentStoreTypeOnly[Eos enabled = false] STARTED
org.apache.kafka.streams.TopologyTestDriverTest > shouldReturnCorrectPersistentStoreTypeOnly[Eos enabled = false] PASSED
org.apache.kafka.streams.TopologyTestDriverTest > shouldUseSourceSpecificDeserializers[Eos enabled = false] STARTED
org.apache.kafka.streams.TopologyTestDriverTest > shouldUseSourceSpecificDeserializers[Eos enabled = false] PASSED
org.apache.kafka.streams.TopologyTestDriverTest > shouldReturnAllStores[Eos enabled = false] STARTED
org.apache.kafka.streams.TopologyTestDriverTest > shouldReturnAllStores[Eos enabled = false] PASSED
org.apache.kafka.streams.TopologyTestDriverTest > shouldSendRecordViaCorrectSourceTopic[Eos enabled = false] STARTED
org.apache.kafka.streams.TopologyTestDriverTest > shouldSendRecordViaCorrectSourceTopic[Eos enabled = false] PASSED
org.apache.kafka.streams.TopologyTestDriverTest > shouldNotCreateStateDirectoryForStatelessTopology[Eos enabled = false] STARTED
org.apache.kafka.streams.TopologyTestDriverTest > shouldNotCreateStateDirectoryForStatelessTopology[Eos enabled = false] PASSED
org.apache.kafka.streams.TopologyTestDriverTest > shouldReturnAllStoresNames[Eos enabled = false] STARTED
org.apache.kafka.streams.TopologyTestDriverTest > shouldReturnAllStoresNames[Eos enabled = false] PASSED
org.apache.kafka.streams.TopologyTestDriverTest > shouldProcessConsumerRecordList[Eos enabled = false] STARTED
org.apache.kafka.streams.TopologyTestDriverTest > shouldProcessConsumerRecordList[Eos enabled = false] PASSED
org.apache.kafka.streams.TopologyTestDriverTest > shouldUseSinkSpecificSerializers[Eos enabled = false] STARTED
org.apache.kafka.streams.TopologyTestDriverTest > shouldUseSinkSpecificSerializers[Eos enabled = false] PASSED
org.apache.kafka.streams.TopologyTestDriverTest > shouldFlushStoreForFirstInput[Eos enabled = false] STARTED
org.apache.kafka.streams.TopologyTestDriverTest > shouldFlushStoreForFirstInput[Eos enabled = false] PASSED
org.apache.kafka.streams.TopologyTestDriverTest > shouldProcessFromSourceThatMatchPattern[Eos enabled = false] STARTED
org.apache.kafka.streams.TopologyTestDriverTest > shouldProcessFromSourceThatMatchPattern[Eos enabled = false] PASSED
org.apache.kafka.streams.TopologyTestDriverTest > shouldUpdateStoreForNewKey[Eos enabled = false] STARTED
org.apache.kafka.streams.TopologyTestDriverTest > shouldUpdateStoreForNewKey[Eos enabled = false] PASSED
org.apache.kafka.streams.TopologyTestDriverTest > shouldPunctuateOnWallClockTime[Eos enabled = false] STARTED
org.apache.kafka.streams.TopologyTestDriverTest > shouldPunctuateOnWallClockTime[Eos enabled = false] PASSED
org.apache.kafka.streams.TopologyTestDriverTest > shouldSetRecordMetadata[Eos enabled = false] STARTED
org.apache.kafka.streams.TopologyTestDriverTest > shouldSetRecordMetadata[Eos enabled = false] PASSED
org.apache.kafka.streams.TopologyTestDriverTest > shouldNotUpdateStoreForLargerValue[Eos enabled = false] STARTED
org.apache.kafka.streams.TopologyTestDriverTest > shouldNotUpdateStoreForLargerValue[Eos enabled = false] PASSED
org.apache.kafka.streams.TopologyTestDriverTest > shouldReturnCorrectInMemoryStoreTypeOnly[Eos enabled = false] STARTED
org.apache.kafka.streams.TopologyTestDriverTest > shouldReturnCorrectInMemoryStoreTypeOnly[Eos enabled = false] PASSED
org.apache.kafka.streams.TopologyTestDriverTest > shouldProcessRecordForTopic[Eos enabled = false] STARTED
org.apache.kafka.streams.TopologyTestDriverTest > shouldProcessRecordForTopic[Eos enabled = false] PASSED
org.apache.kafka.streams.TopologyTestDriverTest > shouldForwardRecordsFromSubtopologyToSubtopology[Eos enabled = false] STARTED
org.apache.kafka.streams.TopologyTestDriverTest > shouldForwardRecordsFromSubtopologyToSubtopology[Eos enabled = false] PASSED
org.apache.kafka.streams.TopologyTestDriverTest > shouldNotUpdateStoreForSmallerValue[Eos enabled = false] STARTED
org.apache.kafka.streams.TopologyTestDriverTest > shouldNotUpdateStoreForSmallerValue[Eos enabled = false] PASSED
org.apache.kafka.streams.TopologyTestDriverTest > shouldCreateStateDirectoryForStatefulTopology[Eos enabled = false] STARTED
org.apache.kafka.streams.TopologyTestDriverTest > shouldCreateStateDirectoryForStatefulTopology[Eos enabled = false] PASSED
org.apache.kafka.streams.TopologyTestDriverTest > shouldPunctuateIfWallClockTimeAdvances[Eos enabled = false] STARTED
org.apache.kafka.streams.TopologyTestDriverTest > shouldPunctuateIfWallClockTimeAdvances[Eos enabled = false] PASSED
FAILURE: Build failed with an exception.
* What went wrong:
Execution failed for task ':core:test'.
> There were failing tests. See the report at: file://<https://builds.apache.org/job/kafka-trunk-jdk11/ws/core/build/reports/tests/test/index.html>
* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
* Get more help at https://help.gradle.org
Deprecated Gradle features were used in this build, making it incompatible with Gradle 6.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/5.4.1/userguide/command_line_interface.html#sec:command_line_warnings
BUILD FAILED in 4h 12m 37s
178 actionable tasks: 175 executed, 3 up-to-date
Build step 'Execute shell' marked build as failure
Recording test results
ERROR: No tool found matching GRADLE_4_10_2_HOME
Setting GRADLE_4_10_3_HOME=/home/jenkins/tools/gradle/4.10.3
ERROR: No tool found matching GRADLE_4_10_2_HOME
Setting GRADLE_4_10_3_HOME=/home/jenkins/tools/gradle/4.10.3
Not sending mail to unregistered user ismael@juma.me.uk
Not sending mail to unregistered user vahid.hashemian@gmail.com
Not sending mail to unregistered user noreply@github.com
Not sending mail to unregistered user gwen@confluent.io
Build failed in Jenkins: kafka-trunk-jdk11 #750
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/kafka-trunk-jdk11/750/display/redirect?page=changes>
Changes:
[rhauch] KAFKA-8774: Regex can be found anywhere in config value (#7197)
[rhauch] KAFKA-7941: Catch TimeoutException in KafkaBasedLog worker thread
------------------------------------------
[...truncated 6.51 MB...]
org.apache.kafka.connect.transforms.TimestampConverterTest > testWithSchemaNullFieldToTime PASSED
org.apache.kafka.connect.transforms.TimestampConverterTest > testWithSchemaNullFieldToUnix STARTED
org.apache.kafka.connect.transforms.TimestampConverterTest > testWithSchemaNullFieldToUnix PASSED
org.apache.kafka.connect.transforms.TimestampConverterTest > testConfigNoTargetType STARTED
org.apache.kafka.connect.transforms.TimestampConverterTest > testConfigNoTargetType PASSED
org.apache.kafka.connect.transforms.TimestampConverterTest > testWithSchemaStringToTimestamp STARTED
org.apache.kafka.connect.transforms.TimestampConverterTest > testWithSchemaStringToTimestamp PASSED
org.apache.kafka.connect.transforms.TimestampConverterTest > testWithSchemaTimeToTimestamp STARTED
org.apache.kafka.connect.transforms.TimestampConverterTest > testWithSchemaTimeToTimestamp PASSED
org.apache.kafka.connect.transforms.TimestampConverterTest > testSchemalessUnixToTimestamp STARTED
org.apache.kafka.connect.transforms.TimestampConverterTest > testSchemalessUnixToTimestamp PASSED
org.apache.kafka.connect.transforms.TimestampConverterTest > testWithSchemaTimestampToString STARTED
org.apache.kafka.connect.transforms.TimestampConverterTest > testWithSchemaTimestampToString PASSED
org.apache.kafka.connect.transforms.TimestampConverterTest > testConfigInvalidFormat STARTED
org.apache.kafka.connect.transforms.TimestampConverterTest > testConfigInvalidFormat PASSED
> Task :streams:examples:test
org.apache.kafka.streams.examples.wordcount.WordCountProcessorTest > test STARTED
org.apache.kafka.streams.examples.wordcount.WordCountProcessorTest > test PASSED
> Task :streams:streams-scala:test
org.apache.kafka.streams.scala.kstream.KTableTest > filter a KTable should filter records satisfying the predicate STARTED
org.apache.kafka.streams.scala.kstream.KTableTest > filter a KTable should filter records satisfying the predicate PASSED
org.apache.kafka.streams.scala.kstream.KTableTest > filterNot a KTable should filter records not satisfying the predicate STARTED
org.apache.kafka.streams.scala.kstream.KTableTest > filterNot a KTable should filter records not satisfying the predicate PASSED
org.apache.kafka.streams.scala.kstream.KTableTest > join 2 KTables should join correctly records STARTED
org.apache.kafka.streams.scala.kstream.KTableTest > join 2 KTables should join correctly records PASSED
org.apache.kafka.streams.scala.kstream.KTableTest > join 2 KTables with a Materialized should join correctly records and state store STARTED
org.apache.kafka.streams.scala.kstream.KTableTest > join 2 KTables with a Materialized should join correctly records and state store PASSED
org.apache.kafka.streams.scala.kstream.KTableTest > windowed KTable#suppress should correctly suppress results using Suppressed.untilTimeLimit STARTED
org.apache.kafka.streams.scala.kstream.KTableTest > windowed KTable#suppress should correctly suppress results using Suppressed.untilTimeLimit PASSED
org.apache.kafka.streams.scala.kstream.KTableTest > windowed KTable#suppress should correctly suppress results using Suppressed.untilWindowCloses STARTED
org.apache.kafka.streams.scala.kstream.KTableTest > windowed KTable#suppress should correctly suppress results using Suppressed.untilWindowCloses PASSED
org.apache.kafka.streams.scala.kstream.KTableTest > session windowed KTable#suppress should correctly suppress results using Suppressed.untilWindowCloses STARTED
org.apache.kafka.streams.scala.kstream.KTableTest > session windowed KTable#suppress should correctly suppress results using Suppressed.untilWindowCloses PASSED
org.apache.kafka.streams.scala.kstream.KTableTest > non-windowed KTable#suppress should correctly suppress results using Suppressed.untilTimeLimit STARTED
org.apache.kafka.streams.scala.kstream.KTableTest > non-windowed KTable#suppress should correctly suppress results using Suppressed.untilTimeLimit PASSED
org.apache.kafka.streams.scala.kstream.ProducedTest > Create a Produced should create a Produced with Serdes STARTED
org.apache.kafka.streams.scala.kstream.ProducedTest > Create a Produced should create a Produced with Serdes PASSED
org.apache.kafka.streams.scala.kstream.ProducedTest > Create a Produced with timestampExtractor and resetPolicy should create a Consumed with Serdes, timestampExtractor and resetPolicy STARTED
org.apache.kafka.streams.scala.kstream.ProducedTest > Create a Produced with timestampExtractor and resetPolicy should create a Consumed with Serdes, timestampExtractor and resetPolicy PASSED
org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed should create a Consumed with Serdes STARTED
org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed should create a Consumed with Serdes PASSED
org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed with timestampExtractor and resetPolicy should create a Consumed with Serdes, timestampExtractor and resetPolicy STARTED
org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed with timestampExtractor and resetPolicy should create a Consumed with Serdes, timestampExtractor and resetPolicy PASSED
org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed with timestampExtractor should create a Consumed with Serdes and timestampExtractor STARTED
org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed with timestampExtractor should create a Consumed with Serdes and timestampExtractor PASSED
org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed with resetPolicy should create a Consumed with Serdes and resetPolicy STARTED
org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed with resetPolicy should create a Consumed with Serdes and resetPolicy PASSED
org.apache.kafka.streams.scala.kstream.JoinedTest > Create a Joined should create a Joined with Serdes STARTED
org.apache.kafka.streams.scala.kstream.JoinedTest > Create a Joined should create a Joined with Serdes PASSED
org.apache.kafka.streams.scala.kstream.JoinedTest > Create a Joined should create a Joined with Serdes and repartition topic name STARTED
org.apache.kafka.streams.scala.kstream.JoinedTest > Create a Joined should create a Joined with Serdes and repartition topic name PASSED
org.apache.kafka.streams.scala.kstream.KStreamTest > filter a KStream should filter records satisfying the predicate STARTED
org.apache.kafka.streams.scala.kstream.KStreamTest > filter a KStream should filter records satisfying the predicate PASSED
org.apache.kafka.streams.scala.kstream.KStreamTest > filterNot a KStream should filter records not satisfying the predicate STARTED
org.apache.kafka.streams.scala.kstream.KStreamTest > filterNot a KStream should filter records not satisfying the predicate PASSED
org.apache.kafka.streams.scala.kstream.KStreamTest > foreach a KStream should run foreach actions on records STARTED
org.apache.kafka.streams.scala.kstream.KStreamTest > foreach a KStream should run foreach actions on records PASSED
org.apache.kafka.streams.scala.kstream.KStreamTest > peek a KStream should run peek actions on records STARTED
org.apache.kafka.streams.scala.kstream.KStreamTest > peek a KStream should run peek actions on records PASSED
org.apache.kafka.streams.scala.kstream.KStreamTest > selectKey a KStream should select a new key STARTED
org.apache.kafka.streams.scala.kstream.KStreamTest > selectKey a KStream should select a new key PASSED
org.apache.kafka.streams.scala.kstream.KStreamTest > join 2 KStreams should join correctly records STARTED
org.apache.kafka.streams.scala.kstream.KStreamTest > join 2 KStreams should join correctly records PASSED
org.apache.kafka.streams.scala.kstream.SuppressedTest > Suppressed.untilWindowCloses should produce the correct suppression STARTED
org.apache.kafka.streams.scala.kstream.SuppressedTest > Suppressed.untilWindowCloses should produce the correct suppression PASSED
org.apache.kafka.streams.scala.kstream.SuppressedTest > Suppressed.untilTimeLimit should produce the correct suppression STARTED
org.apache.kafka.streams.scala.kstream.SuppressedTest > Suppressed.untilTimeLimit should produce the correct suppression PASSED
org.apache.kafka.streams.scala.kstream.SuppressedTest > BufferConfig.maxRecords should produce the correct buffer config STARTED
org.apache.kafka.streams.scala.kstream.SuppressedTest > BufferConfig.maxRecords should produce the correct buffer config PASSED
org.apache.kafka.streams.scala.kstream.SuppressedTest > BufferConfig.maxBytes should produce the correct buffer config STARTED
org.apache.kafka.streams.scala.kstream.SuppressedTest > BufferConfig.maxBytes should produce the correct buffer config PASSED
org.apache.kafka.streams.scala.kstream.SuppressedTest > BufferConfig.unbounded should produce the correct buffer config STARTED
org.apache.kafka.streams.scala.kstream.SuppressedTest > BufferConfig.unbounded should produce the correct buffer config PASSED
org.apache.kafka.streams.scala.kstream.SuppressedTest > BufferConfig should support very long chains of factory methods STARTED
org.apache.kafka.streams.scala.kstream.SuppressedTest > BufferConfig should support very long chains of factory methods PASSED
org.apache.kafka.streams.scala.kstream.GroupedTest > Create a Grouped should create a Grouped with Serdes STARTED
org.apache.kafka.streams.scala.kstream.GroupedTest > Create a Grouped should create a Grouped with Serdes PASSED
org.apache.kafka.streams.scala.kstream.GroupedTest > Create a Grouped with repartition topic name should create a Grouped with Serdes, and repartition topic name STARTED
org.apache.kafka.streams.scala.kstream.GroupedTest > Create a Grouped with repartition topic name should create a Grouped with Serdes, and repartition topic name PASSED
org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialized should create a Materialized with Serdes STARTED
org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialized should create a Materialized with Serdes PASSED
org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize with a store name should create a Materialized with Serdes and a store name STARTED
org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize with a store name should create a Materialized with Serdes and a store name PASSED
org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize with a window store supplier should create a Materialized with Serdes and a store supplier STARTED
org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize with a window store supplier should create a Materialized with Serdes and a store supplier PASSED
org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize with a key value store supplier should create a Materialized with Serdes and a store supplier STARTED
org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize with a key value store supplier should create a Materialized with Serdes and a store supplier PASSED
org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize with a session store supplier should create a Materialized with Serdes and a store supplier STARTED
org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize with a session store supplier should create a Materialized with Serdes and a store supplier PASSED
org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaJoin STARTED
org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaJoin PASSED
org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaSimple STARTED
org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaSimple PASSED
org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaAggregate STARTED
org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaAggregate PASSED
org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaProperties STARTED
org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaProperties PASSED
org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaTransform STARTED
org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaTransform PASSED
org.apache.kafka.streams.scala.StreamToTableJoinScalaIntegrationTestImplicitSerdes > testShouldCountClicksPerRegionWithNamedRepartitionTopic STARTED
org.apache.kafka.streams.scala.StreamToTableJoinScalaIntegrationTestImplicitSerdes > testShouldCountClicksPerRegionWithNamedRepartitionTopic PASSED
org.apache.kafka.streams.scala.StreamToTableJoinScalaIntegrationTestImplicitSerdes > testShouldCountClicksPerRegionJava STARTED
org.apache.kafka.streams.scala.StreamToTableJoinScalaIntegrationTestImplicitSerdes > testShouldCountClicksPerRegionJava PASSED
org.apache.kafka.streams.scala.StreamToTableJoinScalaIntegrationTestImplicitSerdes > testShouldCountClicksPerRegion STARTED
org.apache.kafka.streams.scala.StreamToTableJoinScalaIntegrationTestImplicitSerdes > testShouldCountClicksPerRegion PASSED
org.apache.kafka.streams.scala.WordCountTest > testShouldCountWordsMaterialized STARTED
org.apache.kafka.streams.scala.WordCountTest > testShouldCountWordsMaterialized PASSED
org.apache.kafka.streams.scala.WordCountTest > testShouldCountWordsJava STARTED
org.apache.kafka.streams.scala.WordCountTest > testShouldCountWordsJava PASSED
org.apache.kafka.streams.scala.WordCountTest > testShouldCountWords STARTED
org.apache.kafka.streams.scala.WordCountTest > testShouldCountWords PASSED
FAILURE: Build failed with an exception.
* What went wrong:
Execution failed for task ':connect:runtime:test'.
> There were failing tests. See the report at: file://<https://builds.apache.org/job/kafka-trunk-jdk11/ws/connect/runtime/build/reports/tests/test/index.html>
* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
* Get more help at https://help.gradle.org
Deprecated Gradle features were used in this build, making it incompatible with Gradle 6.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/5.4.1/userguide/command_line_interface.html#sec:command_line_warnings
BUILD FAILED in 3h 7m 25s
178 actionable tasks: 175 executed, 3 up-to-date
Build step 'Execute shell' marked build as failure
Recording test results
ERROR: No tool found matching GRADLE_4_10_2_HOME
Setting GRADLE_4_10_3_HOME=/home/jenkins/tools/gradle/4.10.3
ERROR: No tool found matching GRADLE_4_10_2_HOME
Setting GRADLE_4_10_3_HOME=/home/jenkins/tools/gradle/4.10.3
Not sending mail to unregistered user ismael@juma.me.uk
Not sending mail to unregistered user vahid.hashemian@gmail.com
Not sending mail to unregistered user noreply@github.com
Build failed in Jenkins: kafka-trunk-jdk11 #749
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/kafka-trunk-jdk11/749/display/redirect?page=changes>
Changes:
[cmccabe] MINOR: Clean up the sticky partitioner code a bit (#7151)
------------------------------------------
[...truncated 6.38 MB...]
org.apache.kafka.streams.TopologyTest > multipleSourcesWithSinksShouldHaveDistinctSubtopologies STARTED
org.apache.kafka.streams.TopologyTest > multipleSourcesWithSinksShouldHaveDistinctSubtopologies PASSED
org.apache.kafka.streams.TopologyTest > kGroupedStreamZeroArgCountShouldPreserveTopologyStructure STARTED
org.apache.kafka.streams.TopologyTest > kGroupedStreamZeroArgCountShouldPreserveTopologyStructure PASSED
org.apache.kafka.streams.TopologyTest > sessionWindowNamedMaterializedCountShouldPreserveTopologyStructure STARTED
org.apache.kafka.streams.TopologyTest > sessionWindowNamedMaterializedCountShouldPreserveTopologyStructure PASSED
org.apache.kafka.streams.internals.ApiUtilsTest > shouldReturnMillisecondsOnValidDuration STARTED
org.apache.kafka.streams.internals.ApiUtilsTest > shouldReturnMillisecondsOnValidDuration PASSED
org.apache.kafka.streams.internals.ApiUtilsTest > shouldThrowArithmeticExceptionForMaxInstant STARTED
org.apache.kafka.streams.internals.ApiUtilsTest > shouldThrowArithmeticExceptionForMaxInstant PASSED
org.apache.kafka.streams.internals.ApiUtilsTest > shouldThrowArithmeticExceptionForMaxDuration STARTED
org.apache.kafka.streams.internals.ApiUtilsTest > shouldThrowArithmeticExceptionForMaxDuration PASSED
org.apache.kafka.streams.internals.ApiUtilsTest > shouldReturnMillisecondsOnValidInstant STARTED
org.apache.kafka.streams.internals.ApiUtilsTest > shouldReturnMillisecondsOnValidInstant PASSED
org.apache.kafka.streams.internals.ApiUtilsTest > shouldThrowNullPointerExceptionForNullInstant STARTED
org.apache.kafka.streams.internals.ApiUtilsTest > shouldThrowNullPointerExceptionForNullInstant PASSED
org.apache.kafka.streams.internals.ApiUtilsTest > shouldContainsNameAndValueInFailMsgPrefix STARTED
org.apache.kafka.streams.internals.ApiUtilsTest > shouldContainsNameAndValueInFailMsgPrefix PASSED
org.apache.kafka.streams.internals.ApiUtilsTest > shouldThrowNullPointerExceptionForNullDuration STARTED
org.apache.kafka.streams.internals.ApiUtilsTest > shouldThrowNullPointerExceptionForNullDuration PASSED
org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testLeftInner[caching enabled = true] STARTED
org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testLeftInner[caching enabled = true] PASSED
org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testLeftOuter[caching enabled = true] STARTED
org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testLeftOuter[caching enabled = true] PASSED
org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testLeftLeft[caching enabled = true] STARTED
org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testLeftLeft[caching enabled = true] PASSED
org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testOuterLeft[caching enabled = true] STARTED
org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testOuterLeft[caching enabled = true] PASSED
org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testInner[caching enabled = true] STARTED
org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testInner[caching enabled = true] PASSED
org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testOuter[caching enabled = true] STARTED
org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testOuter[caching enabled = true] PASSED
org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testLeft[caching enabled = true] STARTED
org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testLeft[caching enabled = true] PASSED
org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testInnerInner[caching enabled = true] STARTED
org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testInnerInner[caching enabled = true] PASSED
org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testInnerOuter[caching enabled = true] STARTED
org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testInnerOuter[caching enabled = true] PASSED
org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testInnerLeft[caching enabled = true] STARTED
org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testInnerLeft[caching enabled = true] PASSED
org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testOuterInner[caching enabled = true] STARTED
org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testOuterInner[caching enabled = true] PASSED
org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testOuterOuter[caching enabled = true] STARTED
org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testOuterOuter[caching enabled = true] PASSED
org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testLeftInner[caching enabled = false] STARTED
org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testLeftInner[caching enabled = false] PASSED
org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testLeftOuter[caching enabled = false] STARTED
org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testLeftOuter[caching enabled = false] PASSED
org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testLeftLeft[caching enabled = false] STARTED
org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testLeftLeft[caching enabled = false] PASSED
org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testOuterLeft[caching enabled = false] STARTED
org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testOuterLeft[caching enabled = false] PASSED
org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testInner[caching enabled = false] STARTED
org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testInner[caching enabled = false] PASSED
org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testOuter[caching enabled = false] STARTED
org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testOuter[caching enabled = false] PASSED
org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testLeft[caching enabled = false] STARTED
org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testLeft[caching enabled = false] PASSED
org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testInnerInner[caching enabled = false] STARTED
org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testInnerInner[caching enabled = false] PASSED
org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testInnerOuter[caching enabled = false] STARTED
org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testInnerOuter[caching enabled = false] PASSED
org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testInnerLeft[caching enabled = false] STARTED
org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testInnerLeft[caching enabled = false] PASSED
org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testOuterInner[caching enabled = false] STARTED
org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testOuterInner[caching enabled = false] PASSED
org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testOuterOuter[caching enabled = false] STARTED
org.apache.kafka.streams.integration.TableTableJoinIntegrationTest > testOuterOuter[caching enabled = false] PASSED
org.apache.kafka.streams.integration.InternalTopicIntegrationTest > shouldCompactTopicsForKeyValueStoreChangelogs STARTED
org.apache.kafka.streams.integration.InternalTopicIntegrationTest > shouldCompactTopicsForKeyValueStoreChangelogs PASSED
org.apache.kafka.streams.integration.InternalTopicIntegrationTest > shouldCompactAndDeleteTopicsForWindowStoreChangelogs STARTED
org.apache.kafka.streams.integration.InternalTopicIntegrationTest > shouldCompactAndDeleteTopicsForWindowStoreChangelogs PASSED
org.apache.kafka.streams.integration.KStreamTransformIntegrationTest > shouldFlatTransform STARTED
org.apache.kafka.streams.integration.KStreamTransformIntegrationTest > shouldFlatTransform PASSED
org.apache.kafka.streams.integration.KStreamTransformIntegrationTest > shouldFlatTransformValuesWithValueTransformerWithoutKey STARTED
org.apache.kafka.streams.integration.KStreamTransformIntegrationTest > shouldFlatTransformValuesWithValueTransformerWithoutKey PASSED
org.apache.kafka.streams.integration.KStreamTransformIntegrationTest > shouldFlatTransformValuesWithKey STARTED
org.apache.kafka.streams.integration.KStreamTransformIntegrationTest > shouldFlatTransformValuesWithKey PASSED
org.apache.kafka.streams.integration.KStreamTransformIntegrationTest > shouldTransform STARTED
org.apache.kafka.streams.integration.KStreamTransformIntegrationTest > shouldTransform PASSED
org.apache.kafka.streams.integration.KStreamTransformIntegrationTest > shouldTransformValuesWithValueTransformerWithKey STARTED
org.apache.kafka.streams.integration.KStreamTransformIntegrationTest > shouldTransformValuesWithValueTransformerWithKey PASSED
org.apache.kafka.streams.integration.KStreamTransformIntegrationTest > shouldTransformValuesWithValueTransformerWithoutKey STARTED
org.apache.kafka.streams.integration.KStreamTransformIntegrationTest > shouldTransformValuesWithValueTransformerWithoutKey PASSED
org.apache.kafka.streams.integration.QueryableStateIntegrationTest > shouldBeAbleToQueryMapValuesState STARTED
org.apache.kafka.streams.integration.QueryableStateIntegrationTest > shouldBeAbleToQueryMapValuesState PASSED
org.apache.kafka.streams.integration.QueryableStateIntegrationTest > shouldNotMakeStoreAvailableUntilAllStoresAvailable STARTED
org.apache.kafka.streams.integration.QueryableStateIntegrationTest > shouldNotMakeStoreAvailableUntilAllStoresAvailable PASSED
org.apache.kafka.streams.integration.QueryableStateIntegrationTest > queryOnRebalance STARTED
org.apache.kafka.streams.integration.QueryableStateIntegrationTest > queryOnRebalance PASSED
org.apache.kafka.streams.integration.QueryableStateIntegrationTest > concurrentAccesses STARTED
org.apache.kafka.streams.integration.QueryableStateIntegrationTest > concurrentAccesses PASSED
org.apache.kafka.streams.integration.QueryableStateIntegrationTest > shouldAllowToQueryAfterThreadDied STARTED
org.apache.kafka.streams.integration.QueryableStateIntegrationTest > shouldAllowToQueryAfterThreadDied PASSED
org.apache.kafka.streams.integration.QueryableStateIntegrationTest > shouldBeAbleToQueryStateWithZeroSizedCache STARTED
org.apache.kafka.streams.integration.QueryableStateIntegrationTest > shouldBeAbleToQueryStateWithZeroSizedCache PASSED
org.apache.kafka.streams.integration.QueryableStateIntegrationTest > shouldBeAbleToQueryMapValuesAfterFilterState STARTED
org.apache.kafka.streams.integration.QueryableStateIntegrationTest > shouldBeAbleToQueryMapValuesAfterFilterState PASSED
org.apache.kafka.streams.integration.QueryableStateIntegrationTest > shouldBeAbleToQueryFilterState STARTED
org.apache.kafka.streams.integration.QueryableStateIntegrationTest > shouldBeAbleToQueryFilterState PASSED
org.apache.kafka.streams.integration.QueryableStateIntegrationTest > shouldBeAbleToQueryStateWithNonZeroSizedCache STARTED
org.apache.kafka.streams.integration.QueryableStateIntegrationTest > shouldBeAbleToQueryStateWithNonZeroSizedCache PASSED
org.apache.kafka.streams.integration.KStreamAggregationDedupIntegrationTest > shouldReduce STARTED
org.apache.kafka.streams.integration.KStreamAggregationDedupIntegrationTest > shouldReduce PASSED
org.apache.kafka.streams.integration.KStreamAggregationDedupIntegrationTest > shouldGroupByKey STARTED
org.apache.kafka.streams.integration.KStreamAggregationDedupIntegrationTest > shouldGroupByKey PASSED
org.apache.kafka.streams.integration.KStreamAggregationDedupIntegrationTest > shouldReduceWindowed STARTED
org.apache.kafka.streams.integration.KStreamAggregationDedupIntegrationTest > shouldReduceWindowed PASSED
org.apache.kafka.streams.integration.PurgeRepartitionTopicIntegrationTest > shouldRestoreState STARTED
org.apache.kafka.streams.integration.PurgeRepartitionTopicIntegrationTest > shouldRestoreState PASSED
org.apache.kafka.streams.integration.RepartitionOptimizingIntegrationTest > shouldSendCorrectRecords_OPTIMIZED STARTED
org.apache.kafka.streams.integration.RepartitionOptimizingIntegrationTest > shouldSendCorrectRecords_OPTIMIZED PASSED
org.apache.kafka.streams.integration.RepartitionOptimizingIntegrationTest > shouldSendCorrectResults_NO_OPTIMIZATION STARTED
org.apache.kafka.streams.integration.RepartitionOptimizingIntegrationTest > shouldSendCorrectResults_NO_OPTIMIZATION PASSED
org.apache.kafka.streams.integration.MetricsIntegrationTest > shouldAddMetricsOnAllLevels STARTED
org.apache.kafka.streams.integration.MetricsIntegrationTest > shouldAddMetricsOnAllLevels PASSED
org.apache.kafka.streams.integration.MetricsIntegrationTest > shouldAddMetricsForWindowStore STARTED
org.apache.kafka.streams.integration.MetricsIntegrationTest > shouldAddMetricsForWindowStore PASSED
org.apache.kafka.streams.integration.MetricsIntegrationTest > shouldAddMetricsForSessionStore STARTED
org.apache.kafka.streams.integration.MetricsIntegrationTest > shouldAddMetricsForSessionStore PASSED
Build timed out (after 480 minutes). Marking the build as failed.
Build was aborted
Recording test results
ERROR: No tool found matching GRADLE_4_10_2_HOME
Setting GRADLE_4_10_3_HOME=/home/jenkins/tools/gradle/4.10.3
ERROR: No tool found matching GRADLE_4_10_2_HOME
Setting GRADLE_4_10_3_HOME=/home/jenkins/tools/gradle/4.10.3
Not sending mail to unregistered user ismael@juma.me.uk
Not sending mail to unregistered user vahid.hashemian@gmail.com
Not sending mail to unregistered user noreply@github.com