You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@kafka.apache.org by Apache Jenkins Server <je...@builds.apache.org> on 2019/08/26 17:21:11 UTC
Build failed in Jenkins: kafka-2.1-jdk8 #226
See <https://builds.apache.org/job/kafka-2.1-jdk8/226/display/redirect?page=changes>
Changes:
[wangguoz] KAFKA-8412: Fix nullpointer exception thrown on flushing before closing
------------------------------------------
[...truncated 91.06 KB...]
return valueSerde != null ? valueSerde.deserializer() : null;
^
required: Deserializer<V>
found: Deserializer
where V is a type-variable:
V extends Object declared in class OptimizableRepartitionNode
<https://builds.apache.org/job/kafka-2.1-jdk8/ws/streams/src/main/java/org/apache/kafka/streams/kstream/internals/graph/OptimizableRepartitionNode.java>:78: warning: [unchecked] unchecked conversion
final Serializer<K> keySerializer = keySerde != null ? keySerde.serializer() : null;
^
required: Serializer<K>
found: Serializer
where K is a type-variable:
K extends Object declared in class OptimizableRepartitionNode
<https://builds.apache.org/job/kafka-2.1-jdk8/ws/streams/src/main/java/org/apache/kafka/streams/kstream/internals/graph/OptimizableRepartitionNode.java>:79: warning: [unchecked] unchecked conversion
final Deserializer<K> keyDeserializer = keySerde != null ? keySerde.deserializer() : null;
^
required: Deserializer<K>
found: Deserializer
where K is a type-variable:
K extends Object declared in class OptimizableRepartitionNode
<https://builds.apache.org/job/kafka-2.1-jdk8/ws/streams/src/main/java/org/apache/kafka/streams/kstream/internals/graph/TableSourceNode.java>:80: warning: [unchecked] unchecked cast
topologyBuilder.addGlobalStore((StoreBuilder<KeyValueStore>) storeBuilder,
^
required: StoreBuilder<KeyValueStore>
found: StoreBuilder<S>
where S is a type-variable:
S extends StateStore declared in class TableSourceNode
<https://builds.apache.org/job/kafka-2.1-jdk8/ws/streams/src/main/java/org/apache/kafka/streams/kstream/internals/graph/TableSourceNode.java>:133: warning: [unchecked] unchecked conversion
this.consumedInternal = consumedInternal;
^
required: ConsumedInternal<K,V>
found: ConsumedInternal
where K,V are type-variables:
K extends Object declared in class TableSourceNodeBuilder
V extends Object declared in class TableSourceNodeBuilder
<https://builds.apache.org/job/kafka-2.1-jdk8/ws/streams/src/main/java/org/apache/kafka/streams/kstream/internals/graph/StreamSinkNode.java>:61: warning: [unchecked] unchecked method invocation: constructor <init> in class WindowedStreamPartitioner is applied to given types
final StreamPartitioner<K, V> windowedPartitioner = (StreamPartitioner<K, V>) new WindowedStreamPartitioner<Object, V>((WindowedSerializer) keySerializer);
^
required: WindowedSerializer<K>
found: WindowedSerializer
where K is a type-variable:
K extends Object declared in class WindowedStreamPartitioner
<https://builds.apache.org/job/kafka-2.1-jdk8/ws/streams/src/main/java/org/apache/kafka/streams/kstream/internals/graph/StreamSinkNode.java>:61: warning: [unchecked] unchecked conversion
final StreamPartitioner<K, V> windowedPartitioner = (StreamPartitioner<K, V>) new WindowedStreamPartitioner<Object, V>((WindowedSerializer) keySerializer);
^
required: WindowedSerializer<K>
found: WindowedSerializer
where K is a type-variable:
K extends Object declared in class WindowedStreamPartitioner
<https://builds.apache.org/job/kafka-2.1-jdk8/ws/streams/src/main/java/org/apache/kafka/streams/kstream/internals/graph/StreamSinkNode.java>:61: warning: [unchecked] unchecked cast
final StreamPartitioner<K, V> windowedPartitioner = (StreamPartitioner<K, V>) new WindowedStreamPartitioner<Object, V>((WindowedSerializer) keySerializer);
^
required: StreamPartitioner<K,V>
found: WindowedStreamPartitioner<Object,V>
where V,K are type-variables:
V extends Object declared in class StreamSinkNode
K extends Object declared in class StreamSinkNode
<https://builds.apache.org/job/kafka-2.1-jdk8/ws/streams/src/main/java/org/apache/kafka/streams/kstream/internals/graph/KTableKTableJoinNode.java>:88: warning: [unchecked] unchecked method invocation: constructor <init> in class KeyValueStoreMaterializer is applied to given types
= new KeyValueStoreMaterializer<>(materializedInternal).materialize();
^
required: MaterializedInternal<K,V,KeyValueStore<Bytes,byte[]>>
found: MaterializedInternal
where K,V are type-variables:
K extends Object declared in class KeyValueStoreMaterializer
V extends Object declared in class KeyValueStoreMaterializer
<https://builds.apache.org/job/kafka-2.1-jdk8/ws/streams/src/main/java/org/apache/kafka/streams/kstream/internals/graph/KTableKTableJoinNode.java>:88: warning: [unchecked] unchecked conversion
= new KeyValueStoreMaterializer<>(materializedInternal).materialize();
^
required: MaterializedInternal<K,V,KeyValueStore<Bytes,byte[]>>
found: MaterializedInternal
where K,V are type-variables:
K extends Object declared in class KeyValueStoreMaterializer
V extends Object declared in class KeyValueStoreMaterializer
<https://builds.apache.org/job/kafka-2.1-jdk8/ws/streams/src/main/java/org/apache/kafka/streams/kstream/internals/graph/KTableKTableJoinNode.java>:88: warning: [unchecked] unchecked conversion
= new KeyValueStoreMaterializer<>(materializedInternal).materialize();
^
required: StoreBuilder<KeyValueStore<K,VR>>
found: StoreBuilder
where K,VR are type-variables:
K extends Object declared in class KTableKTableJoinNode
VR extends Object declared in class KTableKTableJoinNode
<https://builds.apache.org/job/kafka-2.1-jdk8/ws/streams/src/main/java/org/apache/kafka/streams/kstream/internals/graph/TableProcessorNode.java>:68: warning: [unchecked] unchecked cast
topologyBuilder.addStateStore(new KeyValueStoreMaterializer<>((MaterializedInternal<K, V, KeyValueStore<Bytes, byte[]>>) materializedInternal).materialize(), processorName);
^
required: MaterializedInternal<K,V,KeyValueStore<Bytes,byte[]>>
found: MaterializedInternal<K,V,S>
where K,V,S are type-variables:
K extends Object declared in class TableProcessorNode
V extends Object declared in class TableProcessorNode
S extends StateStore declared in class TableProcessorNode
<https://builds.apache.org/job/kafka-2.1-jdk8/ws/streams/src/main/java/org/apache/kafka/streams/state/internals/CompositeReadOnlyWindowStore.java>:98: warning: [deprecation] fetch(K,K,long,long) in ReadOnlyWindowStore has been deprecated
public KeyValueIterator<Windowed<K>, V> fetch(final K from, final K to, final long timeFrom, final long timeTo) {
^
where K,V are type-variables:
K extends Object declared in interface ReadOnlyWindowStore
V extends Object declared in interface ReadOnlyWindowStore
<https://builds.apache.org/job/kafka-2.1-jdk8/ws/streams/src/main/java/org/apache/kafka/streams/state/internals/CompositeReadOnlyWindowStore.java>:104: warning: [deprecation] fetch(K,K,long,long) in ReadOnlyWindowStore has been deprecated
return store.fetch(from, to, timeFrom, timeTo);
^
where K,V are type-variables:
K extends Object declared in interface ReadOnlyWindowStore
V extends Object declared in interface ReadOnlyWindowStore
<https://builds.apache.org/job/kafka-2.1-jdk8/ws/streams/src/main/java/org/apache/kafka/streams/state/internals/RocksDBWindowStore.java>:107: warning: [deprecation] fetchAll(long,long) in ReadOnlyWindowStore has been deprecated
public KeyValueIterator<Windowed<K>, V> fetchAll(final long timeFrom, final long timeTo) {
^
where K,V are type-variables:
K extends Object declared in interface ReadOnlyWindowStore
V extends Object declared in interface ReadOnlyWindowStore
<https://builds.apache.org/job/kafka-2.1-jdk8/ws/streams/src/main/java/org/apache/kafka/streams/state/internals/RocksDBWindowStore.java>:95: warning: [deprecation] fetch(K,K,long,long) in ReadOnlyWindowStore has been deprecated
public KeyValueIterator<Windowed<K>, V> fetch(final K from, final K to, final long timeFrom, final long timeTo) {
^
where K,V are type-variables:
K extends Object declared in interface ReadOnlyWindowStore
V extends Object declared in interface ReadOnlyWindowStore
<https://builds.apache.org/job/kafka-2.1-jdk8/ws/streams/src/main/java/org/apache/kafka/streams/state/internals/RocksDBWindowStore.java>:89: warning: [deprecation] fetch(K,long,long) in ReadOnlyWindowStore has been deprecated
public WindowStoreIterator<V> fetch(final K key, final long timeFrom, final long timeTo) {
^
where K,V are type-variables:
K extends Object declared in interface ReadOnlyWindowStore
V extends Object declared in interface ReadOnlyWindowStore
<https://builds.apache.org/job/kafka-2.1-jdk8/ws/streams/src/main/java/org/apache/kafka/streams/state/internals/MeteredWindowStore.java>:156: warning: [deprecation] fetchAll(long,long) in ReadOnlyWindowStore has been deprecated
public KeyValueIterator<Windowed<K>, V> fetchAll(final long timeFrom, final long timeTo) {
^
where K,V are type-variables:
K extends Object declared in interface ReadOnlyWindowStore
V extends Object declared in interface ReadOnlyWindowStore
<https://builds.apache.org/job/kafka-2.1-jdk8/ws/streams/src/main/java/org/apache/kafka/streams/state/internals/MeteredWindowStore.java>:165: warning: [deprecation] fetch(K,K,long,long) in ReadOnlyWindowStore has been deprecated
public KeyValueIterator<Windowed<K>, V> fetch(final K from, final K to, final long timeFrom, final long timeTo) {
^
where K,V are type-variables:
K extends Object declared in interface ReadOnlyWindowStore
V extends Object declared in interface ReadOnlyWindowStore
<https://builds.apache.org/job/kafka-2.1-jdk8/ws/streams/src/main/java/org/apache/kafka/streams/state/internals/MeteredWindowStore.java>:142: warning: [deprecation] fetch(K,long,long) in ReadOnlyWindowStore has been deprecated
public WindowStoreIterator<V> fetch(final K key, final long timeFrom, final long timeTo) {
^
where K,V are type-variables:
K extends Object declared in interface ReadOnlyWindowStore
V extends Object declared in interface ReadOnlyWindowStore
<https://builds.apache.org/job/kafka-2.1-jdk8/ws/streams/src/main/java/org/apache/kafka/streams/state/internals/CachingWindowStore.java>:259: warning: [deprecation] fetchAll(long,long) in ReadOnlyWindowStore has been deprecated
public KeyValueIterator<Windowed<Bytes>, byte[]> fetchAll(final long timeFrom, final long timeTo) {
^
where K,V are type-variables:
K extends Object declared in interface ReadOnlyWindowStore
V extends Object declared in interface ReadOnlyWindowStore
<https://builds.apache.org/job/kafka-2.1-jdk8/ws/streams/src/main/java/org/apache/kafka/streams/state/internals/CachingWindowStore.java>:206: warning: [deprecation] fetch(K,K,long,long) in ReadOnlyWindowStore has been deprecated
public KeyValueIterator<Windowed<Bytes>, byte[]> fetch(final Bytes from, final Bytes to, final long timeFrom, final long timeTo) {
^
where K,V are type-variables:
K extends Object declared in interface ReadOnlyWindowStore
V extends Object declared in interface ReadOnlyWindowStore
<https://builds.apache.org/job/kafka-2.1-jdk8/ws/streams/src/main/java/org/apache/kafka/streams/state/internals/CachingWindowStore.java>:181: warning: [deprecation] fetch(K,long,long) in ReadOnlyWindowStore has been deprecated
public synchronized WindowStoreIterator<byte[]> fetch(final Bytes key, final long timeFrom, final long timeTo) {
^
where K,V are type-variables:
K extends Object declared in interface ReadOnlyWindowStore
V extends Object declared in interface ReadOnlyWindowStore
<https://builds.apache.org/job/kafka-2.1-jdk8/ws/streams/src/main/java/org/apache/kafka/streams/state/internals/ChangeLoggingWindowBytesStore.java>:70: warning: [deprecation] fetchAll(long,long) in ReadOnlyWindowStore has been deprecated
public KeyValueIterator<Windowed<Bytes>, byte[]> fetchAll(final long timeFrom, final long timeTo) {
^
where K,V are type-variables:
K extends Object declared in interface ReadOnlyWindowStore
V extends Object declared in interface ReadOnlyWindowStore
<https://builds.apache.org/job/kafka-2.1-jdk8/ws/streams/src/main/java/org/apache/kafka/streams/state/internals/ChangeLoggingWindowBytesStore.java>:60: warning: [deprecation] fetch(K,K,long,long) in ReadOnlyWindowStore has been deprecated
public KeyValueIterator<Windowed<Bytes>, byte[]> fetch(final Bytes keyFrom, final Bytes keyTo, final long from, final long to) {
^
where K,V are type-variables:
K extends Object declared in interface ReadOnlyWindowStore
V extends Object declared in interface ReadOnlyWindowStore
<https://builds.apache.org/job/kafka-2.1-jdk8/ws/streams/src/main/java/org/apache/kafka/streams/state/internals/ChangeLoggingWindowBytesStore.java>:55: warning: [deprecation] fetch(K,long,long) in ReadOnlyWindowStore has been deprecated
public WindowStoreIterator<byte[]> fetch(final Bytes key, final long from, final long to) {
^
where K,V are type-variables:
K extends Object declared in interface ReadOnlyWindowStore
V extends Object declared in interface ReadOnlyWindowStore
37 warnings
> Task :kafka-2.1-jdk8:streams:processResources NO-SOURCE
> Task :kafka-2.1-jdk8:streams:classes
> Task :kafka-2.1-jdk8:streams:copyDependantLibs
> Task :kafka-2.1-jdk8:streams:jar
> Task :kafka-2.1-jdk8:streams:test-utils:compileJava
> Task :kafka-2.1-jdk8:streams:test-utils:processResources NO-SOURCE
> Task :kafka-2.1-jdk8:streams:test-utils:classes
> Task :kafka-2.1-jdk8:streams:test-utils:copyDependantLibs
> Task :kafka-2.1-jdk8:streams:test-utils:jar
> Task :kafka-2.1-jdk8:core:spotbugsMain
> Task :kafka-2.1-jdk8:streams:compileTestJava FAILED
<https://builds.apache.org/job/kafka-2.1-jdk8/ws/streams/src/test/java/org/apache/kafka/streams/processor/internals/InternalTopologyBuilderTest.java>:172: warning: non-varargs call of varargs method with inexact argument type for last parameter;
builder.addProcessor("processor", new MockProcessorSupplier(), null);
^
cast to String for a varargs call
cast to String[] for a non-varargs call and to suppress this warning
<https://builds.apache.org/job/kafka-2.1-jdk8/ws/streams/src/test/java/org/apache/kafka/streams/processor/internals/InternalTopologyBuilderTest.java>:203: warning: non-varargs call of varargs method with inexact argument type for last parameter;
builder.addSink("sink", "topic", null, null, null, null);
^
cast to String for a varargs call
cast to String[] for a non-varargs call and to suppress this warning
<https://builds.apache.org/job/kafka-2.1-jdk8/ws/streams/src/test/java/org/apache/kafka/streams/processor/internals/AssignedStreamsTasksTest.java>:504: error: constructor StateDirectory in class StateDirectory cannot be applied to given types;
final StateDirectory stateDirectory = new StateDirectory(
^
required: StreamsConfig,Time
found: StreamsConfig,MockTime,boolean
reason: actual and formal argument lists differ in length
<https://builds.apache.org/job/kafka-2.1-jdk8/ws/streams/src/test/java/org/apache/kafka/streams/processor/internals/AssignedStreamsTasksTest.java>:509: error: no suitable constructor found for StreamTask(TaskId,Set<TopicPartition>,ProcessorTopology,MockConsumer<byte[],byte[]>,ChangelogReader,StreamsConfig,StreamsMetricsImpl,StateDirectory,<null>,MockTime,()->producer)
final StreamTask task = new StreamTask(
^
constructor StreamTask.StreamTask(TaskId,Collection<TopicPartition>,ProcessorTopology,Consumer<byte[],byte[]>,ChangelogReader,StreamsConfig,StreamsMetricsImpl,StateDirectory,ThreadCache,Time,ProducerSupplier,Sensor) is not applicable
(actual and formal argument lists differ in length)
constructor StreamTask.StreamTask(TaskId,Collection<TopicPartition>,ProcessorTopology,Consumer<byte[],byte[]>,ChangelogReader,StreamsConfig,StreamsMetricsImpl,StateDirectory,ThreadCache,Time,ProducerSupplier,RecordCollector,Sensor) is not applicable
(actual and formal argument lists differ in length)
Note: Some input files use or override a deprecated API.
Note: Recompile with -Xlint:deprecation for details.
Note: Some input files use unchecked or unsafe operations.
Note: Recompile with -Xlint:unchecked for details.
2 errors
2 warnings
> Task :testScala_2_11 FAILED
FAILURE: Build failed with an exception.
* What went wrong:
Execution failed for task ':kafka-2.1-jdk8:streams:compileTestJava'.
> Compilation failed; see the compiler error output for details.
* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
* Get more help at https://help.gradle.org
BUILD FAILED in 5m 46s
26 actionable tasks: 3 executed, 23 up-to-date
Build step 'Execute shell' marked build as failure
[FINDBUGS] Collecting findbugs analysis files...
Setting GRADLE_4_8_1_HOME=/home/jenkins/tools/gradle/4.8.1
[FINDBUGS] Searching for all files in <https://builds.apache.org/job/kafka-2.1-jdk8/ws/> that match the pattern **/build/reports/findbugs/*.xml
[FINDBUGS] No files found. Configuration error?
Setting GRADLE_4_8_1_HOME=/home/jenkins/tools/gradle/4.8.1
No credentials specified
Setting GRADLE_4_8_1_HOME=/home/jenkins/tools/gradle/4.8.1
<Git Blamer> Using GitBlamer to create author and commit information for all warnings.
<Git Blamer> GIT_COMMIT=8cdc2d6d1b2a227f5e62f148179a52684b230bd6, workspace=<https://builds.apache.org/job/kafka-2.1-jdk8/ws/>
[FINDBUGS] Computing warning deltas based on reference build #225
Recording test results
Setting GRADLE_4_8_1_HOME=/home/jenkins/tools/gradle/4.8.1
ERROR: Step ?Publish JUnit test result report? failed: No test report files were found. Configuration error?
Setting GRADLE_4_8_1_HOME=/home/jenkins/tools/gradle/4.8.1
Not sending mail to unregistered user wangguoz@gmail.com
Jenkins build is back to normal : kafka-2.1-jdk8 #230
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/kafka-2.1-jdk8/230/display/redirect?page=changes>
Build failed in Jenkins: kafka-2.1-jdk8 #229
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/kafka-2.1-jdk8/229/display/redirect?page=changes>
Changes:
[rhauch] HOTFIX: fix compile error in 2.1 branch (#7328)
------------------------------------------
[...truncated 316.00 KB...]
<https://builds.apache.org/job/kafka-2.1-jdk8/ws/streams/streams-scala/src/test/scala/org/apache/kafka/streams/scala/TopologyTest.scala>:139: local val wordCounts in method getTopologyJava is never used
val wordCounts: KTableJ[String, java.lang.Long] = grouped.count()
^
<https://builds.apache.org/job/kafka-2.1-jdk8/ws/streams/streams-scala/src/test/scala/org/apache/kafka/streams/scala/TopologyTest.scala>:160: local val clicksPerRegion in method getTopologyScala is never used
val clicksPerRegion: KTable[String, Long] =
^
<https://builds.apache.org/job/kafka-2.1-jdk8/ws/streams/streams-scala/src/test/scala/org/apache/kafka/streams/scala/TopologyTest.scala>:204: local val clicksPerRegion in method getTopologyJava is never used
val clicksPerRegion: KTableJ[String, JLong] = clicksByRegion
^
<https://builds.apache.org/job/kafka-2.1-jdk8/ws/streams/streams-scala/src/test/scala/org/apache/kafka/streams/scala/TopologyTest.scala>:274: local val wordCounts in method getTopologyJava is never used
val wordCounts: KTableJ[String, java.lang.Long] = grouped.count()
^
6 warnings found
> Task :kafka-2.1-jdk8:streams:streams-scala:processTestResources UP-TO-DATE
> Task :kafka-2.1-jdk8:streams:streams-scala:testClasses
> Task :kafka-2.1-jdk8:streams:streams-scala:checkstyleTest NO-SOURCE
> Task :clients:checkstyleTest
> Task :testScala_2_11 FAILED
> Task :kafka-2.1-jdk8:streams:streams-scala:spotbugsMain
> Task :kafka-2.1-jdk8:streams:streams-scala:test
org.apache.kafka.streams.scala.kstream.KStreamTest > filter a KStream should filter records satisfying the predicate STARTED
org.apache.kafka.streams.scala.kstream.KStreamTest > filter a KStream should filter records satisfying the predicate PASSED
org.apache.kafka.streams.scala.kstream.KStreamTest > filterNot a KStream should filter records not satisfying the predicate STARTED
org.apache.kafka.streams.scala.kstream.KStreamTest > filterNot a KStream should filter records not satisfying the predicate PASSED
org.apache.kafka.streams.scala.kstream.KStreamTest > foreach a KStream should run foreach actions on records STARTED
org.apache.kafka.streams.scala.kstream.KStreamTest > foreach a KStream should run foreach actions on records PASSED
org.apache.kafka.streams.scala.kstream.KStreamTest > peek a KStream should run peek actions on records STARTED
org.apache.kafka.streams.scala.kstream.KStreamTest > peek a KStream should run peek actions on records PASSED
org.apache.kafka.streams.scala.kstream.KStreamTest > selectKey a KStream should select a new key STARTED
org.apache.kafka.streams.scala.kstream.KStreamTest > selectKey a KStream should select a new key PASSED
org.apache.kafka.streams.scala.kstream.KStreamTest > join 2 KStreams should join correctly records STARTED
> Task :clients:spotbugsMain
> Task :kafka-2.1-jdk8:streams:streams-scala:test
org.apache.kafka.streams.scala.kstream.KStreamTest > join 2 KStreams should join correctly records PASSED
org.apache.kafka.streams.scala.kstream.JoinedTest > Create a Joined should create a Joined with Serdes STARTED
org.apache.kafka.streams.scala.kstream.JoinedTest > Create a Joined should create a Joined with Serdes PASSED
org.apache.kafka.streams.scala.kstream.JoinedTest > Create a Joined should create a Joined with Serdes and repartition topic name STARTED
org.apache.kafka.streams.scala.kstream.JoinedTest > Create a Joined should create a Joined with Serdes and repartition topic name PASSED
org.apache.kafka.streams.scala.kstream.GroupedTest > Create a Grouped should create a Grouped with Serdes STARTED
org.apache.kafka.streams.scala.kstream.GroupedTest > Create a Grouped should create a Grouped with Serdes PASSED
org.apache.kafka.streams.scala.kstream.GroupedTest > Create a Grouped with repartition topic name should create a Grouped with Serdes, and repartition topic name STARTED
org.apache.kafka.streams.scala.kstream.GroupedTest > Create a Grouped with repartition topic name should create a Grouped with Serdes, and repartition topic name PASSED
org.apache.kafka.streams.scala.kstream.ProducedTest > Create a Produced should create a Produced with Serdes STARTED
org.apache.kafka.streams.scala.kstream.ProducedTest > Create a Produced should create a Produced with Serdes PASSED
org.apache.kafka.streams.scala.kstream.ProducedTest > Create a Produced with timestampExtractor and resetPolicy should create a Consumed with Serdes, timestampExtractor and resetPolicy STARTED
org.apache.kafka.streams.scala.kstream.ProducedTest > Create a Produced with timestampExtractor and resetPolicy should create a Consumed with Serdes, timestampExtractor and resetPolicy PASSED
org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed should create a Consumed with Serdes STARTED
org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed should create a Consumed with Serdes PASSED
org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed with timestampExtractor and resetPolicy should create a Consumed with Serdes, timestampExtractor and resetPolicy STARTED
org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed with timestampExtractor and resetPolicy should create a Consumed with Serdes, timestampExtractor and resetPolicy PASSED
org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed with timestampExtractor should create a Consumed with Serdes and timestampExtractor STARTED
org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed with timestampExtractor should create a Consumed with Serdes and timestampExtractor PASSED
org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed with resetPolicy should create a Consumed with Serdes and resetPolicy STARTED
org.apache.kafka.streams.scala.kstream.ConsumedTest > Create a Consumed with resetPolicy should create a Consumed with Serdes and resetPolicy PASSED
org.apache.kafka.streams.scala.kstream.KTableTest > filter a KTable should filter records satisfying the predicate STARTED
org.apache.kafka.streams.scala.kstream.KTableTest > filter a KTable should filter records satisfying the predicate PASSED
org.apache.kafka.streams.scala.kstream.KTableTest > filterNot a KTable should filter records not satisfying the predicate STARTED
org.apache.kafka.streams.scala.kstream.KTableTest > filterNot a KTable should filter records not satisfying the predicate PASSED
org.apache.kafka.streams.scala.kstream.KTableTest > join 2 KTables should join correctly records STARTED
org.apache.kafka.streams.scala.kstream.KTableTest > join 2 KTables should join correctly records PASSED
org.apache.kafka.streams.scala.kstream.KTableTest > join 2 KTables with a Materialized should join correctly records and state store STARTED
org.apache.kafka.streams.scala.kstream.KTableTest > join 2 KTables with a Materialized should join correctly records and state store PASSED
org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialized should create a Materialized with Serdes STARTED
org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialized should create a Materialized with Serdes PASSED
org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize with a store name should create a Materialized with Serdes and a store name STARTED
org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize with a store name should create a Materialized with Serdes and a store name PASSED
org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize with a window store supplier should create a Materialized with Serdes and a store supplier STARTED
org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize with a window store supplier should create a Materialized with Serdes and a store supplier PASSED
org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize with a key value store supplier should create a Materialized with Serdes and a store supplier STARTED
org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize with a key value store supplier should create a Materialized with Serdes and a store supplier PASSED
org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize with a session store supplier should create a Materialized with Serdes and a store supplier STARTED
org.apache.kafka.streams.scala.kstream.MaterializedTest > Create a Materialize with a session store supplier should create a Materialized with Serdes and a store supplier PASSED
org.apache.kafka.streams.scala.StreamToTableJoinScalaIntegrationTestImplicitSerdes > testShouldCountClicksPerRegionWithNamedRepartitionTopic STARTED
org.apache.kafka.streams.scala.StreamToTableJoinScalaIntegrationTestImplicitSerdes > testShouldCountClicksPerRegionWithNamedRepartitionTopic FAILED
java.lang.AssertionError: Condition not met within timeout 30000. Did not receive all [KeyValue(americas, 101), KeyValue(europe, 109), KeyValue(asia, 124)] records from topic output-topic
at org.apache.kafka.test.TestUtils.waitForCondition(TestUtils.java:278)
at org.apache.kafka.streams.integration.utils.IntegrationTestUtils.waitUntilFinalKeyValueRecordsReceived(IntegrationTestUtils.java:452)
at org.apache.kafka.streams.integration.utils.IntegrationTestUtils.waitUntilFinalKeyValueRecordsReceived(IntegrationTestUtils.java:419)
at org.apache.kafka.streams.scala.utils.StreamToTableJoinScalaIntegrationTestBase.produceNConsume(StreamToTableJoinScalaIntegrationTestBase.scala:132)
at org.apache.kafka.streams.scala.StreamToTableJoinScalaIntegrationTestImplicitSerdes.testShouldCountClicksPerRegionWithNamedRepartitionTopic(StreamToTableJoinScalaIntegrationTestImplicitSerdes.scala:117)
org.apache.kafka.streams.scala.StreamToTableJoinScalaIntegrationTestImplicitSerdes > testShouldCountClicksPerRegionJava STARTED
org.apache.kafka.streams.scala.StreamToTableJoinScalaIntegrationTestImplicitSerdes > testShouldCountClicksPerRegionJava FAILED
java.lang.AssertionError: Condition not met within timeout 30000. Did not receive all [KeyValue(americas, 101), KeyValue(europe, 109), KeyValue(asia, 124)] records from topic output-topic-j
at org.apache.kafka.test.TestUtils.waitForCondition(TestUtils.java:278)
at org.apache.kafka.streams.integration.utils.IntegrationTestUtils.waitUntilFinalKeyValueRecordsReceived(IntegrationTestUtils.java:452)
at org.apache.kafka.streams.integration.utils.IntegrationTestUtils.waitUntilFinalKeyValueRecordsReceived(IntegrationTestUtils.java:419)
at org.apache.kafka.streams.scala.utils.StreamToTableJoinScalaIntegrationTestBase.produceNConsume(StreamToTableJoinScalaIntegrationTestBase.scala:132)
at org.apache.kafka.streams.scala.StreamToTableJoinScalaIntegrationTestImplicitSerdes.testShouldCountClicksPerRegionJava(StreamToTableJoinScalaIntegrationTestImplicitSerdes.scala:176)
org.apache.kafka.streams.scala.StreamToTableJoinScalaIntegrationTestImplicitSerdes > testShouldCountClicksPerRegion STARTED
org.apache.kafka.streams.scala.StreamToTableJoinScalaIntegrationTestImplicitSerdes > testShouldCountClicksPerRegion FAILED
java.lang.AssertionError: Condition not met within timeout 30000. Did not receive all [KeyValue(americas, 101), KeyValue(europe, 109), KeyValue(asia, 124)] records from topic output-topic
at org.apache.kafka.test.TestUtils.waitForCondition(TestUtils.java:278)
at org.apache.kafka.streams.integration.utils.IntegrationTestUtils.waitUntilFinalKeyValueRecordsReceived(IntegrationTestUtils.java:452)
at org.apache.kafka.streams.integration.utils.IntegrationTestUtils.waitUntilFinalKeyValueRecordsReceived(IntegrationTestUtils.java:419)
at org.apache.kafka.streams.scala.utils.StreamToTableJoinScalaIntegrationTestBase.produceNConsume(StreamToTableJoinScalaIntegrationTestBase.scala:132)
at org.apache.kafka.streams.scala.StreamToTableJoinScalaIntegrationTestImplicitSerdes.testShouldCountClicksPerRegion(StreamToTableJoinScalaIntegrationTestImplicitSerdes.scala:77)
org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaJoin STARTED
org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaJoin PASSED
org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaSimple STARTED
org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaSimple PASSED
org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaAggregate STARTED
org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaAggregate PASSED
org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaProperties STARTED
org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaProperties PASSED
org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaTransform STARTED
org.apache.kafka.streams.scala.TopologyTest > shouldBuildIdenticalTopologyInJavaNScalaTransform PASSED
org.apache.kafka.streams.scala.WordCountTest > testShouldCountWordsMaterialized STARTED
org.apache.kafka.streams.scala.WordCountTest > testShouldCountWordsMaterialized FAILED
java.lang.AssertionError: Condition not met within timeout 30000. Did not receive all 13 records from topic outputTopic
at org.apache.kafka.test.TestUtils.waitForCondition(TestUtils.java:278)
at org.apache.kafka.streams.integration.utils.IntegrationTestUtils.waitUntilMinKeyValueRecordsReceived(IntegrationTestUtils.java:383)
at org.apache.kafka.streams.integration.utils.IntegrationTestUtils.waitUntilMinKeyValueRecordsReceived(IntegrationTestUtils.java:357)
at org.apache.kafka.streams.scala.WordCountTest.produceNConsume(WordCountTest.scala:221)
at org.apache.kafka.streams.scala.WordCountTest.testShouldCountWordsMaterialized(WordCountTest.scala:127)
org.apache.kafka.streams.scala.WordCountTest > testShouldCountWordsJava STARTED
org.apache.kafka.streams.scala.WordCountTest > testShouldCountWordsJava FAILED
java.lang.AssertionError: Condition not met within timeout 30000. Did not receive all 13 records from topic outputTopicJ
at org.apache.kafka.test.TestUtils.waitForCondition(TestUtils.java:278)
at org.apache.kafka.streams.integration.utils.IntegrationTestUtils.waitUntilMinKeyValueRecordsReceived(IntegrationTestUtils.java:383)
at org.apache.kafka.streams.integration.utils.IntegrationTestUtils.waitUntilMinKeyValueRecordsReceived(IntegrationTestUtils.java:357)
at org.apache.kafka.streams.scala.WordCountTest.produceNConsume(WordCountTest.scala:221)
at org.apache.kafka.streams.scala.WordCountTest.testShouldCountWordsJava(WordCountTest.scala:174)
org.apache.kafka.streams.scala.WordCountTest > testShouldCountWords STARTED
org.apache.kafka.streams.scala.WordCountTest > testShouldCountWords PASSED
36 tests completed, 5 failed
> Task :kafka-2.1-jdk8:streams:streams-scala:test FAILED
> Task :testScala_2_12 FAILED
FAILURE: Build completed with 2 failures.
1: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':kafka-2.1-jdk8:core:spotbugsMain'.
> A failure occurred while executing com.github.spotbugs.internal.spotbugs.SpotBugsRunner
> SpotBugs rule violations were found. See the report at: file://<https://builds.apache.org/job/kafka-2.1-jdk8/ws/core/build/reports/spotbugs/main.xml>
* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================
2: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':kafka-2.1-jdk8:streams:streams-scala:test'.
> There were failing tests. See the report at: file://<https://builds.apache.org/job/kafka-2.1-jdk8/ws/streams/streams-scala/build/reports/tests/test/index.html>
* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================
* Get more help at https://help.gradle.org
BUILD FAILED in 25m 7s
34 actionable tasks: 7 executed, 27 up-to-date
Build step 'Execute shell' marked build as failure
[FINDBUGS] Collecting findbugs analysis files...
Setting GRADLE_4_8_1_HOME=/home/jenkins/tools/gradle/4.8.1
[FINDBUGS] Searching for all files in <https://builds.apache.org/job/kafka-2.1-jdk8/ws/> that match the pattern **/build/reports/findbugs/*.xml
[FINDBUGS] No files found. Configuration error?
Setting GRADLE_4_8_1_HOME=/home/jenkins/tools/gradle/4.8.1
No credentials specified
Setting GRADLE_4_8_1_HOME=/home/jenkins/tools/gradle/4.8.1
<Git Blamer> Using GitBlamer to create author and commit information for all warnings.
<Git Blamer> GIT_COMMIT=7f68a626102b49ff05a1eee0656f21116646d21a, workspace=<https://builds.apache.org/job/kafka-2.1-jdk8/ws/>
[FINDBUGS] Computing warning deltas based on reference build #225
Recording test results
Setting GRADLE_4_8_1_HOME=/home/jenkins/tools/gradle/4.8.1
Setting GRADLE_4_8_1_HOME=/home/jenkins/tools/gradle/4.8.1
Not sending mail to unregistered user bill@confluent.io
Not sending mail to unregistered user wangguoz@gmail.com
Build failed in Jenkins: kafka-2.1-jdk8 #228
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/kafka-2.1-jdk8/228/display/redirect?page=changes>
Changes:
[wangguoz] KAFKA-8816: Make offsets immutable to users of RecordCollector.offsets
------------------------------------------
Started by an SCM change
[EnvInject] - Loading node environment variables.
Building remotely on H31 (ubuntu xenial) in workspace <https://builds.apache.org/job/kafka-2.1-jdk8/ws/>
[WS-CLEANUP] Deleting project workspace...
[WS-CLEANUP] Deferred wipeout is used...
No credentials specified
Cloning the remote Git repository
Cloning repository https://github.com/apache/kafka.git
> git init <https://builds.apache.org/job/kafka-2.1-jdk8/ws/> # timeout=10
Fetching upstream changes from https://github.com/apache/kafka.git
> git --version # timeout=10
> git fetch --tags --progress https://github.com/apache/kafka.git +refs/heads/*:refs/remotes/origin/*
> git config remote.origin.url https://github.com/apache/kafka.git # timeout=10
> git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
> git config remote.origin.url https://github.com/apache/kafka.git # timeout=10
Fetching upstream changes from https://github.com/apache/kafka.git
> git fetch --tags --progress https://github.com/apache/kafka.git +refs/heads/*:refs/remotes/origin/*
> git rev-parse refs/remotes/origin/2.1^{commit} # timeout=10
> git rev-parse refs/remotes/origin/origin/2.1^{commit} # timeout=10
Checking out Revision 3242843021933ea9a51e0633e413bc24f6c9165c (refs/remotes/origin/2.1)
> git config core.sparsecheckout # timeout=10
> git checkout -f 3242843021933ea9a51e0633e413bc24f6c9165c
Commit message: "KAFKA-8816: Make offsets immutable to users of RecordCollector.offsets (#7223)"
> git rev-list --no-walk 6560203d20e818515d5b8c961fa648788c684cd6 # timeout=10
Setting GRADLE_4_8_1_HOME=/home/jenkins/tools/gradle/4.8.1
Setting GRADLE_4_8_1_HOME=/home/jenkins/tools/gradle/4.8.1
[kafka-2.1-jdk8] $ /bin/bash -xe /tmp/jenkins245598649444072459.sh
+ rm -rf <https://builds.apache.org/job/kafka-2.1-jdk8/ws/.gradle>
+ /home/jenkins/tools/gradle/4.8.1/bin/gradle
/tmp/jenkins245598649444072459.sh: line 4: /home/jenkins/tools/gradle/4.8.1/bin/gradle: No such file or directory
Build step 'Execute shell' marked build as failure
[FINDBUGS] Collecting findbugs analysis files...
Setting GRADLE_4_8_1_HOME=/home/jenkins/tools/gradle/4.8.1
[FINDBUGS] Searching for all files in <https://builds.apache.org/job/kafka-2.1-jdk8/ws/> that match the pattern **/build/reports/findbugs/*.xml
[FINDBUGS] No files found. Configuration error?
Setting GRADLE_4_8_1_HOME=/home/jenkins/tools/gradle/4.8.1
No credentials specified
Setting GRADLE_4_8_1_HOME=/home/jenkins/tools/gradle/4.8.1
<Git Blamer> Using GitBlamer to create author and commit information for all warnings.
<Git Blamer> GIT_COMMIT=3242843021933ea9a51e0633e413bc24f6c9165c, workspace=<https://builds.apache.org/job/kafka-2.1-jdk8/ws/>
[FINDBUGS] Computing warning deltas based on reference build #225
Recording test results
Setting GRADLE_4_8_1_HOME=/home/jenkins/tools/gradle/4.8.1
ERROR: Step ?Publish JUnit test result report? failed: No test report files were found. Configuration error?
Setting GRADLE_4_8_1_HOME=/home/jenkins/tools/gradle/4.8.1
Not sending mail to unregistered user bill@confluent.io
Not sending mail to unregistered user wangguoz@gmail.com
Build failed in Jenkins: kafka-2.1-jdk8 #227
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/kafka-2.1-jdk8/227/display/redirect?page=changes>
Changes:
[bill] KAFKA-8861 Fix flaky
[bill] HOTFIX: The cherry-pick for https://github.com/apache/kafka/pull/7207 to
------------------------------------------
Started by an SCM change
Started by an SCM change
[EnvInject] - Loading node environment variables.
Building remotely on H27 (ubuntu xenial) in workspace <https://builds.apache.org/job/kafka-2.1-jdk8/ws/>
[WS-CLEANUP] Deleting project workspace...
[WS-CLEANUP] Deferred wipeout is used...
No credentials specified
Cloning the remote Git repository
Cloning repository https://github.com/apache/kafka.git
> git init <https://builds.apache.org/job/kafka-2.1-jdk8/ws/> # timeout=10
Fetching upstream changes from https://github.com/apache/kafka.git
> git --version # timeout=10
> git fetch --tags --progress https://github.com/apache/kafka.git +refs/heads/*:refs/remotes/origin/*
> git config remote.origin.url https://github.com/apache/kafka.git # timeout=10
> git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
> git config remote.origin.url https://github.com/apache/kafka.git # timeout=10
Fetching upstream changes from https://github.com/apache/kafka.git
> git fetch --tags --progress https://github.com/apache/kafka.git +refs/heads/*:refs/remotes/origin/*
> git rev-parse refs/remotes/origin/2.1^{commit} # timeout=10
> git rev-parse refs/remotes/origin/origin/2.1^{commit} # timeout=10
Checking out Revision 6560203d20e818515d5b8c961fa648788c684cd6 (refs/remotes/origin/2.1)
> git config core.sparsecheckout # timeout=10
> git checkout -f 6560203d20e818515d5b8c961fa648788c684cd6
Commit message: "HOTFIX: The cherry-pick for https://github.com/apache/kafka/pull/7207 to 2.1 causes a failure in AssignedStreamsTasksTest"
> git rev-list --no-walk 8cdc2d6d1b2a227f5e62f148179a52684b230bd6 # timeout=10
Setting GRADLE_4_8_1_HOME=/home/jenkins/tools/gradle/4.8.1
Setting GRADLE_4_8_1_HOME=/home/jenkins/tools/gradle/4.8.1
[kafka-2.1-jdk8] $ /bin/bash -xe /tmp/jenkins6125630457162380728.sh
+ rm -rf <https://builds.apache.org/job/kafka-2.1-jdk8/ws/.gradle>
+ /home/jenkins/tools/gradle/4.8.1/bin/gradle
/tmp/jenkins6125630457162380728.sh: line 4: /home/jenkins/tools/gradle/4.8.1/bin/gradle: No such file or directory
Build step 'Execute shell' marked build as failure
[FINDBUGS] Collecting findbugs analysis files...
Setting GRADLE_4_8_1_HOME=/home/jenkins/tools/gradle/4.8.1
[FINDBUGS] Searching for all files in <https://builds.apache.org/job/kafka-2.1-jdk8/ws/> that match the pattern **/build/reports/findbugs/*.xml
[FINDBUGS] No files found. Configuration error?
Setting GRADLE_4_8_1_HOME=/home/jenkins/tools/gradle/4.8.1
No credentials specified
Setting GRADLE_4_8_1_HOME=/home/jenkins/tools/gradle/4.8.1
<Git Blamer> Using GitBlamer to create author and commit information for all warnings.
<Git Blamer> GIT_COMMIT=6560203d20e818515d5b8c961fa648788c684cd6, workspace=<https://builds.apache.org/job/kafka-2.1-jdk8/ws/>
[FINDBUGS] Computing warning deltas based on reference build #225
Recording test results
Setting GRADLE_4_8_1_HOME=/home/jenkins/tools/gradle/4.8.1
ERROR: Step ?Publish JUnit test result report? failed: No test report files were found. Configuration error?
Setting GRADLE_4_8_1_HOME=/home/jenkins/tools/gradle/4.8.1
Not sending mail to unregistered user bill@confluent.io
Not sending mail to unregistered user wangguoz@gmail.com