You are viewing a plain text version of this content. The canonical link for it is here.
Posted to builds@beam.apache.org by Apache Jenkins Server <je...@builds.apache.org> on 2023/01/26 19:38:44 UTC

Build failed in Jenkins: beam_PerformanceTests_Kafka_IO #3577

See <https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/3577/display/redirect>

Changes:


------------------------------------------
[...truncated 571.81 KB...]
Starting process 'command 'docker''. Working directory: <https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/src/runners/google-cloud-dataflow-java> Command: docker tag apache/beam_java8_sdk:2.46.0.dev us.gcr.io/apache-beam-testing/java-postcommit-it/java:20230126193630
Successfully started process 'command 'docker''
Starting process 'command 'gcloud''. Working directory: <https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/src/runners/google-cloud-dataflow-java> Command: gcloud docker -- push us.gcr.io/apache-beam-testing/java-postcommit-it/java:20230126193630
Successfully started process 'command 'gcloud''
WARNING: `gcloud docker` will not be supported for Docker client versions above 18.03.

As an alternative, use `gcloud auth configure-docker` to configure `docker` to
use `gcloud` as a credential helper, then use `docker` as you would for non-GCR
registries, e.g. `docker pull gcr.io/project-id/my-image`. Add
`--verbosity=error` to silence this warning: `gcloud docker
--verbosity=error -- pull gcr.io/project-id/my-image`.

See: https://cloud.google.com/container-registry/docs/support/deprecation-notices#gcloud-docker

The push refers to repository [us.gcr.io/apache-beam-testing/java-postcommit-it/java]
74f10aa70a84: Preparing
6a04190f0767: Preparing
90c278cbd2c3: Preparing
86e0adf1436e: Preparing
a67c58c87eb3: Preparing
063a7bd06c37: Preparing
3c53a4a0d694: Preparing
398fd40fd4ff: Preparing
c62df2c229fd: Preparing
0e5939309fd4: Preparing
9bbf01176cb3: Preparing
36862bdf2f63: Preparing
1d9811cfc4b9: Preparing
63584fe387d4: Preparing
80fac19d119c: Preparing
cf232862b790: Preparing
6b5aaff44254: Preparing
53a0b163e995: Preparing
0e5939309fd4: Waiting
b626401ef603: Preparing
9b55156abf26: Preparing
293d5db30c9f: Preparing
36862bdf2f63: Waiting
03127cdb479b: Preparing
9bbf01176cb3: Waiting
9c742cd6c7a5: Preparing
063a7bd06c37: Waiting
3c53a4a0d694: Waiting
398fd40fd4ff: Waiting
cf232862b790: Waiting
b626401ef603: Waiting
6b5aaff44254: Waiting
9b55156abf26: Waiting
53a0b163e995: Waiting
1d9811cfc4b9: Waiting
293d5db30c9f: Waiting
63584fe387d4: Waiting
80fac19d119c: Waiting
03127cdb479b: Waiting
9c742cd6c7a5: Waiting
6a04190f0767: Pushed
74f10aa70a84: Pushed
90c278cbd2c3: Pushed
a67c58c87eb3: Pushed
86e0adf1436e: Pushed
0e5939309fd4: Pushed
398fd40fd4ff: Pushed
3c53a4a0d694: Pushed
063a7bd06c37: Pushed
c62df2c229fd: Pushed
36862bdf2f63: Pushed
1d9811cfc4b9: Pushed
6b5aaff44254: Layer already exists
53a0b163e995: Layer already exists
b626401ef603: Layer already exists
9b55156abf26: Layer already exists
293d5db30c9f: Layer already exists
03127cdb479b: Layer already exists
9c742cd6c7a5: Layer already exists
9bbf01176cb3: Pushed
80fac19d119c: Pushed
63584fe387d4: Pushed
cf232862b790: Pushed
20230126193630: digest: sha256:d5da322790575201132280270a40a6fb2b1ea4e1fe4fff5de63f26d3fc05ecae size: 5140
:runners:google-cloud-dataflow-java:buildAndPushDockerJavaContainer (Thread[Execution **** Thread 2,5,main]) completed. Took 9.544 secs.
Resolve mutations for :runners:google-cloud-dataflow-java:testClasses (Thread[Execution **** Thread 3,5,main]) started.
Resolve mutations for :runners:google-cloud-dataflow-java:testClasses (Thread[Execution **** Thread 3,5,main]) completed. Took 0.0 secs.
:runners:google-cloud-dataflow-java:testClasses (Thread[Execution ****,5,main]) started.

> Task :runners:google-cloud-dataflow-java:testClasses
Skipping task ':runners:google-cloud-dataflow-java:testClasses' as it has no actions.
:runners:google-cloud-dataflow-java:testClasses (Thread[Execution ****,5,main]) completed. Took 0.0 secs.
Resolve mutations for :runners:google-cloud-dataflow-java:testJar (Thread[Execution **** Thread 2,5,main]) started.
Resolve mutations for :runners:google-cloud-dataflow-java:testJar (Thread[Execution **** Thread 2,5,main]) completed. Took 0.0 secs.
:runners:google-cloud-dataflow-java:testJar (Thread[Execution **** Thread 3,5,main]) started.

> Task :runners:google-cloud-dataflow-java:testJar
Caching disabled for task ':runners:google-cloud-dataflow-java:testJar' because:
  Not worth caching
Task ':runners:google-cloud-dataflow-java:testJar' is not up-to-date because:
  No history is available.
file or directory '<https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/src/runners/google-cloud-dataflow-java/build/resources/test',> not found
:runners:google-cloud-dataflow-java:testJar (Thread[Execution **** Thread 3,5,main]) completed. Took 0.042 secs.
work action resolve beam-runners-google-cloud-dataflow-java-tests.jar (project :runners:google-cloud-dataflow-java) (Thread[Execution **** Thread 7,5,main]) started.
work action null (Thread[Execution **** Thread 7,5,main]) completed. Took 0.0 secs.

> Task :runners:google-cloud-dataflow-java:****:compileJava
Note: Some input files use or override a deprecated API.
Note: Recompile with -Xlint:deprecation for details.
Note: Some input files use unchecked or unsafe operations.
Note: Recompile with -Xlint:unchecked for details.
Class dependency analysis for incremental compilation took 0.357 secs.
Created classpath snapshot for incremental compilation in 0.066 secs.
Stored cache entry for task ':runners:google-cloud-dataflow-java:****:compileJava' with cache key 6f6838b835d819f08fc8f2384714b0a7
:runners:google-cloud-dataflow-java:****:compileJava (Thread[Execution **** Thread 5,5,main]) completed. Took 1 mins 38.338 secs.
Resolve mutations for :runners:google-cloud-dataflow-java:****:classes (Thread[Execution **** Thread 7,5,main]) started.
Resolve mutations for :runners:google-cloud-dataflow-java:****:classes (Thread[Execution **** Thread 7,5,main]) completed. Took 0.0 secs.
:runners:google-cloud-dataflow-java:****:classes (Thread[Execution **** Thread 4,5,main]) started.

> Task :runners:google-cloud-dataflow-java:****:classes
Skipping task ':runners:google-cloud-dataflow-java:****:classes' as it has no actions.
:runners:google-cloud-dataflow-java:****:classes (Thread[Execution **** Thread 4,5,main]) completed. Took 0.0 secs.
Resolve mutations for :runners:google-cloud-dataflow-java:****:shadowJar (Thread[included builds,5,main]) started.
Resolve mutations for :runners:google-cloud-dataflow-java:****:shadowJar (Thread[included builds,5,main]) completed. Took 0.0 secs.
:runners:google-cloud-dataflow-java:****:shadowJar (Thread[Execution **** Thread 3,5,main]) started.

> Task :runners:google-cloud-dataflow-java:****:shadowJar
Custom actions are attached to task ':runners:google-cloud-dataflow-java:****:shadowJar'.
Build cache key for task ':runners:google-cloud-dataflow-java:****:shadowJar' is 37c5ed4b8035f31a4e8beb4160668d3c
Task ':runners:google-cloud-dataflow-java:****:shadowJar' is not up-to-date because:
  No history is available.
file or directory '<https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/src/runners/google-cloud-dataflow-java/****/build/original_sources_to_package',> not found
*******************
GRADLE SHADOW STATS

Total Jars: 15 (includes project)
Total Time: 3.376s [3376ms]
Average Time/Jar: 0.2250666666667s [225.0666666667ms]
*******************
Stored cache entry for task ':runners:google-cloud-dataflow-java:****:shadowJar' with cache key 37c5ed4b8035f31a4e8beb4160668d3c
:runners:google-cloud-dataflow-java:****:shadowJar (Thread[Execution **** Thread 3,5,main]) completed. Took 5.475 secs.
work action resolve beam-runners-google-cloud-dataflow-java-legacy-****.jar (project :runners:google-cloud-dataflow-java:****) (Thread[Execution **** Thread 4,5,main]) started.
work action null (Thread[Execution **** Thread 4,5,main]) completed. Took 0.0 secs.
work action resolve beam-runners-google-cloud-dataflow-java-legacy-****.jar (project :runners:google-cloud-dataflow-java:****) (Thread[included builds,5,main]) started.
work action null (Thread[included builds,5,main]) completed. Took 0.0 secs.
Resolve mutations for :sdks:java:io:kafka:integrationTest (Thread[Execution **** Thread 3,5,main]) started.
Resolve mutations for :sdks:java:io:kafka:integrationTest (Thread[Execution **** Thread 3,5,main]) completed. Took 0.001 secs.
:sdks:java:io:kafka:integrationTest (Thread[Execution **** Thread 4,5,main]) started.
producer locations for task group 0 (Thread[Execution **** Thread 3,5,main]) started.
producer locations for task group 0 (Thread[Execution **** Thread 3,5,main]) completed. Took 0.0 secs.
Gradle Test Executor 3 started executing tests.

> Task :sdks:java:io:kafka:integrationTest
Custom actions are attached to task ':sdks:java:io:kafka:integrationTest'.
Build cache key for task ':sdks:java:io:kafka:integrationTest' is 17774cdff3ef923498b0bf74bb48d50c
Task ':sdks:java:io:kafka:integrationTest' is not up-to-date because:
  Task.upToDateWhen is false.
Starting process 'Gradle Test Executor 3'. Working directory: <https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/src/sdks/java/io/kafka> Command: /usr/lib/jvm/java-8-openjdk-amd64/bin/java -DbeamTestPipelineOptions=["--tempRoot=gs://temp-storage-for-perf-tests","--project=apache-beam-testing","--runner=DataflowRunner","--usePublicIPs=false","--sourceOptions={\"numRecords\":\"100000000\",\"keySizeBytes\":\"10\",\"valueSizeBytes\":\"90\"}","--bigQueryDataset=beam_performance","--bigQueryTable=kafkaioit_results_runner_v2","--influxMeasurement=kafkaioit_results_runner_v2","--influxDatabase=beam_test_metrics","--influxHost=http://10.128.0.96:8086","--kafkaBootstrapServerAddresses=$NODE_IP_0:32401,$NODE_IP_1:32402,$NODE_IP_2:32403","--kafkaTopic=beam-sdf","--readTimeout=1500","--numWorkers=5","--autoscalingAlgorithm=NONE","--experiments=use_runner_v2,use_unified_****","--sdkContainerImage=us.gcr.io/apache-beam-testing/java-postcommit-it/java:20230126193630","--region=us-central1"] -Djava.security.manager=****.org.gradle.process.internal.****.child.BootstrapSecurityManager -Dorg.gradle.internal.****.tmpdir=<https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/src/sdks/java/io/kafka/build/tmp/integrationTest/work> -Dorg.gradle.native=false -Xmx2g -Dfile.encoding=UTF-8 -Duser.country=US -Duser.language=en -Duser.variant -ea -cp /home/jenkins/.gradle/caches/7.5.1/****Main/gradle-****.jar ****.org.gradle.process.internal.****.GradleWorkerMain 'Gradle Test Executor 3'
Successfully started process 'Gradle Test Executor 3'

org.apache.beam.sdk.io.kafka.KafkaIOIT STANDARD_ERROR
    SLF4J: Class path contains multiple SLF4J bindings.
    SLF4J: Found binding in [jar:<https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/src/runners/google-cloud-dataflow-java/****/build/libs/beam-runners-google-cloud-dataflow-java-legacy-****-2.46.0-SNAPSHOT.jar!/org/slf4j/impl/StaticLoggerBinder.class]>
    SLF4J: Found binding in [jar:file:/home/jenkins/.gradle/caches/modules-2/files-2.1/org.slf4j/slf4j-jdk14/1.7.30/d35953dd2fe54ebe39fdf18cfd82fe6eb35b25ed/slf4j-jdk14-1.7.30.jar!/org/slf4j/impl/StaticLoggerBinder.class]
    SLF4J: Found binding in [jar:file:/home/jenkins/.gradle/caches/modules-2/files-2.1/org.slf4j/slf4j-simple/1.7.30/e606eac955f55ecf1d8edcccba04eb8ac98088dd/slf4j-simple-1.7.30.jar!/org/slf4j/impl/StaticLoggerBinder.class]
    SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation.
    SLF4J: Actual binding is of type [org.slf4j.impl.JDK14LoggerFactory]

Gradle Test Executor 3 finished executing tests.

> Task :sdks:java:io:kafka:integrationTest

org.apache.beam.sdk.io.kafka.KafkaIOIT > classMethod FAILED
    java.lang.IllegalArgumentException: Class interface org.apache.beam.sdk.testing.TestPipelineOptions missing a property named 'usePublicIPs'. Did you mean 'usePublicIps'?
        at org.apache.beam.sdk.options.PipelineOptionsFactory.parseObjects(PipelineOptionsFactory.java:1888)
        at org.apache.beam.sdk.options.PipelineOptionsFactory.access$400(PipelineOptionsFactory.java:146)
        at org.apache.beam.sdk.options.PipelineOptionsFactory$Builder.as(PipelineOptionsFactory.java:329)
        at org.apache.beam.sdk.testing.TestPipeline.testingPipelineOptions(TestPipeline.java:500)
        at org.apache.beam.sdk.io.common.IOITHelper.readIOTestPipelineOptions(IOITHelper.java:48)
        at org.apache.beam.sdk.io.kafka.KafkaIOIT.setup(KafkaIOIT.java:174)

1 test completed, 1 failed
Finished generating test XML results (0.025 secs) into: <https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/src/sdks/java/io/kafka/build/test-results/integrationTest>
Generating HTML test report...
Finished generating test html results (0.03 secs) into: <https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/src/sdks/java/io/kafka/build/reports/tests/integrationTest>

> Task :sdks:java:io:kafka:integrationTest FAILED
:sdks:java:io:kafka:integrationTest (Thread[Execution **** Thread 4,5,main]) completed. Took 4.595 secs.
Resolve mutations for :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages (Thread[Execution **** Thread 2,5,main]) started.
Resolve mutations for :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages (Thread[Execution **** Thread 2,5,main]) completed. Took 0.0 secs.
:runners:google-cloud-dataflow-java:cleanUpDockerJavaImages (Thread[Execution **** Thread 6,5,main]) started.

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages
Custom actions are attached to task ':runners:google-cloud-dataflow-java:cleanUpDockerJavaImages'.
Caching disabled for task ':runners:google-cloud-dataflow-java:cleanUpDockerJavaImages' because:
  Gradle would require more information to cache this task
Task ':runners:google-cloud-dataflow-java:cleanUpDockerJavaImages' is not up-to-date because:
  Task has not declared any outputs despite executing actions.
Starting process 'command 'docker''. Working directory: <https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/src/runners/google-cloud-dataflow-java> Command: docker rmi --force us.gcr.io/apache-beam-testing/java-postcommit-it/java:20230126193630
Successfully started process 'command 'docker''
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20230126193630
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:d5da322790575201132280270a40a6fb2b1ea4e1fe4fff5de63f26d3fc05ecae
Starting process 'command 'gcloud''. Working directory: <https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/src/runners/google-cloud-dataflow-java> Command: gcloud --quiet container images untag us.gcr.io/apache-beam-testing/java-postcommit-it/java:20230126193630
Successfully started process 'command 'gcloud''
WARNING: Successfully resolved tag to sha256, but it is recommended to use sha256 directly.
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20230126193630]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:d5da322790575201132280270a40a6fb2b1ea4e1fe4fff5de63f26d3fc05ecae]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20230126193630] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:d5da322790575201132280270a40a6fb2b1ea4e1fe4fff5de63f26d3fc05ecae])].
Starting process 'command './scripts/cleanup_untagged_gcr_images.sh''. Working directory: <https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/src/runners/google-cloud-dataflow-java> Command: ./scripts/cleanup_untagged_gcr_images.sh us.gcr.io/apache-beam-testing/java-postcommit-it/java
Successfully started process 'command './scripts/cleanup_untagged_gcr_images.sh''
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:d5da322790575201132280270a40a6fb2b1ea4e1fe4fff5de63f26d3fc05ecae
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:d5da322790575201132280270a40a6fb2b1ea4e1fe4fff5de63f26d3fc05ecae
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:d5da322790575201132280270a40a6fb2b1ea4e1fe4fff5de63f26d3fc05ecae].
:runners:google-cloud-dataflow-java:cleanUpDockerJavaImages (Thread[Execution **** Thread 6,5,main]) completed. Took 4.695 secs.
Resolve mutations for :sdks:java:io:kafka:cleanUp (Thread[Execution **** Thread 4,5,main]) started.
Resolve mutations for :sdks:java:io:kafka:cleanUp (Thread[Execution **** Thread 4,5,main]) completed. Took 0.0 secs.
:sdks:java:io:kafka:cleanUp (Thread[Execution **** Thread 2,5,main]) started.

> Task :sdks:java:io:kafka:cleanUp
Skipping task ':sdks:java:io:kafka:cleanUp' as it has no actions.
:sdks:java:io:kafka:cleanUp (Thread[Execution **** Thread 2,5,main]) completed. Took 0.0 secs.

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:io:kafka:integrationTest'.
> There were failing tests. See the report at: file://<https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/src/sdks/java/io/kafka/build/reports/tests/integrationTest/index.html>

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --debug option to get more log output.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.5.1/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 2m 31s
158 actionable tasks: 103 executed, 51 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/4utnodniltchy

Stopped 2 **** daemon(s).
Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Jenkins build is back to normal : beam_PerformanceTests_Kafka_IO #3582

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/3582/display/redirect?page=changes>


---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PerformanceTests_Kafka_IO #3581

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/3581/display/redirect>

Changes:


------------------------------------------
[...truncated 9.31 KB...]
+ read -r usedPort
+ '[' 32402 = 32383 ']'
+ IFS=
+ read -r usedPort
+ false
+ echo 32402
+ return 0
[EnvInject] - Injecting environment variables from a build step.
[EnvInject] - Injecting as environment variables the properties file path 'job.properties'
[EnvInject] - Variables injected successfully.
[beam_PerformanceTests_Kafka_IO] $ /bin/bash -xe /tmp/jenkins8646096636002379869.sh
+ set -xo pipefail
+ eval <https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/src/.test-infra/kubernetes/kubernetes.sh> getAvailablePort 32402 32767
++ <https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/src/.test-infra/kubernetes/kubernetes.sh> getAvailablePort 32402 32767
+ KUBECONFIG=<https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/config-beam-performancetests-kafka-io-3581>
+ KUBERNETES_NAMESPACE=beam-performancetests-kafka-io-3581
+ KUBECTL='kubectl --kubeconfig=<https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/config-beam-performancetests-kafka-io-3581> --namespace=beam-performancetests-kafka-io-3581'
+ getAvailablePort 32402 32767
+ local lowRangePort=32403
+ local highRangePort=32767
+ local used=false
+ local 'command=kubectl --kubeconfig=<https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/config-beam-performancetests-kafka-io-3581> --namespace=beam-performancetests-kafka-io-3581 get svc --all-namespaces -o   go-template='\''{{range .items}}{{range.spec.ports}}{{if .nodePort}}{{.nodePort}}{{"\n"}}{{end}}{{end}}{{end}}'\'''
+ local usedPorts
+ sed 's/^/KAFKA_SERVICE_PORT_2=/'
++ eval 'kubectl --kubeconfig=<https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/config-beam-performancetests-kafka-io-3581> --namespace=beam-performancetests-kafka-io-3581 get svc --all-namespaces -o   go-template='\''{{range .items}}{{range.spec.ports}}{{if .nodePort}}{{.nodePort}}{{"\n"}}{{end}}{{end}}{{end}}'\'''
+++ kubectl --kubeconfig=<https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/config-beam-performancetests-kafka-io-3581> --namespace=beam-performancetests-kafka-io-3581 get svc --all-namespaces -o 'go-template={{range .items}}{{range.spec.ports}}{{if .nodePort}}{{.nodePort}}{{"\n"}}{{end}}{{end}}{{end}}'
+ usedPorts='31977
30409
32383'
+ local availablePort=32403
++ seq 32403 32767
+ for i in $(seq $lowRangePort $highRangePort)
+ IFS=
+ read -r usedPort
+ '[' 32403 = 31977 ']'
+ IFS=
+ read -r usedPort
+ '[' 32403 = 30409 ']'
+ IFS=
+ read -r usedPort
+ '[' 32403 = 32383 ']'
+ IFS=
+ read -r usedPort
+ false
+ echo 32403
+ return 0
[EnvInject] - Injecting environment variables from a build step.
[EnvInject] - Injecting as environment variables the properties file path 'job.properties'
[EnvInject] - Variables injected successfully.
[beam_PerformanceTests_Kafka_IO] $ /bin/bash -xe /tmp/jenkins7139051424248758724.sh
+ sed -i -e s/32400/32401/ <https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/src/.test-infra/kubernetes/kafka-cluster/04-outside-services/outside-0.yml>
[beam_PerformanceTests_Kafka_IO] $ /bin/bash -xe /tmp/jenkins4570473093850874041.sh
+ sed -i -e s/32401/32402/ <https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/src/.test-infra/kubernetes/kafka-cluster/04-outside-services/outside-1.yml>
[beam_PerformanceTests_Kafka_IO] $ /bin/bash -xe /tmp/jenkins1917301965231362792.sh
+ sed -i -e s/32402/32403/ <https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/src/.test-infra/kubernetes/kafka-cluster/04-outside-services/outside-2.yml>
[beam_PerformanceTests_Kafka_IO] $ /bin/bash -xe /tmp/jenkins8044922198775079476.sh
+ <https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/src/.test-infra/kubernetes/kubernetes.sh> apply <https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/src/.test-infra/kubernetes/kafka-cluster>
+ KUBECONFIG=<https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/config-beam-performancetests-kafka-io-3581>
+ KUBERNETES_NAMESPACE=beam-performancetests-kafka-io-3581
+ KUBECTL='kubectl --kubeconfig=<https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/config-beam-performancetests-kafka-io-3581> --namespace=beam-performancetests-kafka-io-3581'
+ apply <https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/src/.test-infra/kubernetes/kafka-cluster>
+ eval 'kubectl --kubeconfig=<https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/config-beam-performancetests-kafka-io-3581> --namespace=beam-performancetests-kafka-io-3581 apply -R -f <https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/src/.test-infra/kubernetes/kafka-cluster'>
++ kubectl --kubeconfig=<https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/config-beam-performancetests-kafka-io-3581> --namespace=beam-performancetests-kafka-io-3581 apply -R -f <https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/src/.test-infra/kubernetes/kafka-cluster>
storageclass.storage.k8s.io/kafka-broker unchanged
storageclass.storage.k8s.io/kafka-zookeeper unchanged
clusterrole.rbac.authorization.k8s.io/node-reader unchanged
clusterrolebinding.rbac.authorization.k8s.io/kafka-node-reader configured
role.rbac.authorization.k8s.io/pod-labler created
rolebinding.rbac.authorization.k8s.io/kafka-pod-labler created
configmap/zookeeper-config created
service/pzoo created
service/zookeeper created
statefulset.apps/pzoo created
service/outside-0 created
service/outside-1 created
service/outside-2 created
configmap/broker-config created
service/broker created
service/bootstrap created
statefulset.apps/kafka created
configmap/kafka-config created
job.batch/kafka-config-eff079ec created
[beam_PerformanceTests_Kafka_IO] $ /bin/bash -xe /tmp/jenkins7909273448409796808.sh
+ set -eo pipefail
+ sed 's/^/KAFKA_BROKER_0=/'
+ eval <https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/src/.test-infra/kubernetes/kubernetes.sh> loadBalancerIP outside-0
++ <https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/src/.test-infra/kubernetes/kubernetes.sh> loadBalancerIP outside-0
+ KUBECONFIG=<https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/config-beam-performancetests-kafka-io-3581>
+ KUBERNETES_NAMESPACE=beam-performancetests-kafka-io-3581
+ KUBECTL='kubectl --kubeconfig=<https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/config-beam-performancetests-kafka-io-3581> --namespace=beam-performancetests-kafka-io-3581'
+ loadBalancerIP outside-0
+ local name=outside-0
+ local 'command=kubectl --kubeconfig=<https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/config-beam-performancetests-kafka-io-3581> --namespace=beam-performancetests-kafka-io-3581 get svc outside-0 -ojsonpath='\''{.status.loadBalancer.ingress[0].ip}'\'''
+ retry 'kubectl --kubeconfig=<https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/config-beam-performancetests-kafka-io-3581> --namespace=beam-performancetests-kafka-io-3581 get svc outside-0 -ojsonpath='\''{.status.loadBalancer.ingress[0].ip}'\''' 36 10
+ local 'command=kubectl --kubeconfig=<https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/config-beam-performancetests-kafka-io-3581> --namespace=beam-performancetests-kafka-io-3581 get svc outside-0 -ojsonpath='\''{.status.loadBalancer.ingress[0].ip}'\'''
+ local max_retries=36
+ local sleep_time=10
+ (( i = 1 ))
+ (( i <= max_retries ))
+ local output
++ eval 'kubectl --kubeconfig=<https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/config-beam-performancetests-kafka-io-3581> --namespace=beam-performancetests-kafka-io-3581 get svc outside-0 -ojsonpath='\''{.status.loadBalancer.ingress[0].ip}'\'''
+++ kubectl --kubeconfig=<https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/config-beam-performancetests-kafka-io-3581> --namespace=beam-performancetests-kafka-io-3581 get svc outside-0 '-ojsonpath={.status.loadBalancer.ingress[0].ip}'
+ output=
+ local status=0
+ [[ 0 == 0 ]]
+ [[ -n '' ]]
+ [[ 1 == \3\6 ]]
+ sleep 10
+ (( i++ ))
+ (( i <= max_retries ))
+ local output
++ eval 'kubectl --kubeconfig=<https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/config-beam-performancetests-kafka-io-3581> --namespace=beam-performancetests-kafka-io-3581 get svc outside-0 -ojsonpath='\''{.status.loadBalancer.ingress[0].ip}'\'''
+++ kubectl --kubeconfig=<https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/config-beam-performancetests-kafka-io-3581> --namespace=beam-performancetests-kafka-io-3581 get svc outside-0 '-ojsonpath={.status.loadBalancer.ingress[0].ip}'
+ output=
+ local status=0
+ [[ 0 == 0 ]]
+ [[ -n '' ]]
+ [[ 2 == \3\6 ]]
+ sleep 10
+ (( i++ ))
+ (( i <= max_retries ))
+ local output
++ eval 'kubectl --kubeconfig=<https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/config-beam-performancetests-kafka-io-3581> --namespace=beam-performancetests-kafka-io-3581 get svc outside-0 -ojsonpath='\''{.status.loadBalancer.ingress[0].ip}'\'''
+++ kubectl --kubeconfig=<https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/config-beam-performancetests-kafka-io-3581> --namespace=beam-performancetests-kafka-io-3581 get svc outside-0 '-ojsonpath={.status.loadBalancer.ingress[0].ip}'
+ output=
+ local status=0
+ [[ 0 == 0 ]]
+ [[ -n '' ]]
+ [[ 3 == \3\6 ]]
+ sleep 10
+ (( i++ ))
+ (( i <= max_retries ))
+ local output
++ eval 'kubectl --kubeconfig=<https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/config-beam-performancetests-kafka-io-3581> --namespace=beam-performancetests-kafka-io-3581 get svc outside-0 -ojsonpath='\''{.status.loadBalancer.ingress[0].ip}'\'''
+++ kubectl --kubeconfig=<https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/config-beam-performancetests-kafka-io-3581> --namespace=beam-performancetests-kafka-io-3581 get svc outside-0 '-ojsonpath={.status.loadBalancer.ingress[0].ip}'
+ output=
+ local status=0
+ [[ 0 == 0 ]]
+ [[ -n '' ]]
+ [[ 4 == \3\6 ]]
+ sleep 10
+ (( i++ ))
+ (( i <= max_retries ))
+ local output
++ eval 'kubectl --kubeconfig=<https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/config-beam-performancetests-kafka-io-3581> --namespace=beam-performancetests-kafka-io-3581 get svc outside-0 -ojsonpath='\''{.status.loadBalancer.ingress[0].ip}'\'''
+++ kubectl --kubeconfig=<https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/config-beam-performancetests-kafka-io-3581> --namespace=beam-performancetests-kafka-io-3581 get svc outside-0 '-ojsonpath={.status.loadBalancer.ingress[0].ip}'
+ output=
+ local status=0
+ [[ 0 == 0 ]]
+ [[ -n '' ]]
+ [[ 5 == \3\6 ]]
+ sleep 10
+ (( i++ ))
+ (( i <= max_retries ))
+ local output
++ eval 'kubectl --kubeconfig=<https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/config-beam-performancetests-kafka-io-3581> --namespace=beam-performancetests-kafka-io-3581 get svc outside-0 -ojsonpath='\''{.status.loadBalancer.ingress[0].ip}'\'''
+++ kubectl --kubeconfig=<https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/config-beam-performancetests-kafka-io-3581> --namespace=beam-performancetests-kafka-io-3581 get svc outside-0 '-ojsonpath={.status.loadBalancer.ingress[0].ip}'
+ output=
+ local status=0
+ [[ 0 == 0 ]]
+ [[ -n '' ]]
+ [[ 6 == \3\6 ]]
+ sleep 10
+ (( i++ ))
+ (( i <= max_retries ))
+ local output
++ eval 'kubectl --kubeconfig=<https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/config-beam-performancetests-kafka-io-3581> --namespace=beam-performancetests-kafka-io-3581 get svc outside-0 -ojsonpath='\''{.status.loadBalancer.ingress[0].ip}'\'''
+++ kubectl --kubeconfig=<https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/config-beam-performancetests-kafka-io-3581> --namespace=beam-performancetests-kafka-io-3581 get svc outside-0 '-ojsonpath={.status.loadBalancer.ingress[0].ip}'
+ output=34.28.75.192
+ local status=0
+ [[ 0 == 0 ]]
+ [[ -n 34.28.75.192 ]]
+ echo 34.28.75.192
+ return 0
[EnvInject] - Injecting environment variables from a build step.
[EnvInject] - Injecting as environment variables the properties file path 'job.properties'
[EnvInject] - Variables injected successfully.
[beam_PerformanceTests_Kafka_IO] $ /bin/bash -xe /tmp/jenkins3315193389621810233.sh
+ set -eo pipefail
+ eval <https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/src/.test-infra/kubernetes/kubernetes.sh> loadBalancerIP outside-1
++ <https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/src/.test-infra/kubernetes/kubernetes.sh> loadBalancerIP outside-1
+ sed 's/^/KAFKA_BROKER_1=/'
+ KUBECONFIG=<https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/config-beam-performancetests-kafka-io-3581>
+ KUBERNETES_NAMESPACE=beam-performancetests-kafka-io-3581
+ KUBECTL='kubectl --kubeconfig=<https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/config-beam-performancetests-kafka-io-3581> --namespace=beam-performancetests-kafka-io-3581'
+ loadBalancerIP outside-1
+ local name=outside-1
+ local 'command=kubectl --kubeconfig=<https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/config-beam-performancetests-kafka-io-3581> --namespace=beam-performancetests-kafka-io-3581 get svc outside-1 -ojsonpath='\''{.status.loadBalancer.ingress[0].ip}'\'''
+ retry 'kubectl --kubeconfig=<https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/config-beam-performancetests-kafka-io-3581> --namespace=beam-performancetests-kafka-io-3581 get svc outside-1 -ojsonpath='\''{.status.loadBalancer.ingress[0].ip}'\''' 36 10
+ local 'command=kubectl --kubeconfig=<https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/config-beam-performancetests-kafka-io-3581> --namespace=beam-performancetests-kafka-io-3581 get svc outside-1 -ojsonpath='\''{.status.loadBalancer.ingress[0].ip}'\'''
+ local max_retries=36
+ local sleep_time=10
+ (( i = 1 ))
+ (( i <= max_retries ))
+ local output
++ eval 'kubectl --kubeconfig=<https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/config-beam-performancetests-kafka-io-3581> --namespace=beam-performancetests-kafka-io-3581 get svc outside-1 -ojsonpath='\''{.status.loadBalancer.ingress[0].ip}'\'''
+++ kubectl --kubeconfig=<https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/config-beam-performancetests-kafka-io-3581> --namespace=beam-performancetests-kafka-io-3581 get svc outside-1 '-ojsonpath={.status.loadBalancer.ingress[0].ip}'
+ output=34.170.25.67
+ local status=0
+ [[ 0 == 0 ]]
+ [[ -n 34.170.25.67 ]]
+ echo 34.170.25.67
+ return 0
[EnvInject] - Injecting environment variables from a build step.
[EnvInject] - Injecting as environment variables the properties file path 'job.properties'
[EnvInject] - Variables injected successfully.
[beam_PerformanceTests_Kafka_IO] $ /bin/bash -xe /tmp/jenkins6056950572421862507.sh
+ set -eo pipefail
+ eval <https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/src/.test-infra/kubernetes/kubernetes.sh> loadBalancerIP outside-2
++ <https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/src/.test-infra/kubernetes/kubernetes.sh> loadBalancerIP outside-2
+ sed 's/^/KAFKA_BROKER_2=/'
+ KUBECONFIG=<https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/config-beam-performancetests-kafka-io-3581>
+ KUBERNETES_NAMESPACE=beam-performancetests-kafka-io-3581
+ KUBECTL='kubectl --kubeconfig=<https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/config-beam-performancetests-kafka-io-3581> --namespace=beam-performancetests-kafka-io-3581'
+ loadBalancerIP outside-2
+ local name=outside-2
+ local 'command=kubectl --kubeconfig=<https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/config-beam-performancetests-kafka-io-3581> --namespace=beam-performancetests-kafka-io-3581 get svc outside-2 -ojsonpath='\''{.status.loadBalancer.ingress[0].ip}'\'''
+ retry 'kubectl --kubeconfig=<https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/config-beam-performancetests-kafka-io-3581> --namespace=beam-performancetests-kafka-io-3581 get svc outside-2 -ojsonpath='\''{.status.loadBalancer.ingress[0].ip}'\''' 36 10
+ local 'command=kubectl --kubeconfig=<https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/config-beam-performancetests-kafka-io-3581> --namespace=beam-performancetests-kafka-io-3581 get svc outside-2 -ojsonpath='\''{.status.loadBalancer.ingress[0].ip}'\'''
+ local max_retries=36
+ local sleep_time=10
+ (( i = 1 ))
+ (( i <= max_retries ))
+ local output
++ eval 'kubectl --kubeconfig=<https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/config-beam-performancetests-kafka-io-3581> --namespace=beam-performancetests-kafka-io-3581 get svc outside-2 -ojsonpath='\''{.status.loadBalancer.ingress[0].ip}'\'''
+++ kubectl --kubeconfig=<https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/config-beam-performancetests-kafka-io-3581> --namespace=beam-performancetests-kafka-io-3581 get svc outside-2 '-ojsonpath={.status.loadBalancer.ingress[0].ip}'
+ output=35.193.104.48
+ local status=0
+ [[ 0 == 0 ]]
+ [[ -n 35.193.104.48 ]]
+ echo 35.193.104.48
+ return 0
[EnvInject] - Injecting environment variables from a build step.
[EnvInject] - Injecting as environment variables the properties file path 'job.properties'
[EnvInject] - Variables injected successfully.
[beam_PerformanceTests_Kafka_IO] $ /bin/bash -xe /tmp/jenkins8052996834254211554.sh
+ eval <https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/src/.test-infra/kubernetes/kubernetes.sh> waitForJob job.batch/kafka-config-eff079ec 40m
++ <https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/src/.test-infra/kubernetes/kubernetes.sh> waitForJob job.batch/kafka-config-eff079ec 40m
+ KUBECONFIG=<https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/config-beam-performancetests-kafka-io-3581>
+ KUBERNETES_NAMESPACE=beam-performancetests-kafka-io-3581
+ KUBECTL='kubectl --kubeconfig=<https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/config-beam-performancetests-kafka-io-3581> --namespace=beam-performancetests-kafka-io-3581'
+ waitForJob job.batch/kafka-config-eff079ec 40m
+ echo 'Waiting for job completion...'
Waiting for job completion...
+ jobName=job.batch/kafka-config-eff079ec
+ eval 'kubectl --kubeconfig=<https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/config-beam-performancetests-kafka-io-3581> --namespace=beam-performancetests-kafka-io-3581 wait --for=condition=complete --timeout=40m job.batch/kafka-config-eff079ec'
++ kubectl --kubeconfig=<https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/config-beam-performancetests-kafka-io-3581> --namespace=beam-performancetests-kafka-io-3581 wait --for=condition=complete --timeout=40m job.batch/kafka-config-eff079ec
error: timed out waiting for the condition on jobs/kafka-config-eff079ec
Build step 'Execute shell' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PerformanceTests_Kafka_IO #3580

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/3580/display/redirect?page=changes>

Changes:

[nflavour] Initial commit of boilerplate setup of change stream pipeline for

[ahmedabualsaud] Fix SchemaTransform identifiers

[noreply] Update chromedriver-binary requirement in /sdks/python (#25178)

[noreply] Bump google.golang.org/grpc from 1.52.0 to 1.52.3 in /sdks (#25181)

[noreply] Fix a couple typos caught by an internal linter (#25188)

[noreply] Update Dataflow container versions (#25192)


------------------------------------------
[...truncated 953.61 KB...]
    INFO: Adding PAssert$0/GroupGlobally/Reify.Window/ParDo(Anonymous) as step s16
    Jan 27, 2023 1:15:07 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding PAssert$0/GroupGlobally/ParDo(ToSingletonIterables) as step s17
    Jan 27, 2023 1:15:07 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding PAssert$0/GroupGlobally/Create.Values/Read(CreateSource) as step s18
    Jan 27, 2023 1:15:07 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding PAssert$0/GroupGlobally/Flatten.PCollections as step s19
    Jan 27, 2023 1:15:07 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding PAssert$0/GroupGlobally/Window.Into()/Flatten.PCollections as step s20
    Jan 27, 2023 1:15:07 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding PAssert$0/GroupGlobally/WithKeys/AddKeys/Map as step s21
    Jan 27, 2023 1:15:07 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding PAssert$0/GroupGlobally/GroupByKey as step s22
    Jan 27, 2023 1:15:07 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding PAssert$0/GroupGlobally/Values/Values/Map as step s23
    Jan 27, 2023 1:15:07 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding PAssert$0/GroupGlobally/ParDo(Concat) as step s24
    Jan 27, 2023 1:15:07 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding PAssert$0/GetPane/Map as step s25
    Jan 27, 2023 1:15:07 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding PAssert$0/RunChecks as step s26
    Jan 27, 2023 1:15:07 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding PAssert$0/VerifyAssertions/ParDo(DefaultConclude) as step s27
    Jan 27, 2023 1:15:07 AM org.apache.beam.runners.dataflow.DataflowRunner run
    INFO: Dataflow SDK version: 2.46.0-SNAPSHOT
    Jan 27, 2023 1:15:07 AM org.apache.beam.runners.dataflow.DataflowRunner run
    INFO: To access the Dataflow monitoring console, please navigate to https://console.cloud.google.com/dataflow/jobs/us-central1/2023-01-26_17_15_07-11351705346048958909?project=apache-beam-testing
    Jan 27, 2023 1:15:07 AM org.apache.beam.runners.dataflow.DataflowRunner run
    INFO: Submitted job: 2023-01-26_17_15_07-11351705346048958909
    Jan 27, 2023 1:15:07 AM org.apache.beam.runners.dataflow.DataflowRunner run
    INFO: To cancel the job using the 'gcloud' tool, run:
    > gcloud dataflow jobs --project=apache-beam-testing cancel --region=us-central1 2023-01-26_17_15_07-11351705346048958909
    Jan 27, 2023 1:15:11 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    WARNING: 2023-01-27T01:15:10.716Z: The workflow name is not a valid Cloud Label. Labels applied to Cloud resources (such as GCE Instances) for monitoring will be labeled with this modified job name: kafkaioit0testkafkaioreadsandwritescorrectlyinbatch-jenkin-yd73. For the best monitoring experience, please name your job with a valid Cloud Label. For details, see: https://cloud.google.com/compute/docs/labeling-resources#restrictions
    Jan 27, 2023 1:15:24 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2023-01-27T01:15:23.707Z: Worker configuration: e2-standard-2 in us-central1-b.
    Jan 27, 2023 1:15:27 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2023-01-27T01:15:25.140Z: Expanding CoGroupByKey operations into optimizable parts.
    Jan 27, 2023 1:15:27 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2023-01-27T01:15:25.269Z: Expanding GroupByKey operations into optimizable parts.
    Jan 27, 2023 1:15:27 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2023-01-27T01:15:25.299Z: Lifting ValueCombiningMappingFns into MergeBucketsMappingFns
    Jan 27, 2023 1:15:27 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2023-01-27T01:15:25.472Z: Fusing adjacent ParDo, Read, Write, and Flatten operations
    Jan 27, 2023 1:15:27 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2023-01-27T01:15:25.494Z: Elided trivial flatten 
    Jan 27, 2023 1:15:27 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2023-01-27T01:15:25.526Z: Unzipping flatten s19 for input s17.org.apache.beam.sdk.values.PCollection.<init>:405#e63ad18dfc139ac0
    Jan 27, 2023 1:15:27 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2023-01-27T01:15:25.551Z: Fusing unzipped copy of PAssert$0/GroupGlobally/WithKeys/AddKeys/Map, through flatten PAssert$0/GroupGlobally/Flatten.PCollections, into producer PAssert$0/GroupGlobally/ParDo(ToSingletonIterables)
    Jan 27, 2023 1:15:27 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2023-01-27T01:15:25.585Z: Fusing consumer PAssert$0/GroupGlobally/GroupByKey/Write into PAssert$0/GroupGlobally/GroupByKey/Reify
    Jan 27, 2023 1:15:27 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2023-01-27T01:15:25.618Z: Fusing consumer PAssert$0/GroupGlobally/GroupByKey/GroupByWindow into PAssert$0/GroupGlobally/GroupByKey/Read
    Jan 27, 2023 1:15:27 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2023-01-27T01:15:25.649Z: Fusing consumer PAssert$0/GroupGlobally/Values/Values/Map into PAssert$0/GroupGlobally/GroupByKey/GroupByWindow
    Jan 27, 2023 1:15:27 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2023-01-27T01:15:25.685Z: Fusing consumer PAssert$0/GroupGlobally/ParDo(Concat) into PAssert$0/GroupGlobally/Values/Values/Map
    Jan 27, 2023 1:15:27 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2023-01-27T01:15:25.720Z: Fusing consumer PAssert$0/GetPane/Map into PAssert$0/GroupGlobally/ParDo(Concat)
    Jan 27, 2023 1:15:27 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2023-01-27T01:15:25.798Z: Fusing consumer PAssert$0/RunChecks into PAssert$0/GetPane/Map
    Jan 27, 2023 1:15:27 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2023-01-27T01:15:25.833Z: Fusing consumer PAssert$0/VerifyAssertions/ParDo(DefaultConclude) into PAssert$0/RunChecks
    Jan 27, 2023 1:15:27 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2023-01-27T01:15:25.867Z: Unzipping flatten s19-u40 for input s21.org.apache.beam.sdk.values.PCollection.<init>:405#24aa17775bf242ac-c38
    Jan 27, 2023 1:15:27 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2023-01-27T01:15:25.900Z: Fusing unzipped copy of PAssert$0/GroupGlobally/GroupByKey/Reify, through flatten PAssert$0/GroupGlobally/Flatten.PCollections/Unzipped-1, into producer PAssert$0/GroupGlobally/WithKeys/AddKeys/Map
    Jan 27, 2023 1:15:27 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2023-01-27T01:15:25.933Z: Fusing consumer Read from bounded Kafka/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Split into Read from bounded Kafka/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Create/Read(CreateSource)
    Jan 27, 2023 1:15:27 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2023-01-27T01:15:25.969Z: Fusing consumer Read from bounded Kafka/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Reshuffle/Pair with random key into Read from bounded Kafka/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Split
    Jan 27, 2023 1:15:27 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2023-01-27T01:15:26.004Z: Fusing consumer Read from bounded Kafka/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Reshuffle/Reshuffle/Window.Into()/Window.Assign into Read from bounded Kafka/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Reshuffle/Pair with random key
    Jan 27, 2023 1:15:27 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2023-01-27T01:15:26.036Z: Fusing consumer Read from bounded Kafka/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Reshuffle/Reshuffle/GroupByKey/Reify into Read from bounded Kafka/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Reshuffle/Reshuffle/Window.Into()/Window.Assign
    Jan 27, 2023 1:15:27 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2023-01-27T01:15:26.068Z: Fusing consumer Read from bounded Kafka/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Reshuffle/Reshuffle/GroupByKey/Write into Read from bounded Kafka/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Reshuffle/Reshuffle/GroupByKey/Reify
    Jan 27, 2023 1:15:27 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2023-01-27T01:15:26.089Z: Fusing consumer Read from bounded Kafka/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Reshuffle/Reshuffle/GroupByKey/GroupByWindow into Read from bounded Kafka/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Reshuffle/Reshuffle/GroupByKey/Read
    Jan 27, 2023 1:15:27 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2023-01-27T01:15:26.121Z: Fusing consumer Read from bounded Kafka/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Reshuffle/Reshuffle/ExpandIterable into Read from bounded Kafka/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Reshuffle/Reshuffle/GroupByKey/GroupByWindow
    Jan 27, 2023 1:15:27 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2023-01-27T01:15:26.151Z: Fusing consumer Read from bounded Kafka/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Reshuffle/Values/Values/Map into Read from bounded Kafka/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Reshuffle/Reshuffle/ExpandIterable
    Jan 27, 2023 1:15:27 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2023-01-27T01:15:26.183Z: Fusing consumer Read from bounded Kafka/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Read into Read from bounded Kafka/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Reshuffle/Values/Values/Map
    Jan 27, 2023 1:15:27 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2023-01-27T01:15:26.207Z: Fusing consumer Read from bounded Kafka/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/StripIds into Read from bounded Kafka/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Read
    Jan 27, 2023 1:15:27 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2023-01-27T01:15:26.238Z: Fusing consumer Measure read time into Read from bounded Kafka/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/StripIds
    Jan 27, 2023 1:15:27 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2023-01-27T01:15:26.271Z: Fusing consumer Map records to strings/Map into Measure read time
    Jan 27, 2023 1:15:27 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2023-01-27T01:15:26.297Z: Fusing consumer Calculate hashcode/WithKeys/AddKeys/Map into Map records to strings/Map
    Jan 27, 2023 1:15:27 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2023-01-27T01:15:26.333Z: Fusing consumer Calculate hashcode/Combine.perKey(Hashing)/GroupByKey+Calculate hashcode/Combine.perKey(Hashing)/Combine.GroupedValues/Partial into Calculate hashcode/WithKeys/AddKeys/Map
    Jan 27, 2023 1:15:27 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2023-01-27T01:15:26.369Z: Fusing consumer Calculate hashcode/Combine.perKey(Hashing)/GroupByKey/Reify into Calculate hashcode/Combine.perKey(Hashing)/GroupByKey+Calculate hashcode/Combine.perKey(Hashing)/Combine.GroupedValues/Partial
    Jan 27, 2023 1:15:27 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2023-01-27T01:15:26.402Z: Fusing consumer Calculate hashcode/Combine.perKey(Hashing)/GroupByKey/Write into Calculate hashcode/Combine.perKey(Hashing)/GroupByKey/Reify
    Jan 27, 2023 1:15:27 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2023-01-27T01:15:26.434Z: Fusing consumer Calculate hashcode/Combine.perKey(Hashing)/Combine.GroupedValues into Calculate hashcode/Combine.perKey(Hashing)/GroupByKey/Read
    Jan 27, 2023 1:15:27 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2023-01-27T01:15:26.464Z: Fusing consumer Calculate hashcode/Combine.perKey(Hashing)/Combine.GroupedValues/Extract into Calculate hashcode/Combine.perKey(Hashing)/Combine.GroupedValues
    Jan 27, 2023 1:15:27 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2023-01-27T01:15:26.499Z: Fusing consumer Calculate hashcode/Values/Values/Map into Calculate hashcode/Combine.perKey(Hashing)/Combine.GroupedValues/Extract
    Jan 27, 2023 1:15:27 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2023-01-27T01:15:26.521Z: Fusing consumer PAssert$0/GroupGlobally/Reify.Window/ParDo(Anonymous) into Calculate hashcode/Values/Values/Map
    Jan 27, 2023 1:15:27 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2023-01-27T01:15:26.553Z: Fusing consumer PAssert$0/GroupGlobally/ParDo(ToSingletonIterables) into PAssert$0/GroupGlobally/Reify.Window/ParDo(Anonymous)
    Jan 27, 2023 1:15:27 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2023-01-27T01:15:26.584Z: Fusing consumer PAssert$0/GroupGlobally/WithKeys/AddKeys/Map into PAssert$0/GroupGlobally/Create.Values/Read(CreateSource)
    Jan 27, 2023 1:15:27 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2023-01-27T01:15:26.606Z: Fusing consumer PAssert$0/GroupGlobally/GroupByKey/Reify into PAssert$0/GroupGlobally/WithKeys/AddKeys/Map
    Jan 27, 2023 1:15:27 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2023-01-27T01:15:26.925Z: Executing operation Read from bounded Kafka/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Reshuffle/Reshuffle/GroupByKey/Create
    Jan 27, 2023 1:15:27 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2023-01-27T01:15:26.990Z: Starting 5 ****s in us-central1-b...
    Jan 27, 2023 1:15:27 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2023-01-27T01:15:27.226Z: Finished operation Read from bounded Kafka/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Reshuffle/Reshuffle/GroupByKey/Create
    Jan 27, 2023 1:15:27 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2023-01-27T01:15:27.376Z: Executing operation Read from bounded Kafka/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Create/Read(CreateSource)+Read from bounded Kafka/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Split+Read from bounded Kafka/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Reshuffle/Pair with random key+Read from bounded Kafka/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Reshuffle/Reshuffle/Window.Into()/Window.Assign+Read from bounded Kafka/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Reshuffle/Reshuffle/GroupByKey/Reify+Read from bounded Kafka/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Reshuffle/Reshuffle/GroupByKey/Write
    Jan 27, 2023 1:15:29 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2023-01-27T01:15:28.911Z: Your project already contains 100 Dataflow-created metric descriptors, so new user metrics of the form custom.googleapis.com/* will not be created. However, all user metrics are also available in the metric dataflow.googleapis.com/job/user_counter. If you rely on the custom metrics, you can delete old / unused metric descriptors. See https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.list and https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.delete
    Jan 27, 2023 1:16:01 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2023-01-27T01:16:01.482Z: Autoscaling: Raised the number of ****s to 1 based on the rate of progress in the currently running stage(s).
    Jan 27, 2023 1:16:01 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2023-01-27T01:16:01.513Z: Autoscaling: Resized **** pool to 1, though goal was 5.  This could be a quota issue.
    Jan 27, 2023 1:16:12 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2023-01-27T01:16:10.942Z: Autoscaling: Raised the number of ****s to 5 based on the rate of progress in the currently running stage(s).
    Jan 27, 2023 1:16:43 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2023-01-27T01:16:41.414Z: Workers have started successfully.
    Jan 27, 2023 1:17:04 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2023-01-27T01:17:03.716Z: All ****s have finished the startup processes and began to receive work requests.
    Jan 27, 2023 1:17:11 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2023-01-27T01:17:08.975Z: Finished operation Read from bounded Kafka/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Create/Read(CreateSource)+Read from bounded Kafka/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Split+Read from bounded Kafka/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Reshuffle/Pair with random key+Read from bounded Kafka/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Reshuffle/Reshuffle/Window.Into()/Window.Assign+Read from bounded Kafka/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Reshuffle/Reshuffle/GroupByKey/Reify+Read from bounded Kafka/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Reshuffle/Reshuffle/GroupByKey/Write
    Jan 27, 2023 1:17:11 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2023-01-27T01:17:09.039Z: Executing operation Read from bounded Kafka/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Reshuffle/Reshuffle/GroupByKey/Close
    Jan 27, 2023 1:17:11 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2023-01-27T01:17:09.097Z: Finished operation Read from bounded Kafka/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Reshuffle/Reshuffle/GroupByKey/Close
    Jan 27, 2023 1:17:11 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2023-01-27T01:17:09.175Z: Executing operation Calculate hashcode/Combine.perKey(Hashing)/GroupByKey/Create
    Jan 27, 2023 1:17:11 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2023-01-27T01:17:09.308Z: Finished operation Calculate hashcode/Combine.perKey(Hashing)/GroupByKey/Create
    Jan 27, 2023 1:17:11 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2023-01-27T01:17:09.442Z: Executing operation Read from bounded Kafka/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Reshuffle/Reshuffle/GroupByKey/Read+Read from bounded Kafka/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Reshuffle/Reshuffle/GroupByKey/GroupByWindow+Read from bounded Kafka/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Reshuffle/Reshuffle/ExpandIterable+Read from bounded Kafka/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Reshuffle/Values/Values/Map+Read from bounded Kafka/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Read+Read from bounded Kafka/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/StripIds+Measure read time+Map records to strings/Map+Calculate hashcode/WithKeys/AddKeys/Map+Calculate hashcode/Combine.perKey(Hashing)/GroupByKey+Calculate hashcode/Combine.perKey(Hashing)/Combine.GroupedValues/Partial+Calculate hashcode/Combine.perKey(Hashing)/GroupByKey/Reify+Calculate hashcode/Combine.perKey(Hashing)/GroupByKey/Write
    Jan 27, 2023 1:45:08 AM org.apache.beam.runners.dataflow.DataflowPipelineJob waitUntilFinish
    WARNING: No terminal state was returned within allotted timeout. State value RUNNING
    Jan 27, 2023 1:45:08 AM org.apache.kafka.common.config.AbstractConfig logAll
    INFO: AdminClientConfig values: 
    	bootstrap.servers = [34.133.19.41:32401, 34.123.175.172:32402, 34.171.24.82:32403]
    	client.dns.lookup = default
    	client.id = 
    	connections.max.idle.ms = 300000
    	metadata.max.age.ms = 300000
    	metric.reporters = []
    	metrics.num.samples = 2
    	metrics.recording.level = INFO
    	metrics.sample.window.ms = 30000
    	receive.buffer.bytes = 65536
    	reconnect.backoff.max.ms = 1000
    	reconnect.backoff.ms = 50
    	request.timeout.ms = 120000
    	retries = 5
    	retry.backoff.ms = 100
    	sasl.client.callback.handler.class = null
    	sasl.jaas.config = null
    	sasl.kerberos.kinit.cmd = /usr/bin/kinit
    	sasl.kerberos.min.time.before.relogin = 60000
    	sasl.kerberos.service.name = null
    	sasl.kerberos.ticket.renew.jitter = 0.05
    	sasl.kerberos.ticket.renew.window.factor = 0.8
    	sasl.login.callback.handler.class = null
    	sasl.login.class = null
    	sasl.login.refresh.buffer.seconds = 300
    	sasl.login.refresh.min.period.seconds = 60
    	sasl.login.refresh.window.factor = 0.8
    	sasl.login.refresh.window.jitter = 0.05
    	sasl.mechanism = GSSAPI
    	security.protocol = PLAINTEXT
    	security.providers = null
    	send.buffer.bytes = 131072
    	ssl.cipher.suites = null
    	ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1]
    	ssl.endpoint.identification.algorithm = https
    	ssl.key.password = null
    	ssl.keymanager.algorithm = SunX509
    	ssl.keystore.location = null
    	ssl.keystore.password = null
    	ssl.keystore.type = JKS
    	ssl.protocol = TLS
    	ssl.provider = null
    	ssl.secure.random.implementation = null
    	ssl.trustmanager.algorithm = PKIX
    	ssl.truststore.location = null
    	ssl.truststore.password = null
    	ssl.truststore.type = JKS

    Jan 27, 2023 1:45:08 AM org.apache.kafka.common.utils.AppInfoParser$AppInfo <init>
    INFO: Kafka version: 2.4.1
    Jan 27, 2023 1:45:08 AM org.apache.kafka.common.utils.AppInfoParser$AppInfo <init>
    INFO: Kafka commitId: c57222ae8cd7866b
    Jan 27, 2023 1:45:08 AM org.apache.kafka.common.utils.AppInfoParser$AppInfo <init>
    INFO: Kafka startTimeMs: 1674783908173

Gradle Test Executor 1 finished executing tests.

> Task :sdks:java:io:kafka:integrationTest FAILED

org.apache.beam.sdk.io.kafka.KafkaIOIT > testKafkaIOReadsAndWritesCorrectlyInBatch FAILED
    java.lang.AssertionError: expected:<DONE> but was:<null>
        at org.junit.Assert.fail(Assert.java:89)
        at org.junit.Assert.failNotEquals(Assert.java:835)
        at org.junit.Assert.assertEquals(Assert.java:120)
        at org.junit.Assert.assertEquals(Assert.java:146)
        at org.apache.beam.sdk.io.kafka.KafkaIOIT.testKafkaIOReadsAndWritesCorrectlyInBatch(KafkaIOIT.java:275)

1 test completed, 1 failed
Finished generating test XML results (0.026 secs) into: <https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/src/sdks/java/io/kafka/build/test-results/integrationTest>
Generating HTML test report...
Finished generating test html results (0.038 secs) into: <https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/src/sdks/java/io/kafka/build/reports/tests/integrationTest>
:sdks:java:io:kafka:integrationTest (Thread[Execution **** Thread 7,5,main]) completed. Took 37 mins 21.492 secs.

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:io:kafka:integrationTest'.
> There were failing tests. See the report at: file://<https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/src/sdks/java/io/kafka/build/reports/tests/integrationTest/index.html>

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --debug option to get more log output.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.5.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 37m 45s
137 actionable tasks: 1 executed, 136 up-to-date

Publishing build scan...
https://gradle.com/s/lb7azua255h3w

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PerformanceTests_Kafka_IO #3579

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/3579/display/redirect>

Changes:


------------------------------------------
[...truncated 626.88 KB...]
    	at org.apache.beam.sdk.transforms.reflect.DoFnInvokers.tryInvokeSetupFor(DoFnInvokers.java:53)
    	at org.apache.beam.fn.harness.FnApiDoFnRunner.<init>(FnApiDoFnRunner.java:496)
    	at org.apache.beam.fn.harness.FnApiDoFnRunner$Factory.createRunnerForPTransform(FnApiDoFnRunner.java:192)
    	at org.apache.beam.fn.harness.FnApiDoFnRunner$Factory.createRunnerForPTransform(FnApiDoFnRunner.java:161)
    	at org.apache.beam.fn.harness.control.ProcessBundleHandler.createRunnerAndConsumersForPTransformRecursively(ProcessBundleHandler.java:302)
    	at org.apache.beam.fn.harness.control.ProcessBundleHandler.createRunnerAndConsumersForPTransformRecursively(ProcessBundleHandler.java:256)
    	at org.apache.beam.fn.harness.control.ProcessBundleHandler.createRunnerAndConsumersForPTransformRecursively(ProcessBundleHandler.java:256)
    	at org.apache.beam.fn.harness.control.ProcessBundleHandler.createRunnerAndConsumersForPTransformRecursively(ProcessBundleHandler.java:256)
    	at org.apache.beam.fn.harness.control.ProcessBundleHandler.createRunnerAndConsumersForPTransformRecursively(ProcessBundleHandler.java:256)
    	at org.apache.beam.fn.harness.control.ProcessBundleHandler.createBundleProcessor(ProcessBundleHandler.java:860)
    	at org.apache.beam.fn.harness.control.ProcessBundleHandler.lambda$processBundle$0(ProcessBundleHandler.java:506)
    	at org.apache.beam.fn.harness.control.ProcessBundleHandler$BundleProcessorCache.get(ProcessBundleHandler.java:971)
    	at org.apache.beam.fn.harness.control.ProcessBundleHandler.processBundle(ProcessBundleHandler.java:502)
    	at org.apache.beam.fn.harness.control.BeamFnControlClient.delegateOnInstructionRequestType(BeamFnControlClient.java:151)
    	at org.apache.beam.fn.harness.control.BeamFnControlClient$InboundObserver.lambda$onNext$0(BeamFnControlClient.java:116)
    	at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
    	at java.util.concurrent.FutureTask.run(FutureTask.java:266)
    	at org.apache.beam.sdk.util.UnboundedScheduledExecutorService$ScheduledFutureTask.run(UnboundedScheduledExecutorService.java:162)
    	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
    	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
    	at java.lang.Thread.run(Thread.java:750)
    ,

      kafkaioit0testkafkaioread-01261204-e6tm-harness-235c
          Root cause: org.apache.beam.sdk.util.UserCodeException: org.apache.kafka.common.KafkaException: Failed to construct kafka producer
    	at org.apache.beam.sdk.util.UserCodeException.wrap(UserCodeException.java:39)
    	at org.apache.beam.sdk.io.kafka.KafkaWriter$DoFnInvoker.invokeSetup(Unknown Source)
    	at org.apache.beam.sdk.transforms.reflect.DoFnInvokers.tryInvokeSetupFor(DoFnInvokers.java:53)
    	at org.apache.beam.fn.harness.FnApiDoFnRunner.<init>(FnApiDoFnRunner.java:496)
    	at org.apache.beam.fn.harness.FnApiDoFnRunner$Factory.createRunnerForPTransform(FnApiDoFnRunner.java:192)
    	at org.apache.beam.fn.harness.FnApiDoFnRunner$Factory.createRunnerForPTransform(FnApiDoFnRunner.java:161)
    	at org.apache.beam.fn.harness.control.ProcessBundleHandler.createRunnerAndConsumersForPTransformRecursively(ProcessBundleHandler.java:302)
    	at org.apache.beam.fn.harness.control.ProcessBundleHandler.createRunnerAndConsumersForPTransformRecursively(ProcessBundleHandler.java:256)
    	at org.apache.beam.fn.harness.control.ProcessBundleHandler.createRunnerAndConsumersForPTransformRecursively(ProcessBundleHandler.java:256)
    	at org.apache.beam.fn.harness.control.ProcessBundleHandler.createRunnerAndConsumersForPTransformRecursively(ProcessBundleHandler.java:256)
    	at org.apache.beam.fn.harness.control.ProcessBundleHandler.createRunnerAndConsumersForPTransformRecursively(ProcessBundleHandler.java:256)
    	at org.apache.beam.fn.harness.control.ProcessBundleHandler.createBundleProcessor(ProcessBundleHandler.java:860)
    	at org.apache.beam.fn.harness.control.ProcessBundleHandler.lambda$processBundle$0(ProcessBundleHandler.java:506)
    	at org.apache.beam.fn.harness.control.ProcessBundleHandler$BundleProcessorCache.get(ProcessBundleHandler.java:971)
    	at org.apache.beam.fn.harness.control.ProcessBundleHandler.processBundle(ProcessBundleHandler.java:502)
    	at org.apache.beam.fn.harness.control.BeamFnControlClient.delegateOnInstructionRequestType(BeamFnControlClient.java:151)
    	at org.apache.beam.fn.harness.control.BeamFnControlClient$InboundObserver.lambda$onNext$0(BeamFnControlClient.java:116)
    	at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
    	at java.util.concurrent.FutureTask.run(FutureTask.java:266)
    	at org.apache.beam.sdk.util.UnboundedScheduledExecutorService$ScheduledFutureTask.run(UnboundedScheduledExecutorService.java:162)
    	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
    	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
    	at java.lang.Thread.run(Thread.java:750)
    	Suppressed: org.apache.beam.sdk.util.UserCodeException: java.lang.IllegalStateException
    		at org.apache.beam.sdk.util.UserCodeException.wrap(UserCodeException.java:39)
    		at org.apache.beam.sdk.io.kafka.KafkaWriter$DoFnInvoker.invokeTeardown(Unknown Source)
    		at org.apache.beam.sdk.transforms.reflect.DoFnInvokers.tryInvokeSetupFor(DoFnInvokers.java:56)
    		... 20 more
    	Caused by: java.lang.IllegalStateException
    		at org.apache.beam.sdk.util.Preconditions.checkStateNotNull(Preconditions.java:452)
    		at org.apache.beam.sdk.io.kafka.KafkaWriter.teardown(KafkaWriter.java:102)
    Caused by: org.apache.kafka.common.KafkaException: Failed to construct kafka producer
    	at org.apache.kafka.clients.producer.KafkaProducer.<init>(KafkaProducer.java:432)
    	at org.apache.kafka.clients.producer.KafkaProducer.<init>(KafkaProducer.java:270)
    	at org.apache.beam.sdk.io.kafka.KafkaWriter.setup(KafkaWriter.java:52)
    Caused by: org.apache.kafka.common.config.ConfigException: No resolvable bootstrap urls given in bootstrap.servers
    	at org.apache.kafka.clients.ClientUtils.parseAndValidateAddresses(ClientUtils.java:88)
    	at org.apache.kafka.clients.ClientUtils.parseAndValidateAddresses(ClientUtils.java:47)
    	at org.apache.kafka.clients.producer.KafkaProducer.<init>(KafkaProducer.java:407)
    	at org.apache.kafka.clients.producer.KafkaProducer.<init>(KafkaProducer.java:270)
    	at org.apache.beam.sdk.io.kafka.KafkaWriter.setup(KafkaWriter.java:52)
    	at org.apache.beam.sdk.io.kafka.KafkaWriter$DoFnInvoker.invokeSetup(Unknown Source)
    	at org.apache.beam.sdk.transforms.reflect.DoFnInvokers.tryInvokeSetupFor(DoFnInvokers.java:53)
    	at org.apache.beam.fn.harness.FnApiDoFnRunner.<init>(FnApiDoFnRunner.java:496)
    	at org.apache.beam.fn.harness.FnApiDoFnRunner$Factory.createRunnerForPTransform(FnApiDoFnRunner.java:192)
    	at org.apache.beam.fn.harness.FnApiDoFnRunner$Factory.createRunnerForPTransform(FnApiDoFnRunner.java:161)
    	at org.apache.beam.fn.harness.control.ProcessBundleHandler.createRunnerAndConsumersForPTransformRecursively(ProcessBundleHandler.java:302)
    	at org.apache.beam.fn.harness.control.ProcessBundleHandler.createRunnerAndConsumersForPTransformRecursively(ProcessBundleHandler.java:256)
    	at org.apache.beam.fn.harness.control.ProcessBundleHandler.createRunnerAndConsumersForPTransformRecursively(ProcessBundleHandler.java:256)
    	at org.apache.beam.fn.harness.control.ProcessBundleHandler.createRunnerAndConsumersForPTransformRecursively(ProcessBundleHandler.java:256)
    	at org.apache.beam.fn.harness.control.ProcessBundleHandler.createRunnerAndConsumersForPTransformRecursively(ProcessBundleHandler.java:256)
    	at org.apache.beam.fn.harness.control.ProcessBundleHandler.createBundleProcessor(ProcessBundleHandler.java:860)
    	at org.apache.beam.fn.harness.control.ProcessBundleHandler.lambda$processBundle$0(ProcessBundleHandler.java:506)
    	at org.apache.beam.fn.harness.control.ProcessBundleHandler$BundleProcessorCache.get(ProcessBundleHandler.java:971)
    	at org.apache.beam.fn.harness.control.ProcessBundleHandler.processBundle(ProcessBundleHandler.java:502)
    	at org.apache.beam.fn.harness.control.BeamFnControlClient.delegateOnInstructionRequestType(BeamFnControlClient.java:151)
    	at org.apache.beam.fn.harness.control.BeamFnControlClient$InboundObserver.lambda$onNext$0(BeamFnControlClient.java:116)
    	at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
    	at java.util.concurrent.FutureTask.run(FutureTask.java:266)
    	at org.apache.beam.sdk.util.UnboundedScheduledExecutorService$ScheduledFutureTask.run(UnboundedScheduledExecutorService.java:162)
    	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
    	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
    	at java.lang.Thread.run(Thread.java:750)
    ,

      kafkaioit0testkafkaioread-01261204-e6tm-harness-235c
          Root cause: org.apache.beam.sdk.util.UserCodeException: org.apache.kafka.common.KafkaException: Failed to construct kafka producer
    	at org.apache.beam.sdk.util.UserCodeException.wrap(UserCodeException.java:39)
    	at org.apache.beam.sdk.io.kafka.KafkaWriter$DoFnInvoker.invokeSetup(Unknown Source)
    	at org.apache.beam.sdk.transforms.reflect.DoFnInvokers.tryInvokeSetupFor(DoFnInvokers.java:53)
    	at org.apache.beam.fn.harness.FnApiDoFnRunner.<init>(FnApiDoFnRunner.java:496)
    	at org.apache.beam.fn.harness.FnApiDoFnRunner$Factory.createRunnerForPTransform(FnApiDoFnRunner.java:192)
    	at org.apache.beam.fn.harness.FnApiDoFnRunner$Factory.createRunnerForPTransform(FnApiDoFnRunner.java:161)
    	at org.apache.beam.fn.harness.control.ProcessBundleHandler.createRunnerAndConsumersForPTransformRecursively(ProcessBundleHandler.java:302)
    	at org.apache.beam.fn.harness.control.ProcessBundleHandler.createRunnerAndConsumersForPTransformRecursively(ProcessBundleHandler.java:256)
    	at org.apache.beam.fn.harness.control.ProcessBundleHandler.createRunnerAndConsumersForPTransformRecursively(ProcessBundleHandler.java:256)
    	at org.apache.beam.fn.harness.control.ProcessBundleHandler.createRunnerAndConsumersForPTransformRecursively(ProcessBundleHandler.java:256)
    	at org.apache.beam.fn.harness.control.ProcessBundleHandler.createRunnerAndConsumersForPTransformRecursively(ProcessBundleHandler.java:256)
    	at org.apache.beam.fn.harness.control.ProcessBundleHandler.createBundleProcessor(ProcessBundleHandler.java:860)
    	at org.apache.beam.fn.harness.control.ProcessBundleHandler.lambda$processBundle$0(ProcessBundleHandler.java:506)
    	at org.apache.beam.fn.harness.control.ProcessBundleHandler$BundleProcessorCache.get(ProcessBundleHandler.java:971)
    	at org.apache.beam.fn.harness.control.ProcessBundleHandler.processBundle(ProcessBundleHandler.java:502)
    	at org.apache.beam.fn.harness.control.BeamFnControlClient.delegateOnInstructionRequestType(BeamFnControlClient.java:151)
    	at org.apache.beam.fn.harness.control.BeamFnControlClient$InboundObserver.lambda$onNext$0(BeamFnControlClient.java:116)
    	at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
    	at java.util.concurrent.FutureTask.run(FutureTask.java:266)
    	at org.apache.beam.sdk.util.UnboundedScheduledExecutorService$ScheduledFutureTask.run(UnboundedScheduledExecutorService.java:162)
    	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
    	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
    	at java.lang.Thread.run(Thread.java:750)
    	Suppressed: org.apache.beam.sdk.util.UserCodeException: java.lang.IllegalStateException
    		at org.apache.beam.sdk.util.UserCodeException.wrap(UserCodeException.java:39)
    		at org.apache.beam.sdk.io.kafka.KafkaWriter$DoFnInvoker.invokeTeardown(Unknown Source)
    		at org.apache.beam.sdk.transforms.reflect.DoFnInvokers.tryInvokeSetupFor(DoFnInvokers.java:56)
    		... 20 more
    	Caused by: java.lang.IllegalStateException
    		at org.apache.beam.sdk.util.Preconditions.checkStateNotNull(Preconditions.java:452)
    		at org.apache.beam.sdk.io.kafka.KafkaWriter.teardown(KafkaWriter.java:102)
    Caused by: org.apache.kafka.common.KafkaException: Failed to construct kafka producer
    	at org.apache.kafka.clients.producer.KafkaProducer.<init>(KafkaProducer.java:432)
    	at org.apache.kafka.clients.producer.KafkaProducer.<init>(KafkaProducer.java:270)
    	at org.apache.beam.sdk.io.kafka.KafkaWriter.setup(KafkaWriter.java:52)
    Caused by: org.apache.kafka.common.config.ConfigException: No resolvable bootstrap urls given in bootstrap.servers
    	at org.apache.kafka.clients.ClientUtils.parseAndValidateAddresses(ClientUtils.java:88)
    	at org.apache.kafka.clients.ClientUtils.parseAndValidateAddresses(ClientUtils.java:47)
    	at org.apache.kafka.clients.producer.KafkaProducer.<init>(KafkaProducer.java:407)
    	at org.apache.kafka.clients.producer.KafkaProducer.<init>(KafkaProducer.java:270)
    	at org.apache.beam.sdk.io.kafka.KafkaWriter.setup(KafkaWriter.java:52)
    	at org.apache.beam.sdk.io.kafka.KafkaWriter$DoFnInvoker.invokeSetup(Unknown Source)
    	at org.apache.beam.sdk.transforms.reflect.DoFnInvokers.tryInvokeSetupFor(DoFnInvokers.java:53)
    	at org.apache.beam.fn.harness.FnApiDoFnRunner.<init>(FnApiDoFnRunner.java:496)
    	at org.apache.beam.fn.harness.FnApiDoFnRunner$Factory.createRunnerForPTransform(FnApiDoFnRunner.java:192)
    	at org.apache.beam.fn.harness.FnApiDoFnRunner$Factory.createRunnerForPTransform(FnApiDoFnRunner.java:161)
    	at org.apache.beam.fn.harness.control.ProcessBundleHandler.createRunnerAndConsumersForPTransformRecursively(ProcessBundleHandler.java:302)
    	at org.apache.beam.fn.harness.control.ProcessBundleHandler.createRunnerAndConsumersForPTransformRecursively(ProcessBundleHandler.java:256)
    	at org.apache.beam.fn.harness.control.ProcessBundleHandler.createRunnerAndConsumersForPTransformRecursively(ProcessBundleHandler.java:256)
    	at org.apache.beam.fn.harness.control.ProcessBundleHandler.createRunnerAndConsumersForPTransformRecursively(ProcessBundleHandler.java:256)
    	at org.apache.beam.fn.harness.control.ProcessBundleHandler.createRunnerAndConsumersForPTransformRecursively(ProcessBundleHandler.java:256)
    	at org.apache.beam.fn.harness.control.ProcessBundleHandler.createBundleProcessor(ProcessBundleHandler.java:860)
    	at org.apache.beam.fn.harness.control.ProcessBundleHandler.lambda$processBundle$0(ProcessBundleHandler.java:506)
    	at org.apache.beam.fn.harness.control.ProcessBundleHandler$BundleProcessorCache.get(ProcessBundleHandler.java:971)
    	at org.apache.beam.fn.harness.control.ProcessBundleHandler.processBundle(ProcessBundleHandler.java:502)
    	at org.apache.beam.fn.harness.control.BeamFnControlClient.delegateOnInstructionRequestType(BeamFnControlClient.java:151)
    	at org.apache.beam.fn.harness.control.BeamFnControlClient$InboundObserver.lambda$onNext$0(BeamFnControlClient.java:116)
    	at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
    	at java.util.concurrent.FutureTask.run(FutureTask.java:266)
    	at org.apache.beam.sdk.util.UnboundedScheduledExecutorService$ScheduledFutureTask.run(UnboundedScheduledExecutorService.java:162)
    	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
    	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
    	at java.lang.Thread.run(Thread.java:750)

    Jan 26, 2023 8:06:31 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2023-01-26T20:06:30.094Z: Cleaning up.
    Jan 26, 2023 8:06:31 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2023-01-26T20:06:30.169Z: Stopping **** pool...
    Jan 26, 2023 8:08:45 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2023-01-26T20:08:44.857Z: Autoscaling: Resized **** pool from 5 to 0.
    Jan 26, 2023 8:08:45 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2023-01-26T20:08:44.921Z: Worker pool stopped.
    Jan 26, 2023 8:08:53 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
    INFO: Job 2023-01-26_12_04_06-117384652096874970 failed with status FAILED.

Gradle Test Executor 1 finished executing tests.

> Task :sdks:java:io:kafka:integrationTest FAILED

org.apache.beam.sdk.io.kafka.KafkaIOIT > testKafkaIOReadsAndWritesCorrectlyInStreaming FAILED
    java.lang.AssertionError: Values should be different. Actual: FAILED
        at org.junit.Assert.fail(Assert.java:89)
        at org.junit.Assert.failEquals(Assert.java:187)
        at org.junit.Assert.assertNotEquals(Assert.java:163)
        at org.junit.Assert.assertNotEquals(Assert.java:177)
        at org.apache.beam.sdk.io.kafka.KafkaIOIT.testKafkaIOReadsAndWritesCorrectlyInStreaming(KafkaIOIT.java:214)

1 test completed, 1 failed
Finished generating test XML results (0.033 secs) into: <https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/src/sdks/java/io/kafka/build/test-results/integrationTest>
Generating HTML test report...
Finished generating test html results (0.169 secs) into: <https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/src/sdks/java/io/kafka/build/reports/tests/integrationTest>
:sdks:java:io:kafka:integrationTest (Thread[Execution **** Thread 3,5,main]) completed. Took 5 mins 1.687 secs.
Resolve mutations for :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages (Thread[Execution **** Thread 7,5,main]) started.
Resolve mutations for :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages (Thread[Execution **** Thread 7,5,main]) completed. Took 0.0 secs.
:runners:google-cloud-dataflow-java:cleanUpDockerJavaImages (Thread[Execution **** Thread 2,5,main]) started.

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages
Custom actions are attached to task ':runners:google-cloud-dataflow-java:cleanUpDockerJavaImages'.
Caching disabled for task ':runners:google-cloud-dataflow-java:cleanUpDockerJavaImages' because:
  Gradle would require more information to cache this task
Task ':runners:google-cloud-dataflow-java:cleanUpDockerJavaImages' is not up-to-date because:
  Task has not declared any outputs despite executing actions.
Starting process 'command 'docker''. Working directory: <https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/src/runners/google-cloud-dataflow-java> Command: docker rmi --force us.gcr.io/apache-beam-testing/java-postcommit-it/java:20230126200258
Successfully started process 'command 'docker''
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20230126200258
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:b22df430f0cb6ac100190cdcae50c70ba3917f5a83c4ecb586be5058c87db22e
Starting process 'command 'gcloud''. Working directory: <https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/src/runners/google-cloud-dataflow-java> Command: gcloud --quiet container images untag us.gcr.io/apache-beam-testing/java-postcommit-it/java:20230126200258
Successfully started process 'command 'gcloud''
WARNING: Successfully resolved tag to sha256, but it is recommended to use sha256 directly.
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20230126200258]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:b22df430f0cb6ac100190cdcae50c70ba3917f5a83c4ecb586be5058c87db22e]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20230126200258] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:b22df430f0cb6ac100190cdcae50c70ba3917f5a83c4ecb586be5058c87db22e])].
Starting process 'command './scripts/cleanup_untagged_gcr_images.sh''. Working directory: <https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/src/runners/google-cloud-dataflow-java> Command: ./scripts/cleanup_untagged_gcr_images.sh us.gcr.io/apache-beam-testing/java-postcommit-it/java
Successfully started process 'command './scripts/cleanup_untagged_gcr_images.sh''
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:b22df430f0cb6ac100190cdcae50c70ba3917f5a83c4ecb586be5058c87db22e
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:b22df430f0cb6ac100190cdcae50c70ba3917f5a83c4ecb586be5058c87db22e
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:b22df430f0cb6ac100190cdcae50c70ba3917f5a83c4ecb586be5058c87db22e].
:runners:google-cloud-dataflow-java:cleanUpDockerJavaImages (Thread[Execution **** Thread 2,5,main]) completed. Took 4.692 secs.
Resolve mutations for :sdks:java:io:kafka:cleanUp (Thread[Execution **** Thread 6,5,main]) started.
Resolve mutations for :sdks:java:io:kafka:cleanUp (Thread[Execution **** Thread 6,5,main]) completed. Took 0.0 secs.
:sdks:java:io:kafka:cleanUp (Thread[Execution **** Thread 5,5,main]) started.

> Task :sdks:java:io:kafka:cleanUp
Skipping task ':sdks:java:io:kafka:cleanUp' as it has no actions.
:sdks:java:io:kafka:cleanUp (Thread[Execution **** Thread 5,5,main]) completed. Took 0.0 secs.

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:io:kafka:integrationTest'.
> There were failing tests. See the report at: file://<https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/src/sdks/java/io/kafka/build/reports/tests/integrationTest/index.html>

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --debug option to get more log output.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.5.1/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 6m 16s
158 actionable tasks: 99 executed, 55 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/73hqr4dcy3vzq

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_PerformanceTests_Kafka_IO #3578

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/3578/display/redirect>

Changes:


------------------------------------------
[...truncated 567.31 KB...]
 ---> Running in 2d67b2151a51
Removing intermediate container 2d67b2151a51
 ---> 298c02c09975
Step 5/22 : ADD target/slf4j-api.jar /opt/apache/beam/jars/
 ---> 2de85fa1055a
Step 6/22 : ADD target/slf4j-jdk14.jar /opt/apache/beam/jars/
 ---> 05f8e0b60e5f
Step 7/22 : ADD target/jcl-over-slf4j.jar /opt/apache/beam/jars/
 ---> d59ff4f63363
Step 8/22 : ADD target/log4j-over-slf4j.jar /opt/apache/beam/jars/
 ---> 14657512587c
Step 9/22 : ADD target/log4j-to-slf4j.jar /opt/apache/beam/jars/
 ---> 0e695e83822d
Step 10/22 : ADD target/beam-sdks-java-harness.jar /opt/apache/beam/jars/
 ---> 4b50ace4b806
Step 11/22 : COPY target/jamm.jar target/open-module-agent*.jar /opt/apache/beam/jars/
 ---> fa38f247a37a
Step 12/22 : ADD target/linux_amd64/boot /opt/apache/beam/
 ---> 603a800f115a
Step 13/22 : COPY target/LICENSE /opt/apache/beam/
 ---> 7afb7a557c69
Step 14/22 : COPY target/NOTICE /opt/apache/beam/
 ---> b7ab0f3c5401
Step 15/22 : ADD target/third_party_licenses /opt/apache/beam/third_party_licenses/
 ---> 4794edac26fe
Step 16/22 : COPY target/LICENSE target/options/* /opt/apache/beam/options/
 ---> e0a8802fac49
Step 17/22 : RUN rm /opt/apache/beam/options/LICENSE
 ---> Running in 278aedd038cf
Removing intermediate container 278aedd038cf
 ---> 6afa2f67e019
Step 18/22 : COPY target/LICENSE target/go-licenses/* /opt/apache/beam/third_party_licenses/golang/
 ---> 3c855fd49927
Step 19/22 : RUN rm /opt/apache/beam/third_party_licenses/golang/LICENSE
 ---> Running in 11abbbe5f951
Removing intermediate container 11abbbe5f951
 ---> be819c48a114
Step 20/22 : RUN if [ "${pull_licenses}" = "false" ] ; then     rm -rf /opt/apache/beam/third_party_licenses ;    fi
 ---> Running in b2a428fbd865
Removing intermediate container b2a428fbd865
 ---> fe22ec4b8ef9
Step 21/22 : COPY target/profiler/* /opt/google_cloud_profiler/
 ---> 7d8dd69e5a38
Step 22/22 : ENTRYPOINT ["/opt/apache/beam/boot"]
 ---> Running in 8c7ebf3b6bf4
Removing intermediate container 8c7ebf3b6bf4
 ---> 4e09fc5a4ca9
Successfully built 4e09fc5a4ca9
Successfully tagged apache/beam_java8_sdk:2.46.0.dev
:sdks:java:container:java8:docker (Thread[Execution **** Thread 4,5,main]) completed. Took 8.867 secs.
Resolve mutations for :runners:google-cloud-dataflow-java:buildAndPushDockerJavaContainer (Thread[included builds,5,main]) started.
Resolve mutations for :runners:google-cloud-dataflow-java:buildAndPushDockerJavaContainer (Thread[included builds,5,main]) completed. Took 0.0 secs.
:runners:google-cloud-dataflow-java:buildAndPushDockerJavaContainer (Thread[Execution **** Thread 7,5,main]) started.

> Task :runners:google-cloud-dataflow-java:buildAndPushDockerJavaContainer
Custom actions are attached to task ':runners:google-cloud-dataflow-java:buildAndPushDockerJavaContainer'.
Caching disabled for task ':runners:google-cloud-dataflow-java:buildAndPushDockerJavaContainer' because:
  Gradle would require more information to cache this task
Task ':runners:google-cloud-dataflow-java:buildAndPushDockerJavaContainer' is not up-to-date because:
  Task has not declared any outputs despite executing actions.
Starting process 'command 'docker''. Working directory: <https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/src/runners/google-cloud-dataflow-java> Command: docker tag apache/beam_java8_sdk:2.46.0.dev us.gcr.io/apache-beam-testing/java-postcommit-it/java:20230126194835
Successfully started process 'command 'docker''
Starting process 'command 'gcloud''. Working directory: <https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/src/runners/google-cloud-dataflow-java> Command: gcloud docker -- push us.gcr.io/apache-beam-testing/java-postcommit-it/java:20230126194835
Successfully started process 'command 'gcloud''
WARNING: `gcloud docker` will not be supported for Docker client versions above 18.03.

As an alternative, use `gcloud auth configure-docker` to configure `docker` to
use `gcloud` as a credential helper, then use `docker` as you would for non-GCR
registries, e.g. `docker pull gcr.io/project-id/my-image`. Add
`--verbosity=error` to silence this warning: `gcloud docker
--verbosity=error -- pull gcr.io/project-id/my-image`.

See: https://cloud.google.com/container-registry/docs/support/deprecation-notices#gcloud-docker

The push refers to repository [us.gcr.io/apache-beam-testing/java-postcommit-it/java]
f76846f5ea2b: Preparing
d9180cfb7eb9: Preparing
74f8566c0535: Preparing
78ed54059bdd: Preparing
b1332943cafc: Preparing
14be84d8eeee: Preparing
573cb48f1299: Preparing
e9578bcf2c2d: Preparing
55b2b9c9f773: Preparing
bb7e4af585e8: Preparing
44fae74b8434: Preparing
b6648eeb8cd4: Preparing
7bc2bdee4dc5: Preparing
9ba2eac74afd: Preparing
be6901190bd0: Preparing
2556a8e22289: Preparing
6b5aaff44254: Preparing
53a0b163e995: Preparing
b626401ef603: Preparing
9b55156abf26: Preparing
293d5db30c9f: Preparing
573cb48f1299: Waiting
03127cdb479b: Preparing
9c742cd6c7a5: Preparing
e9578bcf2c2d: Waiting
55b2b9c9f773: Waiting
7bc2bdee4dc5: Waiting
bb7e4af585e8: Waiting
b626401ef603: Waiting
44fae74b8434: Waiting
9b55156abf26: Waiting
9ba2eac74afd: Waiting
293d5db30c9f: Waiting
be6901190bd0: Waiting
03127cdb479b: Waiting
14be84d8eeee: Waiting
9c742cd6c7a5: Waiting
b6648eeb8cd4: Waiting
2556a8e22289: Waiting
53a0b163e995: Waiting
6b5aaff44254: Waiting
74f8566c0535: Pushed
b1332943cafc: Pushed
78ed54059bdd: Pushed
d9180cfb7eb9: Pushed
f76846f5ea2b: Pushed
573cb48f1299: Pushed
e9578bcf2c2d: Pushed
bb7e4af585e8: Pushed
55b2b9c9f773: Pushed
b6648eeb8cd4: Pushed
14be84d8eeee: Pushed
44fae74b8434: Pushed
6b5aaff44254: Layer already exists
53a0b163e995: Layer already exists
b626401ef603: Layer already exists
9b55156abf26: Layer already exists
293d5db30c9f: Layer already exists
03127cdb479b: Layer already exists
9c742cd6c7a5: Layer already exists
7bc2bdee4dc5: Pushed
9ba2eac74afd: Pushed
be6901190bd0: Pushed
2556a8e22289: Pushed
20230126194835: digest: sha256:13076b3d473715b439fbc3a29fd68c16a574bdcecc62c61d6e0a5a4e7cac17bc size: 5140
:runners:google-cloud-dataflow-java:buildAndPushDockerJavaContainer (Thread[Execution **** Thread 7,5,main]) completed. Took 7.031 secs.
Resolve mutations for :sdks:java:io:kafka:integrationTest (Thread[Execution **** Thread 3,5,main]) started.
Resolve mutations for :sdks:java:io:kafka:integrationTest (Thread[Execution **** Thread 3,5,main]) completed. Took 0.0 secs.
:sdks:java:io:kafka:integrationTest (Thread[included builds,5,main]) started.
producer locations for task group 0 (Thread[Execution **** Thread 6,5,main]) started.
producer locations for task group 0 (Thread[Execution **** Thread 6,5,main]) completed. Took 0.0 secs.
Gradle Test Executor 1 started executing tests.

> Task :sdks:java:io:kafka:integrationTest
Custom actions are attached to task ':sdks:java:io:kafka:integrationTest'.
Build cache key for task ':sdks:java:io:kafka:integrationTest' is 17774cdff3ef923498b0bf74bb48d50c
Task ':sdks:java:io:kafka:integrationTest' is not up-to-date because:
  Task.upToDateWhen is false.
Starting process 'Gradle Test Executor 1'. Working directory: <https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/src/sdks/java/io/kafka> Command: /usr/lib/jvm/java-8-openjdk-amd64/bin/java -DbeamTestPipelineOptions=["--tempRoot=gs://temp-storage-for-perf-tests","--project=apache-beam-testing","--runner=DataflowRunner","--usePublicIPs=false","--sourceOptions={\"numRecords\":\"100000000\",\"keySizeBytes\":\"10\",\"valueSizeBytes\":\"90\"}","--bigQueryDataset=beam_performance","--bigQueryTable=kafkaioit_results_runner_v2","--influxMeasurement=kafkaioit_results_runner_v2","--influxDatabase=beam_test_metrics","--influxHost=http://10.128.0.96:8086","--kafkaBootstrapServerAddresses=$NODE_IP_0:32401,$NODE_IP_1:32402,$NODE_IP_2:32403","--kafkaTopic=beam-sdf","--readTimeout=1500","--numWorkers=5","--autoscalingAlgorithm=NONE","--experiments=use_runner_v2,use_unified_****","--sdkContainerImage=us.gcr.io/apache-beam-testing/java-postcommit-it/java:20230126194835","--region=us-central1"] -Djava.security.manager=****.org.gradle.process.internal.****.child.BootstrapSecurityManager -Dorg.gradle.internal.****.tmpdir=<https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/src/sdks/java/io/kafka/build/tmp/integrationTest/work> -Dorg.gradle.native=false -Xmx2g -Dfile.encoding=UTF-8 -Duser.country=US -Duser.language=en -Duser.variant -ea -cp /home/jenkins/.gradle/caches/7.5.1/****Main/gradle-****.jar ****.org.gradle.process.internal.****.GradleWorkerMain 'Gradle Test Executor 1'
Successfully started process 'Gradle Test Executor 1'

org.apache.beam.sdk.io.kafka.KafkaIOIT STANDARD_ERROR
    SLF4J: Class path contains multiple SLF4J bindings.
    SLF4J: Found binding in [jar:<https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/src/runners/google-cloud-dataflow-java/****/build/libs/beam-runners-google-cloud-dataflow-java-legacy-****-2.46.0-SNAPSHOT.jar!/org/slf4j/impl/StaticLoggerBinder.class]>
    SLF4J: Found binding in [jar:file:/home/jenkins/.gradle/caches/modules-2/files-2.1/org.slf4j/slf4j-jdk14/1.7.30/d35953dd2fe54ebe39fdf18cfd82fe6eb35b25ed/slf4j-jdk14-1.7.30.jar!/org/slf4j/impl/StaticLoggerBinder.class]
    SLF4J: Found binding in [jar:file:/home/jenkins/.gradle/caches/modules-2/files-2.1/org.slf4j/slf4j-simple/1.7.30/e606eac955f55ecf1d8edcccba04eb8ac98088dd/slf4j-simple-1.7.30.jar!/org/slf4j/impl/StaticLoggerBinder.class]
    SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation.
    SLF4J: Actual binding is of type [org.slf4j.impl.JDK14LoggerFactory]

Gradle Test Executor 1 finished executing tests.

> Task :sdks:java:io:kafka:integrationTest

org.apache.beam.sdk.io.kafka.KafkaIOIT > classMethod FAILED
    java.lang.IllegalArgumentException: Class interface org.apache.beam.sdk.testing.TestPipelineOptions missing a property named 'usePublicIPs'. Did you mean 'usePublicIps'?
        at org.apache.beam.sdk.options.PipelineOptionsFactory.parseObjects(PipelineOptionsFactory.java:1888)
        at org.apache.beam.sdk.options.PipelineOptionsFactory.access$400(PipelineOptionsFactory.java:146)
        at org.apache.beam.sdk.options.PipelineOptionsFactory$Builder.as(PipelineOptionsFactory.java:329)
        at org.apache.beam.sdk.testing.TestPipeline.testingPipelineOptions(TestPipeline.java:500)
        at org.apache.beam.sdk.io.common.IOITHelper.readIOTestPipelineOptions(IOITHelper.java:48)
        at org.apache.beam.sdk.io.kafka.KafkaIOIT.setup(KafkaIOIT.java:174)

1 test completed, 1 failed
Finished generating test XML results (0.023 secs) into: <https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/src/sdks/java/io/kafka/build/test-results/integrationTest>
Generating HTML test report...
Finished generating test html results (0.035 secs) into: <https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/src/sdks/java/io/kafka/build/reports/tests/integrationTest>

> Task :sdks:java:io:kafka:integrationTest FAILED
:sdks:java:io:kafka:integrationTest (Thread[included builds,5,main]) completed. Took 4.215 secs.
Resolve mutations for :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages (Thread[Execution **** Thread 6,5,main]) started.
Resolve mutations for :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages (Thread[Execution **** Thread 6,5,main]) completed. Took 0.0 secs.
:runners:google-cloud-dataflow-java:cleanUpDockerJavaImages (Thread[Execution **** Thread 2,5,main]) started.

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages
Custom actions are attached to task ':runners:google-cloud-dataflow-java:cleanUpDockerJavaImages'.
Caching disabled for task ':runners:google-cloud-dataflow-java:cleanUpDockerJavaImages' because:
  Gradle would require more information to cache this task
Task ':runners:google-cloud-dataflow-java:cleanUpDockerJavaImages' is not up-to-date because:
  Task has not declared any outputs despite executing actions.
Starting process 'command 'docker''. Working directory: <https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/src/runners/google-cloud-dataflow-java> Command: docker rmi --force us.gcr.io/apache-beam-testing/java-postcommit-it/java:20230126194835
Successfully started process 'command 'docker''
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20230126194835
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:13076b3d473715b439fbc3a29fd68c16a574bdcecc62c61d6e0a5a4e7cac17bc
Starting process 'command 'gcloud''. Working directory: <https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/src/runners/google-cloud-dataflow-java> Command: gcloud --quiet container images untag us.gcr.io/apache-beam-testing/java-postcommit-it/java:20230126194835
Successfully started process 'command 'gcloud''
WARNING: Successfully resolved tag to sha256, but it is recommended to use sha256 directly.
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20230126194835]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:13076b3d473715b439fbc3a29fd68c16a574bdcecc62c61d6e0a5a4e7cac17bc]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20230126194835] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:13076b3d473715b439fbc3a29fd68c16a574bdcecc62c61d6e0a5a4e7cac17bc])].
Starting process 'command './scripts/cleanup_untagged_gcr_images.sh''. Working directory: <https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/src/runners/google-cloud-dataflow-java> Command: ./scripts/cleanup_untagged_gcr_images.sh us.gcr.io/apache-beam-testing/java-postcommit-it/java
Successfully started process 'command './scripts/cleanup_untagged_gcr_images.sh''
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:13076b3d473715b439fbc3a29fd68c16a574bdcecc62c61d6e0a5a4e7cac17bc
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:13076b3d473715b439fbc3a29fd68c16a574bdcecc62c61d6e0a5a4e7cac17bc
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:13076b3d473715b439fbc3a29fd68c16a574bdcecc62c61d6e0a5a4e7cac17bc].
:runners:google-cloud-dataflow-java:cleanUpDockerJavaImages (Thread[Execution **** Thread 2,5,main]) completed. Took 5.217 secs.
Resolve mutations for :sdks:java:io:kafka:cleanUp (Thread[included builds,5,main]) started.
Resolve mutations for :sdks:java:io:kafka:cleanUp (Thread[included builds,5,main]) completed. Took 0.0 secs.
:sdks:java:io:kafka:cleanUp (Thread[Execution **** Thread 2,5,main]) started.

> Task :sdks:java:io:kafka:cleanUp
Skipping task ':sdks:java:io:kafka:cleanUp' as it has no actions.
:sdks:java:io:kafka:cleanUp (Thread[Execution **** Thread 2,5,main]) completed. Took 0.0 secs.

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:io:kafka:integrationTest'.
> There were failing tests. See the report at: file://<https://ci-beam.apache.org/job/beam_PerformanceTests_Kafka_IO/ws/src/sdks/java/io/kafka/build/reports/tests/integrationTest/index.html>

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --debug option to get more log output.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.5.1/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 1m 21s
158 actionable tasks: 99 executed, 55 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/yt4nw7vsvnfaq

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org