You are viewing a plain text version of this content. The canonical link for it is here.
Posted to builds@beam.apache.org by Apache Jenkins Server <je...@builds.apache.org> on 2022/01/08 16:03:09 UTC

Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #23

See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/23/display/redirect?page=changes>

Changes:

[Alexey Romanenko] [adhoc] Fix BigTableIO description

[noreply] [BEAM-13015] Remove dead code now that all instances have migrated to

[noreply] [BEAM-13386] Add RLock support for cloudpickle (#16250)

[danthev] Fix overflow


------------------------------------------
[...truncated 49.78 KB...]
59503af7129a: Waiting
11936051f93b: Waiting
bff9fe6e429c: Waiting
f9e18e59a565: Waiting
37922d475df2: Pushed
30116ca7fbc2: Pushed
42d5043edf61: Pushed
35ba2f40021d: Pushed
96503d4b9d7b: Pushed
273e4936d96a: Pushed
e169869b315f: Pushed
7c70d2b00ed8: Pushed
020b1cf345d3: Pushed
bff9fe6e429c: Layer already exists
07329e9a90e3: Pushed
7c245b2fe4f1: Layer already exists
f9e18e59a565: Layer already exists
b81df8826e45: Pushed
26a504e63be4: Layer already exists
8bf42db0de72: Layer already exists
31892cc314cb: Layer already exists
11936051f93b: Layer already exists
a5443b7579be: Pushed
59503af7129a: Pushed
20220108123036: digest: sha256:39d2b42de9099ca1faa6fa24abc467ff09f34951c0ac1cbf2b3892016ea70ec8 size: 4520

> Task :sdks:java:testing:load-tests:run
Jan 08, 2022 12:32:24 PM org.apache.beam.runners.dataflow.options.DataflowPipelineOptions$StagingLocationFactory create
INFO: No stagingLocation provided, falling back to gcpTempLocation
Jan 08, 2022 12:32:24 PM org.apache.beam.runners.dataflow.DataflowRunner fromOptions
INFO: PipelineOptions.filesToStage was not specified. Defaulting to files from the classpath: will stage 203 files. Enable logging at DEBUG level to see which files will be staged.
Jan 08, 2022 12:32:25 PM org.apache.beam.sdk.Pipeline validate
WARNING: The following transforms do not have stable unique names: Window.Into()
Jan 08, 2022 12:32:25 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Executing pipeline on the Dataflow Service, which will have billing implications related to Google Compute Engine usage and other Google Cloud Services.
Jan 08, 2022 12:32:27 PM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Uploading 203 files from PipelineOptions.filesToStage to staging location to prepare for execution.
Jan 08, 2022 12:32:28 PM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Staging files complete: 203 files cached, 0 files newly uploaded in 0 seconds
Jan 08, 2022 12:32:28 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Staging portable pipeline proto to gs://temp-storage-for-perf-tests/loadtests/staging/
Jan 08, 2022 12:32:28 PM org.apache.beam.runners.dataflow.util.PackageUtil tryStagePackage
INFO: Uploading <113691 bytes, hash c2d27c83baada62718938730262fb4f379ab4f2fb00459ffb8715e1ed404d312> to gs://temp-storage-for-perf-tests/loadtests/staging/pipeline-wtJ8g7qtpicYk4cwJi-083mrTy-wBFn_uHFeHtQE0xI.pb
Jan 08, 2022 12:32:30 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read input/DataflowRunner.StreamingUnboundedRead.ReadWithIds as step s1
Jan 08, 2022 12:32:30 PM org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource split
INFO: Split into 20 bundles of sizes: [org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4b039c6d, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7f5b9db, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@507d64aa, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@37045b48, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@60b34931, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4aa21f9d, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@71c17a57, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@640ab13c, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1e0a864d, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@440e3ce6, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3e67f5f2, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@fd53053, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4527f70a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@707b1a44, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7132a9dc, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@57435801, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2da66a44, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@527fc8e, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@61bfc9bf, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2c7106d9]
Jan 08, 2022 12:32:30 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read input/StripIds as step s2
Jan 08, 2022 12:32:30 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect start time metrics (input) as step s3
Jan 08, 2022 12:32:30 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Window.Into()/Window.Assign as step s4
Jan 08, 2022 12:32:30 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read co-input/DataflowRunner.StreamingUnboundedRead.ReadWithIds as step s5
Jan 08, 2022 12:32:30 PM org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource split
INFO: Split into 20 bundles of sizes: [org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4a37191a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5854a18, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@d5556bf, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@791c12e3, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@b112b13, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@24eb65e3, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3ac3f6f, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1abebef3, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@18f55704, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@67cefd84, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5fbe155, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6add8e3f, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@58a2b917, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@48904d5a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@12bbfc54, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1491344a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@59b65dce, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1386313f, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5e922647, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@433c6abb]
Jan 08, 2022 12:32:30 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read co-input/StripIds as step s6
Jan 08, 2022 12:32:30 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect start time metrics (co-input) as step s7
Jan 08, 2022 12:32:30 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Window.Into()2/Window.Assign as step s8
Jan 08, 2022 12:32:30 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/MakeUnionTable0 as step s9
Jan 08, 2022 12:32:30 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/MakeUnionTable1 as step s10
Jan 08, 2022 12:32:30 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/Flatten as step s11
Jan 08, 2022 12:32:30 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/GBK as step s12
Jan 08, 2022 12:32:30 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/ConstructCoGbkResultFn as step s13
Jan 08, 2022 12:32:30 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Ungroup and reiterate as step s14
Jan 08, 2022 12:32:30 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect total bytes as step s15
Jan 08, 2022 12:32:30 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect end time metrics as step s16
Jan 08, 2022 12:32:30 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Dataflow SDK version: 2.37.0-SNAPSHOT
Jan 08, 2022 12:32:31 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To access the Dataflow monitoring console, please navigate to https://console.cloud.google.com/dataflow/jobs/us-central1/2022-01-08_04_32_30-11533624520464033075?project=apache-beam-testing
Jan 08, 2022 12:32:31 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Submitted job: 2022-01-08_04_32_30-11533624520464033075
Jan 08, 2022 12:32:31 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To cancel the job using the 'gcloud' tool, run:
> gcloud dataflow jobs --project=apache-beam-testing cancel --region=us-central1 2022-01-08_04_32_30-11533624520464033075
Jan 08, 2022 12:32:39 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
WARNING: 2022-01-08T12:32:37.762Z: The workflow name is not a valid Cloud Label. Labels applied to Cloud resources (such as GCE Instances) for monitoring will be labeled with this modified job name: load0tests0java170dataflow0v20streaming0cogbk01-jenkins-01-1fsv. For the best monitoring experience, please name your job with a valid Cloud Label. For details, see: https://cloud.google.com/compute/docs/labeling-resources#restrictions
Jan 08, 2022 12:32:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-08T12:32:41.395Z: Worker configuration: e2-standard-2 in us-central1-f.
Jan 08, 2022 12:32:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-08T12:32:42.105Z: Expanding SplittableParDo operations into optimizable parts.
Jan 08, 2022 12:32:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-08T12:32:42.128Z: Expanding CollectionToSingleton operations into optimizable parts.
Jan 08, 2022 12:32:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-08T12:32:42.197Z: Expanding CoGroupByKey operations into optimizable parts.
Jan 08, 2022 12:32:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-08T12:32:42.248Z: Expanding SplittableProcessKeyed operations into optimizable parts.
Jan 08, 2022 12:32:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-08T12:32:42.274Z: Expanding GroupByKey operations into streaming Read/Write steps
Jan 08, 2022 12:32:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-08T12:32:42.343Z: Lifting ValueCombiningMappingFns into MergeBucketsMappingFns
Jan 08, 2022 12:32:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-08T12:32:42.468Z: Fusing adjacent ParDo, Read, Write, and Flatten operations
Jan 08, 2022 12:32:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-08T12:32:42.493Z: Unzipping flatten CoGroupByKey-Flatten for input CoGroupByKey-MakeUnionTable0-ParMultiDo-ConstructUnionTable-.output
Jan 08, 2022 12:32:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-08T12:32:42.526Z: Fusing unzipped copy of CoGroupByKey/GBK/WriteStream, through flatten CoGroupByKey/Flatten, into producer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable)
Jan 08, 2022 12:32:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-08T12:32:42.567Z: Fusing consumer CoGroupByKey/GBK/WriteStream into CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable)
Jan 08, 2022 12:32:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-08T12:32:42.590Z: Fusing consumer Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read input/Impulse
Jan 08, 2022 12:32:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-08T12:32:42.614Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
Jan 08, 2022 12:32:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-08T12:32:42.637Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
Jan 08, 2022 12:32:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-08T12:32:42.696Z: Fusing consumer Read input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
Jan 08, 2022 12:32:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-08T12:32:42.723Z: Fusing consumer Collect start time metrics (input)/ParMultiDo(TimeMonitor) into Read input/ParDo(StripIds)/ParMultiDo(StripIds)
Jan 08, 2022 12:32:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-08T12:32:42.750Z: Fusing consumer Window.Into()/Window.Assign into Collect start time metrics (input)/ParMultiDo(TimeMonitor)
Jan 08, 2022 12:32:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-08T12:32:42.771Z: Fusing consumer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable) into Window.Into()/Window.Assign
Jan 08, 2022 12:32:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-08T12:32:42.799Z: Fusing consumer Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read co-input/Impulse
Jan 08, 2022 12:32:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-08T12:32:42.824Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
Jan 08, 2022 12:32:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-08T12:32:42.846Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
Jan 08, 2022 12:32:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-08T12:32:42.877Z: Fusing consumer Read co-input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
Jan 08, 2022 12:32:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-08T12:32:42.921Z: Fusing consumer Collect start time metrics (co-input)/ParMultiDo(TimeMonitor) into Read co-input/ParDo(StripIds)/ParMultiDo(StripIds)
Jan 08, 2022 12:32:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-08T12:32:42.957Z: Fusing consumer Window.Into()2/Window.Assign into Collect start time metrics (co-input)/ParMultiDo(TimeMonitor)
Jan 08, 2022 12:32:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-08T12:32:42.981Z: Fusing consumer CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable) into Window.Into()2/Window.Assign
Jan 08, 2022 12:32:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-08T12:32:43.016Z: Fusing consumer CoGroupByKey/GBK/MergeBuckets into CoGroupByKey/GBK/ReadStream
Jan 08, 2022 12:32:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-08T12:32:43.050Z: Fusing consumer CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult) into CoGroupByKey/GBK/MergeBuckets
Jan 08, 2022 12:32:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-08T12:32:43.078Z: Fusing consumer Ungroup and reiterate/ParMultiDo(UngroupAndReiterate) into CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult)
Jan 08, 2022 12:32:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-08T12:32:43.103Z: Fusing consumer Collect total bytes/ParMultiDo(ByteMonitor) into Ungroup and reiterate/ParMultiDo(UngroupAndReiterate)
Jan 08, 2022 12:32:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-08T12:32:43.138Z: Fusing consumer Collect end time metrics/ParMultiDo(TimeMonitor) into Collect total bytes/ParMultiDo(ByteMonitor)
Jan 08, 2022 12:32:45 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-08T12:32:43.433Z: Starting 5 ****s in us-central1-f...
Jan 08, 2022 12:33:03 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-08T12:33:02.666Z: Your project already contains 100 Dataflow-created metric descriptors, so new user metrics of the form custom.googleapis.com/* will not be created. However, all user metrics are also available in the metric dataflow.googleapis.com/job/user_counter. If you rely on the custom metrics, you can delete old / unused metric descriptors. See https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.list and https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.delete
Jan 08, 2022 12:33:29 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-08T12:33:29.032Z: Autoscaling: Raised the number of ****s to 5 so that the pipeline can catch up with its backlog and keep up with its input rate.
Jan 08, 2022 12:34:20 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-08T12:34:19.790Z: Workers have started successfully.
Jan 08, 2022 12:34:20 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-08T12:34:19.818Z: Workers have started successfully.
Jan 08, 2022 4:00:28 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-08T16:00:27.624Z: Cancel request is committed for workflow job: 2022-01-08_04_32_30-11533624520464033075.
Jan 08, 2022 4:00:28 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-08T16:00:27.700Z: Cleaning up.
Jan 08, 2022 4:00:28 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-08T16:00:27.793Z: Stopping **** pool...
Jan 08, 2022 4:00:28 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-08T16:00:27.846Z: Stopping **** pool...
Jan 08, 2022 4:02:55 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-08T16:02:55.471Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Jan 08, 2022 4:02:55 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-08T16:02:55.507Z: Worker pool stopped.
Jan 08, 2022 4:03:02 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-01-08_04_32_30-11533624520464033075 finished with status CANCELLED.
Load test results for test (ID): dccc4f03-c330-404d-a972-5a72f2886af8 and timestamp: 2022-01-08T12:32:24.822000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                 12331.272
dataflow_v2_java17_total_bytes_count             3.15654884E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220108123036
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:39d2b42de9099ca1faa6fa24abc467ff09f34951c0ac1cbf2b3892016ea70ec8
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220108123036]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:39d2b42de9099ca1faa6fa24abc467ff09f34951c0ac1cbf2b3892016ea70ec8]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220108123036] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:39d2b42de9099ca1faa6fa24abc467ff09f34951c0ac1cbf2b3892016ea70ec8])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:778efe755b70f21de23933fbc0772b54ab528f99893fab28dc90a2a79f0b0f49
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:778efe755b70f21de23933fbc0772b54ab528f99893fab28dc90a2a79f0b0f49
ERROR: (gcloud.container.images.delete) Not found: response: {'status': '404', 'content-length': '168', 'x-xss-protection': '0', 'transfer-encoding': 'chunked', 'server': 'Docker Registry', '-content-encoding': 'gzip', 'docker-distribution-api-version': 'registry/2.0', 'cache-control': 'private', 'date': 'Sat, 08 Jan 2022 16:03:05 GMT', 'x-frame-options': 'SAMEORIGIN', 'content-type': 'application/json'}
Failed to compute blob liveness for manifest: 'sha256:778efe755b70f21de23933fbc0772b54ab528f99893fab28dc90a2a79f0b0f49': None

> Task :runners:google-cloud-dataflow-java:cleanUpDockerImages FAILED

FAILURE: Build completed with 2 failures.

1: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.
==============================================================================

2: Task failed with an exception.
-----------
* Where:
Build file '<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/runners/google-cloud-dataflow-java/build.gradle'> line: 290

* What went wrong:
Execution failed for task ':runners:google-cloud-dataflow-java:cleanUpDockerImages'.
> Process 'command './scripts/cleanup_untagged_gcr_images.sh'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.
==============================================================================

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 32m 43s
109 actionable tasks: 72 executed, 33 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/va3fdazoysj56

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Jenkins build is back to normal : beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #142

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/142/display/redirect?page=changes>


---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #141

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/141/display/redirect?page=changes>

Changes:

[andyye333] Add extra details to PubSub matcher errors

[noreply] Merge pull request #17559 from [BEAM-14423] Add exception injection

[noreply] [BEAM-11104] Allow self-checkpointing SDFs to return without finishing

[noreply] Merge pull request #17544 from [BEAM-14415] Exception handling tests for

[noreply] Merge pull request #17565 from [BEAM-14413] add Kafka exception test

[noreply] Merge pull request #17555 from [BEAM-14417] Adding exception handling

[noreply] [BEAM-14433] Improve Go split error message. (#17575)

[noreply] [BEAM-14429] Force java load test on dataflow runner v2

[noreply] Merge pull request #17577 from [BEAM-14435] Adding exception handling

[noreply] [BEAM-14347] Add generic registration functions for iters and emitters

[noreply] [BEAM-14169] Add Credentials rotation cron job for clusters (#17383)

[noreply] [BEAM-14347] Add generic registration for accumulators (#17579)


------------------------------------------
[...truncated 50.68 KB...]
8d3c4b8a056c: Preparing
779e60d40a60: Preparing
6e94b4ebcad2: Preparing
3a7cd99b8479: Preparing
c49c9f3cb678: Preparing
5a27e519e570: Waiting
3fe3663cf18e: Preparing
3634662e7534: Preparing
779e60d40a60: Waiting
6e94b4ebcad2: Waiting
267e49f181ed: Preparing
3bc383470c05: Preparing
e93827457889: Preparing
08fa02ce37eb: Preparing
a037458de4e0: Preparing
bafdbe68e4ae: Preparing
a13c519c6361: Preparing
08fa02ce37eb: Waiting
a13c519c6361: Waiting
a037458de4e0: Waiting
8d3c4b8a056c: Waiting
e93827457889: Waiting
3a7cd99b8479: Waiting
267e49f181ed: Waiting
c49c9f3cb678: Waiting
3bc383470c05: Waiting
83bac70d0466: Waiting
aa1c4990f94b: Pushed
5d41c137d1a8: Pushed
7d9293365344: Pushed
38be9522fc38: Pushed
a51548066d49: Pushed
8d3c4b8a056c: Pushed
5a27e519e570: Pushed
779e60d40a60: Pushed
6e94b4ebcad2: Pushed
c49c9f3cb678: Pushed
3a7cd99b8479: Pushed
3bc383470c05: Layer already exists
83bac70d0466: Pushed
e93827457889: Layer already exists
08fa02ce37eb: Layer already exists
a037458de4e0: Layer already exists
bafdbe68e4ae: Layer already exists
a13c519c6361: Layer already exists
3634662e7534: Pushed
267e49f181ed: Pushed
3fe3663cf18e: Pushed
20220510123038: digest: sha256:681c0d5f2a230de72485a61d99b98d34569d0303dda01320734c2ebe64c4a3da size: 4729

> Task :sdks:java:testing:load-tests:run
May 10, 2022 12:32:49 PM org.apache.beam.runners.dataflow.options.DataflowPipelineOptions$StagingLocationFactory create
INFO: No stagingLocation provided, falling back to gcpTempLocation
May 10, 2022 12:32:50 PM org.apache.beam.runners.dataflow.DataflowRunner fromOptions
INFO: PipelineOptions.filesToStage was not specified. Defaulting to files from the classpath: will stage 222 files. Enable logging at DEBUG level to see which files will be staged.
May 10, 2022 12:32:51 PM org.apache.beam.sdk.Pipeline validate
WARNING: The following transforms do not have stable unique names: Window.Into()
May 10, 2022 12:32:51 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Executing pipeline on the Dataflow Service, which will have billing implications related to Google Compute Engine usage and other Google Cloud Services.
May 10, 2022 12:32:53 PM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Uploading 222 files from PipelineOptions.filesToStage to staging location to prepare for execution.
May 10, 2022 12:32:54 PM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Staging files complete: 222 files cached, 0 files newly uploaded in 0 seconds
May 10, 2022 12:32:54 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Staging portable pipeline proto to gs://temp-storage-for-perf-tests/loadtests/staging/
May 10, 2022 12:32:54 PM org.apache.beam.runners.dataflow.util.PackageUtil tryStagePackage
INFO: Uploading <120279 bytes, hash 90d5250ac2615b6f15499695641c4bbc2e25c31d4352ec686073b7925252605d> to gs://temp-storage-for-perf-tests/loadtests/staging/pipeline-kNUlCsJhW28VSZaVZBxLvC4lwx1DUuxoYHO3klJSYF0.pb
May 10, 2022 12:32:56 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read input/DataflowRunner.StreamingUnboundedRead.ReadWithIds as step s1
May 10, 2022 12:32:56 PM org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource split
INFO: Split into 1 bundles of sizes: [org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4232b34a]
May 10, 2022 12:32:56 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read input/StripIds as step s2
May 10, 2022 12:32:56 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect start time metrics (input) as step s3
May 10, 2022 12:32:56 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Window.Into()/Window.Assign as step s4
May 10, 2022 12:32:56 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read co-input/DataflowRunner.StreamingUnboundedRead.ReadWithIds as step s5
May 10, 2022 12:32:56 PM org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource split
INFO: Split into 1 bundles of sizes: [org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@ffd4cba]
May 10, 2022 12:32:56 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read co-input/StripIds as step s6
May 10, 2022 12:32:56 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect start time metrics (co-input) as step s7
May 10, 2022 12:32:56 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Window.Into()2/Window.Assign as step s8
May 10, 2022 12:32:56 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/MakeUnionTable0 as step s9
May 10, 2022 12:32:56 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/MakeUnionTable1 as step s10
May 10, 2022 12:32:56 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/Flatten as step s11
May 10, 2022 12:32:56 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/GBK as step s12
May 10, 2022 12:32:56 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/ConstructCoGbkResultFn as step s13
May 10, 2022 12:32:56 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Ungroup and reiterate as step s14
May 10, 2022 12:32:56 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect total bytes as step s15
May 10, 2022 12:32:56 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect end time metrics as step s16
May 10, 2022 12:32:56 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Dataflow SDK version: 2.40.0-SNAPSHOT
May 10, 2022 12:32:57 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To access the Dataflow monitoring console, please navigate to https://console.cloud.google.com/dataflow/jobs/us-central1/2022-05-10_05_32_56-7371717862294852586?project=apache-beam-testing
May 10, 2022 12:32:57 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Submitted job: 2022-05-10_05_32_56-7371717862294852586
May 10, 2022 12:32:57 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To cancel the job using the 'gcloud' tool, run:
> gcloud dataflow jobs --project=apache-beam-testing cancel --region=us-central1 2022-05-10_05_32_56-7371717862294852586
May 10, 2022 12:33:08 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
WARNING: 2022-05-10T12:33:02.262Z: The workflow name is not a valid Cloud Label. Labels applied to Cloud resources (such as GCE Instances) for monitoring will be labeled with this modified job name: load0tests0java170dataflow0v20streaming0cogbk01-jenkins-05-ahvo. For the best monitoring experience, please name your job with a valid Cloud Label. For details, see: https://cloud.google.com/compute/docs/labeling-resources#restrictions
May 10, 2022 12:33:13 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-10T12:33:12.593Z: Worker configuration: e2-standard-2 in us-central1-b.
May 10, 2022 12:33:15 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-10T12:33:13.927Z: Expanding SplittableParDo operations into optimizable parts.
May 10, 2022 12:33:15 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-10T12:33:13.960Z: Expanding CollectionToSingleton operations into optimizable parts.
May 10, 2022 12:33:15 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-10T12:33:14.024Z: Expanding CoGroupByKey operations into optimizable parts.
May 10, 2022 12:33:15 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-10T12:33:14.093Z: Expanding SplittableProcessKeyed operations into optimizable parts.
May 10, 2022 12:33:15 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-10T12:33:14.125Z: Expanding GroupByKey operations into streaming Read/Write steps
May 10, 2022 12:33:15 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-10T12:33:14.186Z: Lifting ValueCombiningMappingFns into MergeBucketsMappingFns
May 10, 2022 12:33:15 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-10T12:33:14.308Z: Fusing adjacent ParDo, Read, Write, and Flatten operations
May 10, 2022 12:33:15 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-10T12:33:14.336Z: Unzipping flatten CoGroupByKey-Flatten for input CoGroupByKey-MakeUnionTable0-ParMultiDo-ConstructUnionTable-.output
May 10, 2022 12:33:15 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-10T12:33:14.371Z: Fusing unzipped copy of CoGroupByKey/GBK/WriteStream, through flatten CoGroupByKey/Flatten, into producer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable)
May 10, 2022 12:33:15 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-10T12:33:14.405Z: Fusing consumer CoGroupByKey/GBK/WriteStream into CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable)
May 10, 2022 12:33:15 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-10T12:33:14.439Z: Fusing consumer Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read input/Impulse
May 10, 2022 12:33:15 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-10T12:33:14.476Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
May 10, 2022 12:33:15 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-10T12:33:14.506Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
May 10, 2022 12:33:15 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-10T12:33:14.543Z: Fusing consumer Read input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
May 10, 2022 12:33:15 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-10T12:33:14.575Z: Fusing consumer Collect start time metrics (input)/ParMultiDo(TimeMonitor) into Read input/ParDo(StripIds)/ParMultiDo(StripIds)
May 10, 2022 12:33:15 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-10T12:33:14.606Z: Fusing consumer Window.Into()/Window.Assign into Collect start time metrics (input)/ParMultiDo(TimeMonitor)
May 10, 2022 12:33:15 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-10T12:33:14.659Z: Fusing consumer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable) into Window.Into()/Window.Assign
May 10, 2022 12:33:15 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-10T12:33:14.694Z: Fusing consumer Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read co-input/Impulse
May 10, 2022 12:33:15 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-10T12:33:14.738Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
May 10, 2022 12:33:15 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-10T12:33:14.772Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
May 10, 2022 12:33:15 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-10T12:33:14.797Z: Fusing consumer Read co-input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
May 10, 2022 12:33:15 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-10T12:33:14.827Z: Fusing consumer Collect start time metrics (co-input)/ParMultiDo(TimeMonitor) into Read co-input/ParDo(StripIds)/ParMultiDo(StripIds)
May 10, 2022 12:33:15 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-10T12:33:14.864Z: Fusing consumer Window.Into()2/Window.Assign into Collect start time metrics (co-input)/ParMultiDo(TimeMonitor)
May 10, 2022 12:33:15 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-10T12:33:14.889Z: Fusing consumer CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable) into Window.Into()2/Window.Assign
May 10, 2022 12:33:15 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-10T12:33:14.918Z: Fusing consumer CoGroupByKey/GBK/MergeBuckets into CoGroupByKey/GBK/ReadStream
May 10, 2022 12:33:15 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-10T12:33:14.952Z: Fusing consumer CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult) into CoGroupByKey/GBK/MergeBuckets
May 10, 2022 12:33:15 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-10T12:33:14.987Z: Fusing consumer Ungroup and reiterate/ParMultiDo(UngroupAndReiterate) into CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult)
May 10, 2022 12:33:15 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-10T12:33:15.020Z: Fusing consumer Collect total bytes/ParMultiDo(ByteMonitor) into Ungroup and reiterate/ParMultiDo(UngroupAndReiterate)
May 10, 2022 12:33:15 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-10T12:33:15.047Z: Fusing consumer Collect end time metrics/ParMultiDo(TimeMonitor) into Collect total bytes/ParMultiDo(ByteMonitor)
May 10, 2022 12:33:18 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-10T12:33:15.240Z: Running job using Streaming Engine
May 10, 2022 12:33:18 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-10T12:33:15.523Z: Starting 5 ****s in us-central1-b...
May 10, 2022 12:33:22 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-10T12:33:21.076Z: Your project already contains 100 Dataflow-created metric descriptors, so new user metrics of the form custom.googleapis.com/* will not be created. However, all user metrics are also available in the metric dataflow.googleapis.com/job/user_counter. If you rely on the custom metrics, you can delete old / unused metric descriptors. See https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.list and https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.delete
May 10, 2022 12:33:38 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-10T12:33:38.137Z: Autoscaling: Raised the number of ****s to 5 so that the pipeline can catch up with its backlog and keep up with its input rate.
May 10, 2022 12:34:47 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-10T12:34:46.474Z: Workers have started successfully.
May 10, 2022 2:13:37 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-10T14:13:36.440Z: Cleaning up.
May 10, 2022 2:13:37 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-10T14:13:36.646Z: Stopping **** pool...
May 10, 2022 2:13:37 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-10T14:13:36.700Z: Stopping **** pool...
May 10, 2022 2:14:13 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-10T14:14:11.917Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
May 10, 2022 2:14:13 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-10T14:14:11.960Z: Worker pool stopped.
May 10, 2022 2:14:18 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-05-10_05_32_56-7371717862294852586 finished with status DONE.
Load test results for test (ID): 8cdbdbfb-6fcb-49fa-a861-a1bb58fd0ad5 and timestamp: 2022-05-10T12:32:50.673000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                  5813.232
dataflow_v2_java17_total_bytes_count               6.0274294E9

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220510123038
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:681c0d5f2a230de72485a61d99b98d34569d0303dda01320734c2ebe64c4a3da
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220510123038]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:681c0d5f2a230de72485a61d99b98d34569d0303dda01320734c2ebe64c4a3da]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220510123038] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:681c0d5f2a230de72485a61d99b98d34569d0303dda01320734c2ebe64c4a3da])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:e016287d22625030f983cbf6f23e489ff06d1e82e3b1276c012f318dcaa366bf
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:e016287d22625030f983cbf6f23e489ff06d1e82e3b1276c012f318dcaa366bf
ERROR: (gcloud.container.images.delete) Not found: response: {'docker-distribution-api-version': 'registry/2.0', 'content-type': 'application/json', 'date': 'Tue, 10 May 2022 14:14:25 GMT', 'server': 'Docker Registry', 'cache-control': 'private', 'x-xss-protection': '0', 'x-frame-options': 'SAMEORIGIN', 'transfer-encoding': 'chunked', 'status': '404', 'content-length': '168', '-content-encoding': 'gzip'}
Failed to compute blob liveness for manifest: 'sha256:e016287d22625030f983cbf6f23e489ff06d1e82e3b1276c012f318dcaa366bf': None

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages FAILED

FAILURE: Build failed with an exception.

* Where:
Build file '<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/runners/google-cloud-dataflow-java/build.gradle'> line: 297

* What went wrong:
Execution failed for task ':runners:google-cloud-dataflow-java:cleanUpDockerJavaImages'.
> Process 'command './scripts/cleanup_untagged_gcr_images.sh'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.4/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 1h 44m
110 actionable tasks: 73 executed, 33 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/iters5gdlxc5s

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #140

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/140/display/redirect?page=changes>

Changes:

[elias.segundo] Changing elegibility to AllNodeElegibility

[chamikaramj] Adds code reviewers for GCP I/O connectors and KafkaIO to Beam OWNERS


------------------------------------------
[...truncated 50.04 KB...]
e93827457889: Preparing
08fa02ce37eb: Preparing
a037458de4e0: Preparing
bafdbe68e4ae: Preparing
a13c519c6361: Preparing
4558b07d7c6c: Waiting
7572dcf244a7: Waiting
a037458de4e0: Waiting
bafdbe68e4ae: Waiting
a13c519c6361: Waiting
76cbfa6ab6de: Waiting
3bc383470c05: Waiting
212e3bed8e24: Waiting
a1cdc24dc423: Waiting
08fa02ce37eb: Waiting
e93827457889: Waiting
e71737e48960: Waiting
9eded4dea7e1: Waiting
d25a54d106dd: Waiting
65c0e27593f2: Waiting
b0a0cfc0c1c7: Pushed
99b1b0c017c9: Pushed
70d12e40179b: Pushed
74d7ad42a174: Pushed
a7b750c44030: Pushed
76cbfa6ab6de: Pushed
212e3bed8e24: Pushed
65c0e27593f2: Pushed
e71737e48960: Pushed
9eded4dea7e1: Pushed
a1cdc24dc423: Pushed
3bc383470c05: Layer already exists
e93827457889: Layer already exists
08fa02ce37eb: Layer already exists
4fdbf0a79d70: Pushed
a037458de4e0: Layer already exists
bafdbe68e4ae: Layer already exists
a13c519c6361: Layer already exists
4558b07d7c6c: Pushed
7572dcf244a7: Pushed
d25a54d106dd: Pushed
20220509123035: digest: sha256:ae661c08a40bbf8a47368c1570d2cbdb9dbb0865ef587bd35d60f43cbee2a87d size: 4729

> Task :sdks:java:testing:load-tests:run
May 09, 2022 12:32:39 PM org.apache.beam.runners.dataflow.options.DataflowPipelineOptions$StagingLocationFactory create
INFO: No stagingLocation provided, falling back to gcpTempLocation
May 09, 2022 12:32:39 PM org.apache.beam.runners.dataflow.DataflowRunner fromOptions
INFO: PipelineOptions.filesToStage was not specified. Defaulting to files from the classpath: will stage 222 files. Enable logging at DEBUG level to see which files will be staged.
May 09, 2022 12:32:40 PM org.apache.beam.sdk.Pipeline validate
WARNING: The following transforms do not have stable unique names: Window.Into()
May 09, 2022 12:32:40 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Executing pipeline on the Dataflow Service, which will have billing implications related to Google Compute Engine usage and other Google Cloud Services.
May 09, 2022 12:32:43 PM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Uploading 222 files from PipelineOptions.filesToStage to staging location to prepare for execution.
May 09, 2022 12:32:44 PM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Staging files complete: 222 files cached, 0 files newly uploaded in 0 seconds
May 09, 2022 12:32:44 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Staging portable pipeline proto to gs://temp-storage-for-perf-tests/loadtests/staging/
May 09, 2022 12:32:44 PM org.apache.beam.runners.dataflow.util.PackageUtil tryStagePackage
INFO: Uploading <120277 bytes, hash 703fa7d7284d03a6724f18a00f996b599f49c2d0600f6d8db28805bcee16d47a> to gs://temp-storage-for-perf-tests/loadtests/staging/pipeline-cD-n1yhNA6ZyTxigD5lrWZ9JwtBgD22NsogFvO4W1Ho.pb
May 09, 2022 12:32:45 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read input/DataflowRunner.StreamingUnboundedRead.ReadWithIds as step s1
May 09, 2022 12:32:45 PM org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource split
INFO: Split into 20 bundles of sizes: [org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@751ae8a4, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@235d659c, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4232b34a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2da16263, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@f5ce0bb, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@47e51549, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@101a461c, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@360e9c06, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5ebffb44, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@311ff287, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7377781e, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@31db34da, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@109f8c7e, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5edacf20, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@16a5eb6d, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7e307087, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1220ef43, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1a8b81e8, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@234cff57, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5e8507f1]
May 09, 2022 12:32:45 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read input/StripIds as step s2
May 09, 2022 12:32:45 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect start time metrics (input) as step s3
May 09, 2022 12:32:45 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Window.Into()/Window.Assign as step s4
May 09, 2022 12:32:46 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read co-input/DataflowRunner.StreamingUnboundedRead.ReadWithIds as step s5
May 09, 2022 12:32:46 PM org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource split
INFO: Split into 20 bundles of sizes: [org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3bf54172, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2c9a6717, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7b3cde6f, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6d091cad, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7c663eaf, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3bb5ceb, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1e692555, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3ba0ae41, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@76fe6cdc, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2ffb3aec, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@786ff1cb, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@46039a21, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@431e86b1, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@35c4e864, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@32a2a6be, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@682af059, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5f36c8e3, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4da39ca9, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6a9344f5, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5584d9c6]
May 09, 2022 12:32:46 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read co-input/StripIds as step s6
May 09, 2022 12:32:46 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect start time metrics (co-input) as step s7
May 09, 2022 12:32:46 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Window.Into()2/Window.Assign as step s8
May 09, 2022 12:32:46 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/MakeUnionTable0 as step s9
May 09, 2022 12:32:46 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/MakeUnionTable1 as step s10
May 09, 2022 12:32:46 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/Flatten as step s11
May 09, 2022 12:32:46 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/GBK as step s12
May 09, 2022 12:32:46 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/ConstructCoGbkResultFn as step s13
May 09, 2022 12:32:46 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Ungroup and reiterate as step s14
May 09, 2022 12:32:46 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect total bytes as step s15
May 09, 2022 12:32:46 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect end time metrics as step s16
May 09, 2022 12:32:46 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Dataflow SDK version: 2.40.0-SNAPSHOT
May 09, 2022 12:32:46 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To access the Dataflow monitoring console, please navigate to https://console.cloud.google.com/dataflow/jobs/us-central1/2022-05-09_05_32_46-14665142041617253802?project=apache-beam-testing
May 09, 2022 12:32:46 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Submitted job: 2022-05-09_05_32_46-14665142041617253802
May 09, 2022 12:32:46 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To cancel the job using the 'gcloud' tool, run:
> gcloud dataflow jobs --project=apache-beam-testing cancel --region=us-central1 2022-05-09_05_32_46-14665142041617253802
May 09, 2022 12:32:52 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
WARNING: 2022-05-09T12:32:51.088Z: The workflow name is not a valid Cloud Label. Labels applied to Cloud resources (such as GCE Instances) for monitoring will be labeled with this modified job name: load0tests0java170dataflow0v20streaming0cogbk01-jenkins-05-ithr. For the best monitoring experience, please name your job with a valid Cloud Label. For details, see: https://cloud.google.com/compute/docs/labeling-resources#restrictions
May 09, 2022 12:32:57 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-09T12:32:56.474Z: Worker configuration: e2-standard-2 in us-central1-b.
May 09, 2022 12:33:00 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-09T12:32:57.352Z: Expanding SplittableParDo operations into optimizable parts.
May 09, 2022 12:33:00 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-09T12:32:57.408Z: Expanding CollectionToSingleton operations into optimizable parts.
May 09, 2022 12:33:00 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-09T12:32:57.481Z: Expanding CoGroupByKey operations into optimizable parts.
May 09, 2022 12:33:00 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-09T12:32:57.573Z: Expanding SplittableProcessKeyed operations into optimizable parts.
May 09, 2022 12:33:00 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-09T12:32:57.613Z: Expanding GroupByKey operations into streaming Read/Write steps
May 09, 2022 12:33:00 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-09T12:32:57.689Z: Lifting ValueCombiningMappingFns into MergeBucketsMappingFns
May 09, 2022 12:33:00 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-09T12:32:57.803Z: Fusing adjacent ParDo, Read, Write, and Flatten operations
May 09, 2022 12:33:00 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-09T12:32:57.838Z: Unzipping flatten CoGroupByKey-Flatten for input CoGroupByKey-MakeUnionTable0-ParMultiDo-ConstructUnionTable-.output
May 09, 2022 12:33:00 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-09T12:32:57.881Z: Fusing unzipped copy of CoGroupByKey/GBK/WriteStream, through flatten CoGroupByKey/Flatten, into producer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable)
May 09, 2022 12:33:00 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-09T12:32:57.914Z: Fusing consumer CoGroupByKey/GBK/WriteStream into CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable)
May 09, 2022 12:33:00 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-09T12:32:57.954Z: Fusing consumer Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read input/Impulse
May 09, 2022 12:33:00 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-09T12:32:57.985Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
May 09, 2022 12:33:00 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-09T12:32:58.023Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
May 09, 2022 12:33:00 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-09T12:32:58.058Z: Fusing consumer Read input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
May 09, 2022 12:33:00 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-09T12:32:58.093Z: Fusing consumer Collect start time metrics (input)/ParMultiDo(TimeMonitor) into Read input/ParDo(StripIds)/ParMultiDo(StripIds)
May 09, 2022 12:33:00 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-09T12:32:58.124Z: Fusing consumer Window.Into()/Window.Assign into Collect start time metrics (input)/ParMultiDo(TimeMonitor)
May 09, 2022 12:33:00 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-09T12:32:58.158Z: Fusing consumer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable) into Window.Into()/Window.Assign
May 09, 2022 12:33:00 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-09T12:32:58.194Z: Fusing consumer Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read co-input/Impulse
May 09, 2022 12:33:00 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-09T12:32:58.227Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
May 09, 2022 12:33:00 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-09T12:32:58.251Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
May 09, 2022 12:33:00 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-09T12:32:58.284Z: Fusing consumer Read co-input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
May 09, 2022 12:33:00 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-09T12:32:58.319Z: Fusing consumer Collect start time metrics (co-input)/ParMultiDo(TimeMonitor) into Read co-input/ParDo(StripIds)/ParMultiDo(StripIds)
May 09, 2022 12:33:00 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-09T12:32:58.353Z: Fusing consumer Window.Into()2/Window.Assign into Collect start time metrics (co-input)/ParMultiDo(TimeMonitor)
May 09, 2022 12:33:00 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-09T12:32:58.400Z: Fusing consumer CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable) into Window.Into()2/Window.Assign
May 09, 2022 12:33:00 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-09T12:32:58.429Z: Fusing consumer CoGroupByKey/GBK/MergeBuckets into CoGroupByKey/GBK/ReadStream
May 09, 2022 12:33:00 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-09T12:32:58.456Z: Fusing consumer CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult) into CoGroupByKey/GBK/MergeBuckets
May 09, 2022 12:33:00 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-09T12:32:58.492Z: Fusing consumer Ungroup and reiterate/ParMultiDo(UngroupAndReiterate) into CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult)
May 09, 2022 12:33:00 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-09T12:32:58.525Z: Fusing consumer Collect total bytes/ParMultiDo(ByteMonitor) into Ungroup and reiterate/ParMultiDo(UngroupAndReiterate)
May 09, 2022 12:33:00 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-09T12:32:58.575Z: Fusing consumer Collect end time metrics/ParMultiDo(TimeMonitor) into Collect total bytes/ParMultiDo(ByteMonitor)
May 09, 2022 12:33:00 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-09T12:32:58.758Z: Running job using Streaming Engine
May 09, 2022 12:33:00 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-09T12:32:58.982Z: Starting 5 ****s in us-central1-b...
May 09, 2022 12:33:22 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-09T12:33:21.818Z: Autoscaling: Raised the number of ****s to 4 so that the pipeline can catch up with its backlog and keep up with its input rate.
May 09, 2022 12:33:22 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-09T12:33:21.846Z: Resized **** pool to 4, though goal was 5.  This could be a quota issue.
May 09, 2022 12:33:30 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-09T12:33:29.180Z: Your project already contains 100 Dataflow-created metric descriptors, so new user metrics of the form custom.googleapis.com/* will not be created. However, all user metrics are also available in the metric dataflow.googleapis.com/job/user_counter. If you rely on the custom metrics, you can delete old / unused metric descriptors. See https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.list and https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.delete
May 09, 2022 12:33:33 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-09T12:33:32.071Z: Autoscaling: Raised the number of ****s to 5 so that the pipeline can catch up with its backlog and keep up with its input rate.
May 09, 2022 12:34:35 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-09T12:34:33.040Z: Workers have started successfully.
May 09, 2022 4:01:01 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-09T16:01:00.503Z: Cancel request is committed for workflow job: 2022-05-09_05_32_46-14665142041617253802.
May 09, 2022 4:01:01 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-09T16:01:00.620Z: Cleaning up.
May 09, 2022 4:01:01 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-09T16:01:00.718Z: Stopping **** pool...
May 09, 2022 4:01:01 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-09T16:01:00.768Z: Stopping **** pool...
May 09, 2022 4:01:32 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-09T16:01:32.713Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
May 09, 2022 4:01:32 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-09T16:01:32.750Z: Worker pool stopped.
May 09, 2022 4:01:41 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-05-09_05_32_46-14665142041617253802 finished with status CANCELLED.
Load test results for test (ID): a6fbb7f6-1e3f-46ae-9feb-1181e26d3207 and timestamp: 2022-05-09T12:32:40.038000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                 12338.085
dataflow_v2_java17_total_bytes_count             2.86964364E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220509123035
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:ae661c08a40bbf8a47368c1570d2cbdb9dbb0865ef587bd35d60f43cbee2a87d
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220509123035]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:ae661c08a40bbf8a47368c1570d2cbdb9dbb0865ef587bd35d60f43cbee2a87d]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220509123035] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:ae661c08a40bbf8a47368c1570d2cbdb9dbb0865ef587bd35d60f43cbee2a87d])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:ae661c08a40bbf8a47368c1570d2cbdb9dbb0865ef587bd35d60f43cbee2a87d
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:ae661c08a40bbf8a47368c1570d2cbdb9dbb0865ef587bd35d60f43cbee2a87d
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:ae661c08a40bbf8a47368c1570d2cbdb9dbb0865ef587bd35d60f43cbee2a87d].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.4/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 31m 27s
110 actionable tasks: 73 executed, 33 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/5fdkmmdldigby

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #139

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/139/display/redirect>

Changes:


------------------------------------------
[...truncated 50.84 KB...]
08fa02ce37eb: Preparing
a037458de4e0: Preparing
bafdbe68e4ae: Preparing
a13c519c6361: Preparing
fced9607bc47: Waiting
e91b08f16562: Waiting
713082863eee: Waiting
08e8a29c256a: Waiting
00d4a583dbe4: Waiting
875a60575306: Waiting
c92dc92e5b42: Waiting
721b8ef4b4ba: Waiting
b900d4df52f8: Waiting
5b26c8721c69: Waiting
3bc383470c05: Waiting
bafdbe68e4ae: Waiting
a13c519c6361: Waiting
e93827457889: Waiting
08fa02ce37eb: Waiting
a037458de4e0: Waiting
fbe1b45afd3e: Pushed
c8df78b28fbb: Pushed
e023246e7a9a: Pushed
d580cb531531: Pushed
de4ea1bb1e74: Pushed
875a60575306: Pushed
721b8ef4b4ba: Pushed
b900d4df52f8: Pushed
5b26c8721c69: Pushed
713082863eee: Pushed
e91b08f16562: Pushed
fced9607bc47: Pushed
e93827457889: Layer already exists
08fa02ce37eb: Layer already exists
3bc383470c05: Layer already exists
a037458de4e0: Layer already exists
bafdbe68e4ae: Layer already exists
a13c519c6361: Layer already exists
00d4a583dbe4: Pushed
c92dc92e5b42: Pushed
08e8a29c256a: Pushed
20220508123043: digest: sha256:3e2bc6f9b92f30686d7581e866cc3457a9e65cb36f5ee4f62689dd7b0f578fb4 size: 4729

> Task :sdks:java:testing:load-tests:run
May 08, 2022 12:33:00 PM org.apache.beam.runners.dataflow.options.DataflowPipelineOptions$StagingLocationFactory create
INFO: No stagingLocation provided, falling back to gcpTempLocation
May 08, 2022 12:33:01 PM org.apache.beam.runners.dataflow.DataflowRunner fromOptions
INFO: PipelineOptions.filesToStage was not specified. Defaulting to files from the classpath: will stage 222 files. Enable logging at DEBUG level to see which files will be staged.
May 08, 2022 12:33:01 PM org.apache.beam.sdk.Pipeline validate
WARNING: The following transforms do not have stable unique names: Window.Into()
May 08, 2022 12:33:01 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Executing pipeline on the Dataflow Service, which will have billing implications related to Google Compute Engine usage and other Google Cloud Services.
May 08, 2022 12:33:05 PM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Uploading 222 files from PipelineOptions.filesToStage to staging location to prepare for execution.
May 08, 2022 12:33:06 PM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Staging files complete: 222 files cached, 0 files newly uploaded in 0 seconds
May 08, 2022 12:33:06 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Staging portable pipeline proto to gs://temp-storage-for-perf-tests/loadtests/staging/
May 08, 2022 12:33:06 PM org.apache.beam.runners.dataflow.util.PackageUtil tryStagePackage
INFO: Uploading <120277 bytes, hash 6c4231b8b37219127e2cbe678df05f694b012a4e308249a674ec8a16db5c1949> to gs://temp-storage-for-perf-tests/loadtests/staging/pipeline-bEIxuLNyGRJ-LL5njfBfaUsBKk4wgkmmdOyKFttcGUk.pb
May 08, 2022 12:33:08 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read input/DataflowRunner.StreamingUnboundedRead.ReadWithIds as step s1
May 08, 2022 12:33:08 PM org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource split
INFO: Split into 20 bundles of sizes: [org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@751ae8a4, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@235d659c, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4232b34a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2da16263, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@f5ce0bb, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@47e51549, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@101a461c, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@360e9c06, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5ebffb44, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@311ff287, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7377781e, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@31db34da, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@109f8c7e, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5edacf20, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@16a5eb6d, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7e307087, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1220ef43, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1a8b81e8, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@234cff57, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5e8507f1]
May 08, 2022 12:33:08 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read input/StripIds as step s2
May 08, 2022 12:33:08 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect start time metrics (input) as step s3
May 08, 2022 12:33:08 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Window.Into()/Window.Assign as step s4
May 08, 2022 12:33:08 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read co-input/DataflowRunner.StreamingUnboundedRead.ReadWithIds as step s5
May 08, 2022 12:33:08 PM org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource split
INFO: Split into 20 bundles of sizes: [org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3bf54172, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2c9a6717, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7b3cde6f, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6d091cad, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7c663eaf, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3bb5ceb, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1e692555, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3ba0ae41, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@76fe6cdc, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2ffb3aec, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@786ff1cb, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@46039a21, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@431e86b1, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@35c4e864, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@32a2a6be, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@682af059, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5f36c8e3, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4da39ca9, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6a9344f5, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5584d9c6]
May 08, 2022 12:33:08 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read co-input/StripIds as step s6
May 08, 2022 12:33:08 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect start time metrics (co-input) as step s7
May 08, 2022 12:33:08 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Window.Into()2/Window.Assign as step s8
May 08, 2022 12:33:08 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/MakeUnionTable0 as step s9
May 08, 2022 12:33:08 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/MakeUnionTable1 as step s10
May 08, 2022 12:33:08 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/Flatten as step s11
May 08, 2022 12:33:08 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/GBK as step s12
May 08, 2022 12:33:08 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/ConstructCoGbkResultFn as step s13
May 08, 2022 12:33:08 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Ungroup and reiterate as step s14
May 08, 2022 12:33:08 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect total bytes as step s15
May 08, 2022 12:33:08 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect end time metrics as step s16
May 08, 2022 12:33:08 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Dataflow SDK version: 2.40.0-SNAPSHOT
May 08, 2022 12:33:09 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To access the Dataflow monitoring console, please navigate to https://console.cloud.google.com/dataflow/jobs/us-central1/2022-05-08_05_33_08-12380385818833013014?project=apache-beam-testing
May 08, 2022 12:33:09 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Submitted job: 2022-05-08_05_33_08-12380385818833013014
May 08, 2022 12:33:09 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To cancel the job using the 'gcloud' tool, run:
> gcloud dataflow jobs --project=apache-beam-testing cancel --region=us-central1 2022-05-08_05_33_08-12380385818833013014
May 08, 2022 12:33:18 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
WARNING: 2022-05-08T12:33:14.423Z: The workflow name is not a valid Cloud Label. Labels applied to Cloud resources (such as GCE Instances) for monitoring will be labeled with this modified job name: load0tests0java170dataflow0v20streaming0cogbk01-jenkins-05-v0tj. For the best monitoring experience, please name your job with a valid Cloud Label. For details, see: https://cloud.google.com/compute/docs/labeling-resources#restrictions
May 08, 2022 12:33:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-08T12:33:25.065Z: Worker configuration: e2-standard-2 in us-central1-b.
May 08, 2022 12:33:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-08T12:33:25.642Z: Expanding SplittableParDo operations into optimizable parts.
May 08, 2022 12:33:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-08T12:33:25.737Z: Expanding CollectionToSingleton operations into optimizable parts.
May 08, 2022 12:33:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-08T12:33:25.818Z: Expanding CoGroupByKey operations into optimizable parts.
May 08, 2022 12:33:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-08T12:33:25.888Z: Expanding SplittableProcessKeyed operations into optimizable parts.
May 08, 2022 12:33:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-08T12:33:25.917Z: Expanding GroupByKey operations into streaming Read/Write steps
May 08, 2022 12:33:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-08T12:33:25.983Z: Lifting ValueCombiningMappingFns into MergeBucketsMappingFns
May 08, 2022 12:33:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-08T12:33:26.099Z: Fusing adjacent ParDo, Read, Write, and Flatten operations
May 08, 2022 12:33:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-08T12:33:26.130Z: Unzipping flatten CoGroupByKey-Flatten for input CoGroupByKey-MakeUnionTable0-ParMultiDo-ConstructUnionTable-.output
May 08, 2022 12:33:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-08T12:33:26.164Z: Fusing unzipped copy of CoGroupByKey/GBK/WriteStream, through flatten CoGroupByKey/Flatten, into producer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable)
May 08, 2022 12:33:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-08T12:33:26.199Z: Fusing consumer CoGroupByKey/GBK/WriteStream into CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable)
May 08, 2022 12:33:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-08T12:33:26.230Z: Fusing consumer Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read input/Impulse
May 08, 2022 12:33:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-08T12:33:26.270Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
May 08, 2022 12:33:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-08T12:33:26.306Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
May 08, 2022 12:33:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-08T12:33:26.340Z: Fusing consumer Read input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
May 08, 2022 12:33:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-08T12:33:26.371Z: Fusing consumer Collect start time metrics (input)/ParMultiDo(TimeMonitor) into Read input/ParDo(StripIds)/ParMultiDo(StripIds)
May 08, 2022 12:33:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-08T12:33:26.410Z: Fusing consumer Window.Into()/Window.Assign into Collect start time metrics (input)/ParMultiDo(TimeMonitor)
May 08, 2022 12:33:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-08T12:33:26.443Z: Fusing consumer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable) into Window.Into()/Window.Assign
May 08, 2022 12:33:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-08T12:33:26.476Z: Fusing consumer Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read co-input/Impulse
May 08, 2022 12:33:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-08T12:33:26.509Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
May 08, 2022 12:33:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-08T12:33:26.542Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
May 08, 2022 12:33:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-08T12:33:26.568Z: Fusing consumer Read co-input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
May 08, 2022 12:33:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-08T12:33:26.605Z: Fusing consumer Collect start time metrics (co-input)/ParMultiDo(TimeMonitor) into Read co-input/ParDo(StripIds)/ParMultiDo(StripIds)
May 08, 2022 12:33:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-08T12:33:26.639Z: Fusing consumer Window.Into()2/Window.Assign into Collect start time metrics (co-input)/ParMultiDo(TimeMonitor)
May 08, 2022 12:33:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-08T12:33:26.736Z: Fusing consumer CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable) into Window.Into()2/Window.Assign
May 08, 2022 12:33:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-08T12:33:26.771Z: Fusing consumer CoGroupByKey/GBK/MergeBuckets into CoGroupByKey/GBK/ReadStream
May 08, 2022 12:33:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-08T12:33:26.802Z: Fusing consumer CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult) into CoGroupByKey/GBK/MergeBuckets
May 08, 2022 12:33:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-08T12:33:26.836Z: Fusing consumer Ungroup and reiterate/ParMultiDo(UngroupAndReiterate) into CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult)
May 08, 2022 12:33:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-08T12:33:26.872Z: Fusing consumer Collect total bytes/ParMultiDo(ByteMonitor) into Ungroup and reiterate/ParMultiDo(UngroupAndReiterate)
May 08, 2022 12:33:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-08T12:33:26.897Z: Fusing consumer Collect end time metrics/ParMultiDo(TimeMonitor) into Collect total bytes/ParMultiDo(ByteMonitor)
May 08, 2022 12:33:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-08T12:33:27.139Z: Running job using Streaming Engine
May 08, 2022 12:33:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-08T12:33:27.534Z: Starting 5 ****s in us-central1-b...
May 08, 2022 12:33:42 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-08T12:33:41.960Z: Your project already contains 100 Dataflow-created metric descriptors, so new user metrics of the form custom.googleapis.com/* will not be created. However, all user metrics are also available in the metric dataflow.googleapis.com/job/user_counter. If you rely on the custom metrics, you can delete old / unused metric descriptors. See https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.list and https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.delete
May 08, 2022 12:33:51 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-08T12:33:49.913Z: Autoscaling: Raised the number of ****s to 4 so that the pipeline can catch up with its backlog and keep up with its input rate.
May 08, 2022 12:33:51 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-08T12:33:49.939Z: Resized **** pool to 4, though goal was 5.  This could be a quota issue.
May 08, 2022 12:34:01 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-08T12:34:00.147Z: Autoscaling: Raised the number of ****s to 5 so that the pipeline can catch up with its backlog and keep up with its input rate.
May 08, 2022 12:34:58 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-08T12:34:57.660Z: Workers have started successfully.
May 08, 2022 4:01:13 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-08T16:01:12.049Z: Cancel request is committed for workflow job: 2022-05-08_05_33_08-12380385818833013014.
May 08, 2022 4:01:13 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-08T16:01:12.227Z: Cleaning up.
May 08, 2022 4:01:13 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-08T16:01:12.337Z: Stopping **** pool...
May 08, 2022 4:01:13 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-08T16:01:12.380Z: Stopping **** pool...
May 08, 2022 4:01:45 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-08T16:01:45.060Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
May 08, 2022 4:01:45 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-08T16:01:45.128Z: Worker pool stopped.
May 08, 2022 4:01:52 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-05-08_05_33_08-12380385818833013014 finished with status CANCELLED.
Load test results for test (ID): 6898ae29-0c8e-440a-aed1-3e8098cdf3b0 and timestamp: 2022-05-08T12:33:01.540000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                 12302.848
dataflow_v2_java17_total_bytes_count             2.44242771E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220508123043
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:3e2bc6f9b92f30686d7581e866cc3457a9e65cb36f5ee4f62689dd7b0f578fb4
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220508123043]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:3e2bc6f9b92f30686d7581e866cc3457a9e65cb36f5ee4f62689dd7b0f578fb4]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220508123043] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:3e2bc6f9b92f30686d7581e866cc3457a9e65cb36f5ee4f62689dd7b0f578fb4])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:3e2bc6f9b92f30686d7581e866cc3457a9e65cb36f5ee4f62689dd7b0f578fb4
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:3e2bc6f9b92f30686d7581e866cc3457a9e65cb36f5ee4f62689dd7b0f578fb4
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:3e2bc6f9b92f30686d7581e866cc3457a9e65cb36f5ee4f62689dd7b0f578fb4].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.4/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 31m 31s
110 actionable tasks: 73 executed, 33 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/a3nj7pq4szqmc

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #138

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/138/display/redirect?page=changes>

Changes:

[kevinsijo] Setting up a basic directory

[kevinsijo] Mirroring Python SDK's directory structure

[kerrydc] Adds initial tests

[kevinsijo] 'runners' is the correct directory name

[Pablo Estrada] sketching the core API for JS SDK

[jonathanlui] add .gitignore for node/ts project

[Robert Bradshaw] Worker directory.

[Robert Bradshaw] Fix complile errors with explicit any for callables.

[Robert Bradshaw] Add worker entry point.

[Robert Bradshaw] Add proto generation code.

[Robert Bradshaw] Add generated proto files.

[Robert Bradshaw] Attempts to get ts protos to compile.

[Robert Bradshaw] Exclude ts protos for now.

[Robert Bradshaw] More changes to get ts protos working.

[Robert Bradshaw] Update scripts and config to get protos compiling.

[Robert Bradshaw] Update geenrated files.

[jonathanlui] add build and clean script to compile ts

[Robert Bradshaw] Generate server for loopback worker.

[Robert Bradshaw] Generated grpc servers for loopback.

[Robert Bradshaw] Add typescript formatter.

[Robert Bradshaw] Loopback server (that does nothing).

[Robert Bradshaw] Working server.

[Pablo Estrada] Starting expansion of primitive transforms

[Pablo Estrada] Starting to implement and support standard coders

[Robert Bradshaw] Also generate grpc clients.

[Robert Bradshaw] Basic implementation of worker harness.

[Pablo Estrada] fix the build

[Robert Bradshaw] Add some missing files for worker harness.

[Robert Bradshaw] Refactor operators to use registration.

[jonathanlui] enable ts in mocha

[jonathanlui] update readme

[jonathanlui] --save-dev @types/mocha

[jonathanlui] translate core_test.js to typescript

[Robert Bradshaw] Encapsulate worker service in a class.

[Kenneth Knowles] Port standard_coders_test to typescript (superficially)

[Pablo Estrada] Starting the proto translation of Impulse, ParDo, GBK

[Robert Bradshaw] Add some tests for the worker code.

[Robert Bradshaw] Fixing old lock file error.

[Pablo Estrada] Adding transform names and fixing GBK coder issue

[Robert Bradshaw] npx tsfmt -r src/apache_beam/base.ts src/apache_beam/transforms/core.ts

[Kenneth Knowles] switch to import style require() statements

[Kenneth Knowles] Add Coder interface using protobufjs classes

[Kenneth Knowles] BytesCoder with some failures

[noreply] Added GeneralObjectCoder and using it as coder for most transforms (#9)

[Kenneth Knowles] Fix order of arguments to deepEqual

[Kenneth Knowles] Encode expected encoding as binary

[Robert Bradshaw] Refactor API to allow for composites.

[jrmccluskey] Initial setup for automated Java expansion startup

[jrmccluskey] Update exp_service.ts

[Kenneth Knowles] Fix up coder deserialization

[Robert Bradshaw] Simplify GBK coder computation.

[Robert Bradshaw] Remove top-level PValue.

[Pablo Estrada] Make tests green

[Robert Bradshaw] Rename PValueish to PValue.

[jonathanlui] node runner

[jonathanlui] whitespaces

[Robert Bradshaw] Make Runner.run async.

[jonathanlui] bson and fast-deep-equal should not be listed as devdependency

[jrmccluskey] Add basic Dockerfile that starts ExternalWorkerPool

[Robert Bradshaw] Direct runner.

[kevinsijo] Testing expansion service communication

[Robert Bradshaw] Added flatten, assertion checkers.

[Pablo Estrada] progress on basic coders

[Robert Bradshaw] Fixing the build.

[Robert Bradshaw] Cleanup, simplify access.

[Pablo Estrada] Adding limited support for KVCoder and IterableCoder

[Robert Bradshaw] Introduce PipelineContext.

[Robert Bradshaw] Add toProto to all coders.

[Robert Bradshaw] Some work with coders.

[Robert Bradshaw] Remove debug logging.

[Robert Bradshaw] Use coders over data channel.

[Kenneth Knowles] explicitly sequence sub-coder serializations

[Kenneth Knowles] no more need to extend FakeCoder

[Kenneth Knowles] actually advance reader

[Kenneth Knowles] autoformat

[Kenneth Knowles] protobufjs already can write and read signed varints

[Kenneth Knowles] with improved test harness, kv has many more failures

[Kenneth Knowles] read bytescoder from correct position

[Kenneth Knowles] no more fake coders

[Kenneth Knowles] varint examples all work

[Kenneth Knowles] simplify coder value parsing

[Kenneth Knowles] global window coder

[Kenneth Knowles] fix swapEndian32

[Robert Bradshaw] Add P(...) operator.

[kevinsijo] Implementing RowCoder encoding.

[jrmccluskey] remove unused container dir

[kevinsijo] Corrected sorting of encoded positions to reflect an argsort instead.

[Robert Bradshaw] Populate environments.

[kevinsijo] Implementing RowCoder decoding.

[Kenneth Knowles] preliminary unbounded iterable coder

[Kenneth Knowles] friendlier description of standard coder test case

[Kenneth Knowles] fix test harness; iterable works

[jrmccluskey] first pass at boot.go

[jonathanlui] update package-lock.json

[jonathanlui] make NodeRunner a subclass of Runner

[jonathanlui] add waitUntilFinish interface member

[Pablo Estrada] Adding double coder

[Kenneth Knowles] scaffolding for windowed values

[Pablo Estrada] Adding type information to PColleciton and PTransform

[jonathanlui] fix direct runner

[Pablo Estrada] Adding typing information for DoFns

[Kenneth Knowles] add interval window

[Robert Bradshaw] Export PValue.

[Robert Bradshaw] Add CombineFn interface.

[Robert Bradshaw] Typed flatten.

[jonathanlui] add runAsync method to base.Runner

[Kenneth Knowles] add Long package

[Pablo Estrada] Adding more types. Making PValue typed

[Kenneth Knowles] instant coder draft

[Robert Bradshaw] Return job state from direct runner.

[Kenneth Knowles] type instant = long

[jonathanlui] implement NodeRunner.runPipeline

[Kenneth Knowles] autoformat

[kevinsijo] Completed implementation of basic row coder

[Kenneth Knowles] Fix IntervalWindowCoder, almost

[Kenneth Knowles] fix interval window coder

[Kenneth Knowles] autoformat

[Robert Bradshaw] loopback runner works

[Kenneth Knowles] move core element types into values.ts

[Kenneth Knowles] just build object directly to be cool

[Robert Bradshaw] GBK working on ULR.

[Robert Bradshaw] Async transforms.

[Robert Bradshaw] External transform grpah splicing.

[Kenneth Knowles] progress on windowed value: paneinfo encoding

[Robert Bradshaw] Fix merge.

[Robert Bradshaw] autoformat

[Kenneth Knowles] full windowed value coder

[kerrydc] Updates tests to use correct types, adds generics where needed to DoFns

[Robert Bradshaw] Add serialization librarires.'

[Robert Bradshaw] Add Split() PTransform, for producing multiple outputs from a single

[Robert Bradshaw] Schema-encoded external payloads.

[kevinsijo] Adding Schema inference from JSON

[Pablo Estrada] Removing unused directories

[Pablo Estrada] Support for finishBundle and improving typing annotations.

[Pablo Estrada] A base implementation of combiners with GBK/ParDo

[Robert Bradshaw] Fully propagate windowing information in both remote and direct runner.

[Robert Bradshaw] Make args and kwargs optional for python external transform.

[Robert Bradshaw] Infer schema for external transforms.

[Pablo Estrada] Implementing a custom combine fn as an example. Small fixes

[Robert Bradshaw] Fix missing windowing information in combiners.

[Robert Bradshaw] PostShuffle needn't group by key as that's already done.

[Robert Bradshaw] Guard pre-combine for global window only.

[Robert Bradshaw] WindowInto

[Robert Bradshaw] Fix optional kwargs.

[Robert Bradshaw] A couple of tweaks for js + py

[Robert Bradshaw] Add windowing file.

[Robert Bradshaw] CombineBy transform, stand-alone WordCount.

[Robert Bradshaw] cleanup

[Robert Bradshaw] Actually fix optional external kwargs.

[Robert Bradshaw] Demo2, textio read.

[Robert Bradshaw] Add command lines for starting up the servers.

[Robert Bradshaw] Run prettier on the full codebase.

[Robert Bradshaw] Update deps.

[Pablo Estrada] Adding docstrings for core.ts. Prettier dependency

[Pablo Estrada] Documenting coder interfaces

[Pablo Estrada] Added documentation for a few standard coders

[Robert Bradshaw] Unified grouping and combining.

[Robert Bradshaw] Allow PCollection ids to be lazy.

[Robert Bradshaw] Reorganize module structure.

[Robert Bradshaw] A couple more renames.

[Robert Bradshaw] Simplify.

[Robert Bradshaw] Consolidation.

[Robert Bradshaw] Fix build.

[Robert Bradshaw] Add optional context to ParDo.

[Robert Bradshaw] fixup: iterable coder endian sign issue

[Robert Bradshaw] omit context for map(console.log)

[Robert Bradshaw] Fix ReadFromText coders.

[Robert Bradshaw] Flesh out README with overview and current state.

[noreply] Readme typo

[Robert Bradshaw] Two more TODOs.

[noreply] Add a pointer to the example wordcount to the readme.

[Pablo Estrada] Documenting coders and implementing unknown-length method

[Robert Bradshaw] UIID dependency.

[Robert Bradshaw] Artifact handling.

[Robert Bradshaw] Properly wait on data channel for bundle completion.

[Robert Bradshaw] Automatic java expansion service startup.

[Robert Bradshaw] Process promises.

[Robert Bradshaw] Implement side inputs.

[Robert Bradshaw] Cleanup.

[Robert Bradshaw] Put complex constext stuff in its own file.

[Robert Bradshaw] Rename BoundedWindow to just Window.

[Robert Bradshaw] Alternative splitter class.

[Pablo Estrada] Documenting internal functions

[Robert Bradshaw] Take a pass clarifying the TODOs.

[Robert Bradshaw] Sql transform wrapper.

[Robert Bradshaw] Incorporate some feedback into the TODOs.

[Robert Bradshaw] More TODOs.

[Robert Bradshaw] Remove app placeholder.

[Robert Bradshaw] Apache license headers.

[Robert Bradshaw] More TODOs

[jankuehle] Suggestions for TypeScript todos

[dannymccormick] Add actions for typescript sdk

[dannymccormick] Fix test command

[noreply] Add missing version

[dannymccormick] Fix codecovTest command

[noreply] Only do prettier check on linux

[noreply] Only get codecov on linux

[Robert Bradshaw] Resolve some comments.

[Robert Bradshaw] Fix compile errors.

[Robert Bradshaw] Prettier.

[Robert Bradshaw] Re-order expandInternal arguments pending unification.

[Robert Bradshaw] More consistent and stricter PTransform naming.

[Robert Bradshaw] Notes on explicit, if less idiomatic, use of classes.

[Robert Bradshaw] Let DoFn be an interface rather than a class.

[Robert Bradshaw] Provide DoFn context to start and finish bundle.

[Robert Bradshaw] Optional promise code simplification.

[Robert Bradshaw] Cleanup todos.

[Robert Bradshaw] Avoid any type where not needed.

[Robert Bradshaw] Apache RAT excludes for typescript.

[Robert Bradshaw] Remove empty READMEs.

[Robert Bradshaw] Add licences statement to readme files.

[Robert Bradshaw] More RAT fixes.

[Robert Bradshaw] Another unsupported coder.

[Robert Bradshaw] Remove debugging code.

[noreply] Fix automatic naming with code coverage.

[Robert Bradshaw] Coders cleanup.

[Robert Bradshaw] Add tests for RowCoder.

[Robert Bradshaw] Normalize capitalization, comments.

[Robert Bradshaw] Install typescript closure packages.

[Robert Bradshaw] npm audit fix

[Robert Bradshaw] Move more imports out of base.

[Robert Bradshaw] Changes needed to compile with ts closure plugin.

[Robert Bradshaw] Use ttsc and ts-closure-transform plugin.

[Robert Bradshaw] Serialization registration to actually get serialization working.

[Robert Bradshaw] Container images working on local runner.

[Robert Bradshaw] Add a portable job server that proxies the Dataflow backend. (#17189)

[Robert Bradshaw] Improvements to dataflow job service for non-Python jobs.

[Robert Bradshaw] Get dataflow working.

[Robert Bradshaw] User friendly pipeline options.

[Robert Bradshaw] Less classes, more functions.

[Robert Bradshaw] Add new nullable standard coder.

[Robert Bradshaw] Make Apache Rat happy.

[Robert Bradshaw] Disable broken codecov.

[Robert Bradshaw] Remove last uses of base.ts.

[Robert Bradshaw] Remove unneedd file.

[Robert Bradshaw] Remove more uneeded/unused files.

[Robert Bradshaw] Cleanup tests.

[Robert Bradshaw] Minor cleanups to coder tests.

[noreply] Quote pip install package name

[noreply] [BEAM-14374] Fix module import error in FullyQualifiedNamedTransform

[Robert Bradshaw] Addressing issues from the review.

[noreply] Apply suggestions from code review.

[Robert Bradshaw] Post-merge fixes.

[dannymccormick] Delete tags.go

[Robert Bradshaw] Update tests to use our actual serialization libraries.

[Robert Bradshaw] Another pass at TODOs, removing finished items.

[Heejong Lee] [BEAM-14146] Python Streaming job failing to drain with BigQueryIO write

[Kenneth Knowles] Add parameter for service account impersonation in GCP credentials

[Heejong Lee] add test

[noreply] Merge pull request #17490 from [BEAM-14370] [Website] Add new page about

[noreply] [BEAM-14332] Refactored cluster management for Flink on Dataproc

[noreply] [BEAM-13988] Update mtime to use time.UnixMilli() calls (#17578)

[noreply] Fixing patching error on missing dependencies (#17564)

[noreply] Merge pull request #17517 from [BEAM-14383] Improve "FailedRows" errors

[Heejong Lee] add test without mock


------------------------------------------
[...truncated 50.18 KB...]
2d05db1e2a4b: Preparing
3bc383470c05: Preparing
e93827457889: Preparing
08fa02ce37eb: Preparing
a037458de4e0: Preparing
662540b7db15: Waiting
1b1b55724774: Waiting
bafdbe68e4ae: Preparing
0e2e00df1023: Waiting
d507830c9b22: Waiting
a13c519c6361: Preparing
08fa02ce37eb: Waiting
82227587df8f: Waiting
e93827457889: Waiting
3bc383470c05: Waiting
bafdbe68e4ae: Waiting
a13c519c6361: Waiting
1252636fdc5b: Waiting
b0bdfb1e5c67: Waiting
8db9d37c9268: Waiting
2d05db1e2a4b: Waiting
a037458de4e0: Waiting
413656da2079: Pushed
98faf80f3452: Pushed
94899fdeb521: Pushed
3c9402bf3f9b: Pushed
1e44127ca56a: Pushed
662540b7db15: Pushed
104aa3e0eab7: Pushed
82227587df8f: Pushed
d507830c9b22: Pushed
1252636fdc5b: Pushed
1b1b55724774: Pushed
3bc383470c05: Layer already exists
e93827457889: Layer already exists
b0bdfb1e5c67: Pushed
08fa02ce37eb: Layer already exists
a037458de4e0: Layer already exists
bafdbe68e4ae: Layer already exists
a13c519c6361: Layer already exists
8db9d37c9268: Pushed
2d05db1e2a4b: Pushed
0e2e00df1023: Pushed
20220507123039: digest: sha256:a201febb4501343aed27f34781031d37aebfe84ff97646ce1d23d15df7f7d15a size: 4729

> Task :sdks:java:testing:load-tests:run
May 07, 2022 12:32:44 PM org.apache.beam.runners.dataflow.options.DataflowPipelineOptions$StagingLocationFactory create
INFO: No stagingLocation provided, falling back to gcpTempLocation
May 07, 2022 12:32:44 PM org.apache.beam.runners.dataflow.DataflowRunner fromOptions
INFO: PipelineOptions.filesToStage was not specified. Defaulting to files from the classpath: will stage 222 files. Enable logging at DEBUG level to see which files will be staged.
May 07, 2022 12:32:45 PM org.apache.beam.sdk.Pipeline validate
WARNING: The following transforms do not have stable unique names: Window.Into()
May 07, 2022 12:32:45 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Executing pipeline on the Dataflow Service, which will have billing implications related to Google Compute Engine usage and other Google Cloud Services.
May 07, 2022 12:32:48 PM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Uploading 222 files from PipelineOptions.filesToStage to staging location to prepare for execution.
May 07, 2022 12:32:48 PM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Staging files complete: 222 files cached, 0 files newly uploaded in 0 seconds
May 07, 2022 12:32:48 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Staging portable pipeline proto to gs://temp-storage-for-perf-tests/loadtests/staging/
May 07, 2022 12:32:48 PM org.apache.beam.runners.dataflow.util.PackageUtil tryStagePackage
INFO: Uploading <120277 bytes, hash a6ab5cbe90343949e56d702d9a136db2fa719819f168bdcc66299093c41072f1> to gs://temp-storage-for-perf-tests/loadtests/staging/pipeline-pqtcvpA0OUnlbXAtmhNtsvpxmBnxaL3MZimQk8QQcvE.pb
May 07, 2022 12:32:50 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read input/DataflowRunner.StreamingUnboundedRead.ReadWithIds as step s1
May 07, 2022 12:32:50 PM org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource split
INFO: Split into 20 bundles of sizes: [org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@751ae8a4, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@235d659c, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4232b34a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2da16263, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@f5ce0bb, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@47e51549, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@101a461c, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@360e9c06, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5ebffb44, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@311ff287, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7377781e, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@31db34da, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@109f8c7e, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5edacf20, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@16a5eb6d, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7e307087, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1220ef43, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1a8b81e8, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@234cff57, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5e8507f1]
May 07, 2022 12:32:50 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read input/StripIds as step s2
May 07, 2022 12:32:50 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect start time metrics (input) as step s3
May 07, 2022 12:32:50 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Window.Into()/Window.Assign as step s4
May 07, 2022 12:32:50 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read co-input/DataflowRunner.StreamingUnboundedRead.ReadWithIds as step s5
May 07, 2022 12:32:50 PM org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource split
INFO: Split into 20 bundles of sizes: [org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3bf54172, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2c9a6717, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7b3cde6f, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6d091cad, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7c663eaf, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3bb5ceb, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1e692555, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3ba0ae41, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@76fe6cdc, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2ffb3aec, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@786ff1cb, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@46039a21, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@431e86b1, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@35c4e864, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@32a2a6be, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@682af059, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5f36c8e3, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4da39ca9, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6a9344f5, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5584d9c6]
May 07, 2022 12:32:50 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read co-input/StripIds as step s6
May 07, 2022 12:32:50 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect start time metrics (co-input) as step s7
May 07, 2022 12:32:50 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Window.Into()2/Window.Assign as step s8
May 07, 2022 12:32:50 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/MakeUnionTable0 as step s9
May 07, 2022 12:32:50 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/MakeUnionTable1 as step s10
May 07, 2022 12:32:50 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/Flatten as step s11
May 07, 2022 12:32:50 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/GBK as step s12
May 07, 2022 12:32:50 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/ConstructCoGbkResultFn as step s13
May 07, 2022 12:32:50 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Ungroup and reiterate as step s14
May 07, 2022 12:32:50 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect total bytes as step s15
May 07, 2022 12:32:50 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect end time metrics as step s16
May 07, 2022 12:32:50 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Dataflow SDK version: 2.40.0-SNAPSHOT
May 07, 2022 12:32:51 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To access the Dataflow monitoring console, please navigate to https://console.cloud.google.com/dataflow/jobs/us-central1/2022-05-07_05_32_50-18293218377134002814?project=apache-beam-testing
May 07, 2022 12:32:51 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Submitted job: 2022-05-07_05_32_50-18293218377134002814
May 07, 2022 12:32:51 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To cancel the job using the 'gcloud' tool, run:
> gcloud dataflow jobs --project=apache-beam-testing cancel --region=us-central1 2022-05-07_05_32_50-18293218377134002814
May 07, 2022 12:32:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
WARNING: 2022-05-07T12:32:54.377Z: The workflow name is not a valid Cloud Label. Labels applied to Cloud resources (such as GCE Instances) for monitoring will be labeled with this modified job name: load0tests0java170dataflow0v20streaming0cogbk01-jenkins-05-e677. For the best monitoring experience, please name your job with a valid Cloud Label. For details, see: https://cloud.google.com/compute/docs/labeling-resources#restrictions
May 07, 2022 12:33:05 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-07T12:33:03.852Z: Worker configuration: e2-standard-2 in us-central1-b.
May 07, 2022 12:33:05 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-07T12:33:04.691Z: Expanding SplittableParDo operations into optimizable parts.
May 07, 2022 12:33:05 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-07T12:33:04.716Z: Expanding CollectionToSingleton operations into optimizable parts.
May 07, 2022 12:33:05 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-07T12:33:04.786Z: Expanding CoGroupByKey operations into optimizable parts.
May 07, 2022 12:33:05 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-07T12:33:04.935Z: Expanding SplittableProcessKeyed operations into optimizable parts.
May 07, 2022 12:33:05 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-07T12:33:05.134Z: Expanding GroupByKey operations into streaming Read/Write steps
May 07, 2022 12:33:07 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-07T12:33:05.412Z: Lifting ValueCombiningMappingFns into MergeBucketsMappingFns
May 07, 2022 12:33:07 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-07T12:33:05.501Z: Your project already contains 100 Dataflow-created metric descriptors, so new user metrics of the form custom.googleapis.com/* will not be created. However, all user metrics are also available in the metric dataflow.googleapis.com/job/user_counter. If you rely on the custom metrics, you can delete old / unused metric descriptors. See https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.list and https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.delete
May 07, 2022 12:33:07 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-07T12:33:05.614Z: Fusing adjacent ParDo, Read, Write, and Flatten operations
May 07, 2022 12:33:07 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-07T12:33:05.650Z: Unzipping flatten CoGroupByKey-Flatten for input CoGroupByKey-MakeUnionTable0-ParMultiDo-ConstructUnionTable-.output
May 07, 2022 12:33:07 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-07T12:33:05.687Z: Fusing unzipped copy of CoGroupByKey/GBK/WriteStream, through flatten CoGroupByKey/Flatten, into producer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable)
May 07, 2022 12:33:07 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-07T12:33:05.719Z: Fusing consumer CoGroupByKey/GBK/WriteStream into CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable)
May 07, 2022 12:33:07 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-07T12:33:05.753Z: Fusing consumer Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read input/Impulse
May 07, 2022 12:33:07 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-07T12:33:05.786Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
May 07, 2022 12:33:07 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-07T12:33:05.820Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
May 07, 2022 12:33:07 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-07T12:33:05.851Z: Fusing consumer Read input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
May 07, 2022 12:33:07 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-07T12:33:05.884Z: Fusing consumer Collect start time metrics (input)/ParMultiDo(TimeMonitor) into Read input/ParDo(StripIds)/ParMultiDo(StripIds)
May 07, 2022 12:33:07 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-07T12:33:05.917Z: Fusing consumer Window.Into()/Window.Assign into Collect start time metrics (input)/ParMultiDo(TimeMonitor)
May 07, 2022 12:33:07 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-07T12:33:05.942Z: Fusing consumer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable) into Window.Into()/Window.Assign
May 07, 2022 12:33:07 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-07T12:33:05.975Z: Fusing consumer Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read co-input/Impulse
May 07, 2022 12:33:07 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-07T12:33:06.008Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
May 07, 2022 12:33:07 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-07T12:33:06.041Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
May 07, 2022 12:33:07 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-07T12:33:06.101Z: Fusing consumer Read co-input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
May 07, 2022 12:33:07 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-07T12:33:06.123Z: Fusing consumer Collect start time metrics (co-input)/ParMultiDo(TimeMonitor) into Read co-input/ParDo(StripIds)/ParMultiDo(StripIds)
May 07, 2022 12:33:07 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-07T12:33:06.166Z: Fusing consumer Window.Into()2/Window.Assign into Collect start time metrics (co-input)/ParMultiDo(TimeMonitor)
May 07, 2022 12:33:07 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-07T12:33:06.199Z: Fusing consumer CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable) into Window.Into()2/Window.Assign
May 07, 2022 12:33:07 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-07T12:33:06.232Z: Fusing consumer CoGroupByKey/GBK/MergeBuckets into CoGroupByKey/GBK/ReadStream
May 07, 2022 12:33:07 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-07T12:33:06.266Z: Fusing consumer CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult) into CoGroupByKey/GBK/MergeBuckets
May 07, 2022 12:33:07 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-07T12:33:06.300Z: Fusing consumer Ungroup and reiterate/ParMultiDo(UngroupAndReiterate) into CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult)
May 07, 2022 12:33:07 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-07T12:33:06.333Z: Fusing consumer Collect total bytes/ParMultiDo(ByteMonitor) into Ungroup and reiterate/ParMultiDo(UngroupAndReiterate)
May 07, 2022 12:33:07 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-07T12:33:06.365Z: Fusing consumer Collect end time metrics/ParMultiDo(TimeMonitor) into Collect total bytes/ParMultiDo(ByteMonitor)
May 07, 2022 12:33:07 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-07T12:33:06.562Z: Running job using Streaming Engine
May 07, 2022 12:33:07 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-07T12:33:06.802Z: Starting 5 ****s in us-central1-b...
May 07, 2022 12:33:33 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-07T12:33:32.472Z: Autoscaling: Raised the number of ****s to 5 so that the pipeline can catch up with its backlog and keep up with its input rate.
May 07, 2022 12:34:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-07T12:34:39.607Z: Workers have started successfully.
May 07, 2022 1:13:26 PM org.apache.beam.sdk.metrics.MetricsEnvironment getCurrentContainer
WARNING: Reporting metrics are not supported in the current execution environment.
May 07, 2022 4:01:01 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-07T16:00:58.218Z: Cancel request is committed for workflow job: 2022-05-07_05_32_50-18293218377134002814.
May 07, 2022 4:01:01 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-07T16:00:59.490Z: Cleaning up.
May 07, 2022 4:01:01 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-07T16:00:59.579Z: Stopping **** pool...
May 07, 2022 4:01:01 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-07T16:00:59.623Z: Stopping **** pool...
May 07, 2022 4:01:35 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-07T16:01:35.077Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
May 07, 2022 4:01:35 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-07T16:01:35.117Z: Worker pool stopped.
May 07, 2022 4:01:45 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-05-07_05_32_50-18293218377134002814 finished with status CANCELLED.
Load test results for test (ID): d380543f-25c8-4015-8b36-36a9fa03eb93 and timestamp: 2022-05-07T12:32:45.055000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                 12349.746
dataflow_v2_java17_total_bytes_count             3.37999092E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220507123039
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:a201febb4501343aed27f34781031d37aebfe84ff97646ce1d23d15df7f7d15a
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220507123039]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:a201febb4501343aed27f34781031d37aebfe84ff97646ce1d23d15df7f7d15a]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220507123039] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:a201febb4501343aed27f34781031d37aebfe84ff97646ce1d23d15df7f7d15a])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:a201febb4501343aed27f34781031d37aebfe84ff97646ce1d23d15df7f7d15a
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:a201febb4501343aed27f34781031d37aebfe84ff97646ce1d23d15df7f7d15a
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:a201febb4501343aed27f34781031d37aebfe84ff97646ce1d23d15df7f7d15a].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.4/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 31m 26s
110 actionable tasks: 73 executed, 33 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/icdmjakku6vew

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure


---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #137

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/137/display/redirect?page=changes>

Changes:

[zyichi] Move master readme.md to 2.40.0

[noreply] [BEAM-14173] Fix Go Loadtests on Dataflow & partial fix for Flink

[noreply] Upgrade python sdk container requirements. (#17549)

[noreply] Merge pull request #17497: [BEAM-11205] Update GCP Libraries BOM version

[noreply] [BEAM-12603] Add retry on grpc data channel and remove retry from test.

[noreply] Merge pull request #17359: [BEAM-14303] Add a way to exclude output

[noreply] [BEAM-14347] Allow users to optimize DoFn execution with a single

[noreply] [BEAM-5878] Add (failing) kwonly-argument test (#17509)


------------------------------------------
[...truncated 47.03 KB...]
6736907aa128: Pushed
3bc383470c05: Layer already exists
e93827457889: Layer already exists
e0251c2cfd97: Pushed
08fa02ce37eb: Layer already exists
a037458de4e0: Layer already exists
bafdbe68e4ae: Layer already exists
a13c519c6361: Layer already exists
c21785f068bb: Pushed
feb9704976ce: Pushed
a73b38700550: Pushed
20220506123132: digest: sha256:df7a799483a8dfcf0e4964b95e5dc6ef838d5908578aa974f5608b4b2f79498f size: 4729

> Task :sdks:java:testing:load-tests:run
May 06, 2022 12:35:55 PM org.apache.beam.runners.dataflow.options.DataflowPipelineOptions$StagingLocationFactory create
INFO: No stagingLocation provided, falling back to gcpTempLocation
May 06, 2022 12:35:56 PM org.apache.beam.runners.dataflow.DataflowRunner fromOptions
INFO: PipelineOptions.filesToStage was not specified. Defaulting to files from the classpath: will stage 222 files. Enable logging at DEBUG level to see which files will be staged.
May 06, 2022 12:35:57 PM org.apache.beam.sdk.Pipeline validate
WARNING: The following transforms do not have stable unique names: Window.Into()
May 06, 2022 12:35:57 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Executing pipeline on the Dataflow Service, which will have billing implications related to Google Compute Engine usage and other Google Cloud Services.
May 06, 2022 12:36:03 PM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Uploading 222 files from PipelineOptions.filesToStage to staging location to prepare for execution.
May 06, 2022 12:36:05 PM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Staging files complete: 222 files cached, 0 files newly uploaded in 2 seconds
May 06, 2022 12:36:05 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Staging portable pipeline proto to gs://temp-storage-for-perf-tests/loadtests/staging/
May 06, 2022 12:36:05 PM org.apache.beam.runners.dataflow.util.PackageUtil tryStagePackage
INFO: Uploading <120277 bytes, hash 4c05446f1cc0b7cb866bcf3be6df3ca8b2912a2d7b39757dd38920d959660589> to gs://temp-storage-for-perf-tests/loadtests/staging/pipeline-TAVEbxzAt8uGa8875t88qLKRKi17OXV904kg2VlmBYk.pb
May 06, 2022 12:36:08 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read input/DataflowRunner.StreamingUnboundedRead.ReadWithIds as step s1
May 06, 2022 12:36:08 PM org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource split
INFO: Split into 20 bundles of sizes: [org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@751ae8a4, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@235d659c, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4232b34a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2da16263, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@f5ce0bb, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@47e51549, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@101a461c, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@360e9c06, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5ebffb44, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@311ff287, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7377781e, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@31db34da, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@109f8c7e, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5edacf20, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@16a5eb6d, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7e307087, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1220ef43, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1a8b81e8, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@234cff57, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5e8507f1]
May 06, 2022 12:36:08 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read input/StripIds as step s2
May 06, 2022 12:36:08 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect start time metrics (input) as step s3
May 06, 2022 12:36:08 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Window.Into()/Window.Assign as step s4
May 06, 2022 12:36:08 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read co-input/DataflowRunner.StreamingUnboundedRead.ReadWithIds as step s5
May 06, 2022 12:36:08 PM org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource split
INFO: Split into 20 bundles of sizes: [org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3b95a6db, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3bf54172, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2c9a6717, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7b3cde6f, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6d091cad, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7c663eaf, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3bb5ceb, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1e692555, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3ba0ae41, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@76fe6cdc, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2ffb3aec, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@786ff1cb, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@46039a21, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@431e86b1, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@35c4e864, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@32a2a6be, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@682af059, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5f36c8e3, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4da39ca9, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6a9344f5]
May 06, 2022 12:36:08 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read co-input/StripIds as step s6
May 06, 2022 12:36:08 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect start time metrics (co-input) as step s7
May 06, 2022 12:36:08 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Window.Into()2/Window.Assign as step s8
May 06, 2022 12:36:08 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/MakeUnionTable0 as step s9
May 06, 2022 12:36:08 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/MakeUnionTable1 as step s10
May 06, 2022 12:36:08 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/Flatten as step s11
May 06, 2022 12:36:08 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/GBK as step s12
May 06, 2022 12:36:08 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/ConstructCoGbkResultFn as step s13
May 06, 2022 12:36:08 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Ungroup and reiterate as step s14
May 06, 2022 12:36:08 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect total bytes as step s15
May 06, 2022 12:36:08 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect end time metrics as step s16
May 06, 2022 12:36:08 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Dataflow SDK version: 2.40.0-SNAPSHOT
May 06, 2022 12:36:09 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To access the Dataflow monitoring console, please navigate to https://console.cloud.google.com/dataflow/jobs/us-central1/2022-05-06_05_36_09-16672906884944557169?project=apache-beam-testing
May 06, 2022 12:36:09 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Submitted job: 2022-05-06_05_36_09-16672906884944557169
May 06, 2022 12:36:09 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To cancel the job using the 'gcloud' tool, run:
> gcloud dataflow jobs --project=apache-beam-testing cancel --region=us-central1 2022-05-06_05_36_09-16672906884944557169
May 06, 2022 12:36:26 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
WARNING: 2022-05-06T12:36:16.319Z: The workflow name is not a valid Cloud Label. Labels applied to Cloud resources (such as GCE Instances) for monitoring will be labeled with this modified job name: load0tests0java170dataflow0v20streaming0cogbk01-jenkins-05-eke9. For the best monitoring experience, please name your job with a valid Cloud Label. For details, see: https://cloud.google.com/compute/docs/labeling-resources#restrictions
May 06, 2022 12:36:26 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-06T12:36:22.945Z: Worker configuration: e2-standard-2 in us-central1-b.
May 06, 2022 12:36:26 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-06T12:36:23.746Z: Expanding SplittableParDo operations into optimizable parts.
May 06, 2022 12:36:26 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-06T12:36:23.775Z: Expanding CollectionToSingleton operations into optimizable parts.
May 06, 2022 12:36:26 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-06T12:36:23.839Z: Expanding CoGroupByKey operations into optimizable parts.
May 06, 2022 12:36:26 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-06T12:36:23.900Z: Expanding SplittableProcessKeyed operations into optimizable parts.
May 06, 2022 12:36:26 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-06T12:36:23.919Z: Expanding GroupByKey operations into streaming Read/Write steps
May 06, 2022 12:36:26 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-06T12:36:23.983Z: Lifting ValueCombiningMappingFns into MergeBucketsMappingFns
May 06, 2022 12:36:26 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-06T12:36:24.087Z: Fusing adjacent ParDo, Read, Write, and Flatten operations
May 06, 2022 12:36:26 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-06T12:36:24.120Z: Unzipping flatten CoGroupByKey-Flatten for input CoGroupByKey-MakeUnionTable0-ParMultiDo-ConstructUnionTable-.output
May 06, 2022 12:36:26 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-06T12:36:24.154Z: Fusing unzipped copy of CoGroupByKey/GBK/WriteStream, through flatten CoGroupByKey/Flatten, into producer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable)
May 06, 2022 12:36:26 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-06T12:36:24.188Z: Fusing consumer CoGroupByKey/GBK/WriteStream into CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable)
May 06, 2022 12:36:26 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-06T12:36:24.220Z: Fusing consumer Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read input/Impulse
May 06, 2022 12:36:26 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-06T12:36:24.254Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
May 06, 2022 12:36:26 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-06T12:36:24.288Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
May 06, 2022 12:36:26 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-06T12:36:24.332Z: Fusing consumer Read input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
May 06, 2022 12:36:26 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-06T12:36:24.352Z: Fusing consumer Collect start time metrics (input)/ParMultiDo(TimeMonitor) into Read input/ParDo(StripIds)/ParMultiDo(StripIds)
May 06, 2022 12:36:26 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-06T12:36:24.379Z: Fusing consumer Window.Into()/Window.Assign into Collect start time metrics (input)/ParMultiDo(TimeMonitor)
May 06, 2022 12:36:26 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-06T12:36:24.412Z: Fusing consumer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable) into Window.Into()/Window.Assign
May 06, 2022 12:36:26 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-06T12:36:24.448Z: Fusing consumer Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read co-input/Impulse
May 06, 2022 12:36:26 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-06T12:36:24.470Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
May 06, 2022 12:36:26 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-06T12:36:24.502Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
May 06, 2022 12:36:26 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-06T12:36:24.528Z: Fusing consumer Read co-input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
May 06, 2022 12:36:26 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-06T12:36:24.559Z: Fusing consumer Collect start time metrics (co-input)/ParMultiDo(TimeMonitor) into Read co-input/ParDo(StripIds)/ParMultiDo(StripIds)
May 06, 2022 12:36:26 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-06T12:36:24.627Z: Fusing consumer Window.Into()2/Window.Assign into Collect start time metrics (co-input)/ParMultiDo(TimeMonitor)
May 06, 2022 12:36:26 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-06T12:36:24.654Z: Fusing consumer CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable) into Window.Into()2/Window.Assign
May 06, 2022 12:36:26 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-06T12:36:24.697Z: Fusing consumer CoGroupByKey/GBK/MergeBuckets into CoGroupByKey/GBK/ReadStream
May 06, 2022 12:36:26 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-06T12:36:24.733Z: Fusing consumer CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult) into CoGroupByKey/GBK/MergeBuckets
May 06, 2022 12:36:26 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-06T12:36:24.755Z: Fusing consumer Ungroup and reiterate/ParMultiDo(UngroupAndReiterate) into CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult)
May 06, 2022 12:36:26 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-06T12:36:24.787Z: Fusing consumer Collect total bytes/ParMultiDo(ByteMonitor) into Ungroup and reiterate/ParMultiDo(UngroupAndReiterate)
May 06, 2022 12:36:26 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-06T12:36:24.820Z: Fusing consumer Collect end time metrics/ParMultiDo(TimeMonitor) into Collect total bytes/ParMultiDo(ByteMonitor)
May 06, 2022 12:36:26 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-06T12:36:24.993Z: Running job using Streaming Engine
May 06, 2022 12:36:26 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-06T12:36:25.292Z: Starting 5 ****s in us-central1-b...
May 06, 2022 12:36:38 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-06T12:36:35.180Z: Your project already contains 100 Dataflow-created metric descriptors, so new user metrics of the form custom.googleapis.com/* will not be created. However, all user metrics are also available in the metric dataflow.googleapis.com/job/user_counter. If you rely on the custom metrics, you can delete old / unused metric descriptors. See https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.list and https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.delete
May 06, 2022 12:36:48 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-06T12:36:47.859Z: Autoscaling: Raised the number of ****s to 5 so that the pipeline can catch up with its backlog and keep up with its input rate.
May 06, 2022 12:37:49 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-06T12:37:47.668Z: Workers have started successfully.
May 06, 2022 4:01:13 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-06T16:01:11.959Z: Cancel request is committed for workflow job: 2022-05-06_05_36_09-16672906884944557169.
May 06, 2022 4:01:13 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-06T16:01:12.063Z: Cleaning up.
May 06, 2022 4:01:13 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-06T16:01:12.173Z: Stopping **** pool...
May 06, 2022 4:01:13 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-06T16:01:12.236Z: Stopping **** pool...
May 06, 2022 4:01:48 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-06T16:01:46.626Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
May 06, 2022 4:01:48 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-06T16:01:46.663Z: Worker pool stopped.
May 06, 2022 4:01:56 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-05-06_05_36_09-16672906884944557169 finished with status CANCELLED.
Load test results for test (ID): aea7067d-9935-4110-bf99-16a4bf6df520 and timestamp: 2022-05-06T12:35:56.758000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                 12160.557
dataflow_v2_java17_total_bytes_count             3.05775743E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220506123132
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:df7a799483a8dfcf0e4964b95e5dc6ef838d5908578aa974f5608b4b2f79498f
Deleted: sha256:c61e2bf56473ec4563f54698aa2a8e107b9212d2f9cb4249e1b8926eebd47660
Deleted: sha256:9d9130c54f1baeb08b94a07964e9e22ecd4bab33862e7798a10319d1d8d4b08d
Deleted: sha256:8467a3fb70145cc93817e21ee5c868eec7532f4e86ca70e4d38f522af0cb398c
Deleted: sha256:59fbca7d507b996b1d51fe22b67984f3ba7e986f1c53f6e3a14639d6277f637d
Deleted: sha256:c5e291eab10e7283e0ca96918031f1a9e83e8b8411f7ac5b1f40ec2f30ba230e
Deleted: sha256:6f601d14f835a5e2a860e8838164b3f7b7d0e20a6c80cab2e040083a47b0b56f
Deleted: sha256:b495ad196c71a0563ffb223307fb959e64e64a20da8e12016bb4602bf19fa9a5
Deleted: sha256:7730c370c352c50337fe7f2bad2ce7e584bfef3be35d36987d4541fbb7c06a2d
Deleted: sha256:6ee922752cb80e19cfd4beb7f1111c74c090f8d970a315f577ca3ab541447ab4
Deleted: sha256:9e190c203486d33f72fe7036f544e381072667a81abf21b65a53927a8ca3ddd9
Deleted: sha256:f2bd20d129b02ee50849e5c9152aac79942dba3415537ea9cba062da73f12839
Deleted: sha256:2909b2457cefac16105342dbdf59f3e9288bd44393da551bd3665665ede39ac8
Deleted: sha256:00968e40a472f2b5ad0dc5d917c967cc8b28c86c62b3207ac0a6e3f943550041
Deleted: sha256:f4d6879504907af6782a76ee389504614fb7ab9455c4b107eb02f2c19046ed49
Deleted: sha256:4be40ff13744c31b7d2eb8e31c8e9e5dd9b9690ef2a3bec37134190d3c85cdb0
Deleted: sha256:36bcd0ee35e63597b3e643aef706aac4fbef299ba2773de693bc6ce03688f272
Deleted: sha256:da888297f316c3a48e5f9787b1fe1a9384df56871c0e907eeb0fd44a91012df7
Deleted: sha256:58d9461f9b85e6dc8dce3d9061ab1b7690a0a8b13fc305f246ef037183b06bf3
Deleted: sha256:9e3aed6009e7e4f7dbf0492e1853b2a2549a3e81cfd12daeaeabfb6094af94bf
Deleted: sha256:75fb1d5e10ded2a00433da41d70fbc05be0aa093ebdeb90611cba0b377d7fec6
Deleted: sha256:e999ffd0c7debe92dbcad4eabf1788922fefe6e251093e945a669a8f458b1c67
Deleted: sha256:9b05c11211514e7340a5e8839f29b37218653537f63497a47cede3c2125bd15b
Deleted: sha256:65e811058f16c4f2bfe8c6b1a1d910ea4dc0532e1a41c80eb57bbc9fded148cb
Deleted: sha256:b66d07df3cdafc664e5592122dc648cfb3250d33c251f4ec18d46206d483bb23
Deleted: sha256:1427b463374c90c14de7918ef0f0c7b26c7fd8a48ca5e4b9e3ef08203ad7c545
Deleted: sha256:d637282d3e86cd34969b59506e5e78e9bb4e3cde032e3c4f9e6e04c08a078941
Deleted: sha256:89a707572366f6809adb4362ae03d1c016718b4c4f45fe40192369f027375415
Deleted: sha256:4d36ad43cbe6758277403104e5f851d113c19fe8dd2e1cbd383416b8ac45d28d
Deleted: sha256:a7ce6b29ef0448179f5fc9bd58e3ff1eb05009c7417e54905a9dd6b2a10f7fff
Deleted: sha256:632d214ecd9402e629ff5d15ce8a4c318d85c083035e0043a09c7d41bdc02d30
Deleted: sha256:6eb47e9cb58a35946ffaab9ad85d0fbe80def98404dee173c9e679c55c36d57e
Deleted: sha256:affbfb0c7a73592504cf91a2cbde5d28f930f33acc93430d08460a0c68304935
Deleted: sha256:239810707ccfa57dd42da91f0a3261bf3902d2c487fa6c35dc7241f5e86b3335
Deleted: sha256:46e146925a73a269ca21a84f848012801e5f0e57e63c23c8dd5206c23efda7ec
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220506123132]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:df7a799483a8dfcf0e4964b95e5dc6ef838d5908578aa974f5608b4b2f79498f]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220506123132] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:df7a799483a8dfcf0e4964b95e5dc6ef838d5908578aa974f5608b4b2f79498f])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:df7a799483a8dfcf0e4964b95e5dc6ef838d5908578aa974f5608b4b2f79498f
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:df7a799483a8dfcf0e4964b95e5dc6ef838d5908578aa974f5608b4b2f79498f
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:df7a799483a8dfcf0e4964b95e5dc6ef838d5908578aa974f5608b4b2f79498f].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.4/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 31m 24s
110 actionable tasks: 73 executed, 33 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/l5ae3vt3byunc

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #136

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/136/display/redirect?page=changes>

Changes:

[Heejong Lee] [BEAM-9245] Unable to pull datatore Entity which contains dict

[bulat.safiullin] [BEAM-14382] [Website] add banner container for with css, images, html

[Jan Lukavský] [BEAM-14196] add test verifying output watermark propagation in bundle

[Jan Lukavský] [BEAM-14196] Fix FlinkRunner mid-bundle output watermark handling

[nielm] [BEAM-14405] Fix NPE when ProjectID is not specified in a template

[bulat.safiullin] [BEAM-14382] change mobile banner img, add padding to banner section

[ahmedabualsaud] fix test decotrator typo

[noreply] Merge pull request #17440 from [BEAM-14329] Enable exponential backoff

[noreply] [BEAM-11104] Fix output forwarding issue for ProcessContinuations

[noreply] re-add testing package to pydoc (#17524)

[Heejong Lee] add test

[noreply] [BEAM-14250] Amended the workaround (#17531)

[noreply] [BEAM-11104] Fix broken split result validation (#17546)

[noreply] Fixed a SQL and screenshots in the Beam SQL blog (#17545)

[noreply] Merge pull request #17417: [BEAM-14388] Address some performance

[noreply] [BEAM-14386] [Flink] Support for scala 2.12 (#17512)

[noreply] [BEAM-14294] Worker changes to support trivial Batched DoFns (#17384)

[zyichi] Moving to 2.40.0-SNAPSHOT on master branch.

[noreply] [BEAM-14048] [CdapIO] Add ConfigWrapper for building CDAP PluginConfigs


------------------------------------------
[...truncated 50.60 KB...]
a037458de4e0: Preparing
a0aeee58f513: Waiting
d7b56d57c324: Waiting
04d5df9ec8e8: Waiting
dc197617b2ee: Waiting
bafdbe68e4ae: Preparing
c85968a0c042: Waiting
3263c0e06234: Waiting
3bc383470c05: Waiting
673012901f20: Waiting
e93827457889: Waiting
a13c519c6361: Preparing
bafdbe68e4ae: Waiting
a13c519c6361: Waiting
a5830f62fc2b: Waiting
a037458de4e0: Waiting
8996c988e5df: Pushed
882944b20e14: Pushed
c72f16974de9: Pushed
0d42ebd3c794: Pushed
911b1a8b5e57: Pushed
a0aeee58f513: Pushed
a81a04dbaea7: Pushed
d7b56d57c324: Pushed
1d4e401c8952: Pushed
04d5df9ec8e8: Pushed
a5830f62fc2b: Pushed
3bc383470c05: Layer already exists
e93827457889: Layer already exists
c85968a0c042: Pushed
08fa02ce37eb: Layer already exists
a037458de4e0: Layer already exists
bafdbe68e4ae: Layer already exists
a13c519c6361: Layer already exists
3263c0e06234: Pushed
dc197617b2ee: Pushed
673012901f20: Pushed
20220505123041: digest: sha256:131e4451715f7df49079705cc0d0df461d263e958b7b9a5e9d379950b22ba9a0 size: 4729

> Task :sdks:java:testing:load-tests:run
May 05, 2022 12:32:46 PM org.apache.beam.runners.dataflow.options.DataflowPipelineOptions$StagingLocationFactory create
INFO: No stagingLocation provided, falling back to gcpTempLocation
May 05, 2022 12:32:46 PM org.apache.beam.runners.dataflow.DataflowRunner fromOptions
INFO: PipelineOptions.filesToStage was not specified. Defaulting to files from the classpath: will stage 221 files. Enable logging at DEBUG level to see which files will be staged.
May 05, 2022 12:32:47 PM org.apache.beam.sdk.Pipeline validate
WARNING: The following transforms do not have stable unique names: Window.Into()
May 05, 2022 12:32:47 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Executing pipeline on the Dataflow Service, which will have billing implications related to Google Compute Engine usage and other Google Cloud Services.
May 05, 2022 12:32:49 PM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Uploading 221 files from PipelineOptions.filesToStage to staging location to prepare for execution.
May 05, 2022 12:32:49 PM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Staging files complete: 221 files cached, 0 files newly uploaded in 0 seconds
May 05, 2022 12:32:49 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Staging portable pipeline proto to gs://temp-storage-for-perf-tests/loadtests/staging/
May 05, 2022 12:32:50 PM org.apache.beam.runners.dataflow.util.PackageUtil tryStagePackage
INFO: Uploading <119943 bytes, hash d2875933a8f7081a206ea4f4f77ec0d39d59892254cb85fcb163b4db2e98fae1> to gs://temp-storage-for-perf-tests/loadtests/staging/pipeline-0odZM6j3CBogbqT0937A051ZiSJUy4X8sWO02y6Y-uE.pb
May 05, 2022 12:32:51 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read input/DataflowRunner.StreamingUnboundedRead.ReadWithIds as step s1
May 05, 2022 12:32:51 PM org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource split
INFO: Split into 20 bundles of sizes: [org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@463561c5, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@659feb22, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3468ee6e, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2f4b98f6, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@421def93, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@58c1da09, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2b2954e1, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@58d6e55a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@751ae8a4, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@235d659c, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4232b34a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2da16263, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@f5ce0bb, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@47e51549, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@101a461c, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@360e9c06, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5ebffb44, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@311ff287, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7377781e, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@31db34da]
May 05, 2022 12:32:51 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read input/StripIds as step s2
May 05, 2022 12:32:51 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect start time metrics (input) as step s3
May 05, 2022 12:32:51 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Window.Into()/Window.Assign as step s4
May 05, 2022 12:32:51 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read co-input/DataflowRunner.StreamingUnboundedRead.ReadWithIds as step s5
May 05, 2022 12:32:51 PM org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource split
INFO: Split into 20 bundles of sizes: [org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3a2e9f5b, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7da34b26, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@10c47c79, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3b2553d9, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@753fd7a1, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@12a2585b, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@21b6c9c2, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@309cedb6, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3b95a6db, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3bf54172, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2c9a6717, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7b3cde6f, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6d091cad, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7c663eaf, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3bb5ceb, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1e692555, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3ba0ae41, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@76fe6cdc, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2ffb3aec, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@786ff1cb]
May 05, 2022 12:32:51 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read co-input/StripIds as step s6
May 05, 2022 12:32:51 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect start time metrics (co-input) as step s7
May 05, 2022 12:32:51 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Window.Into()2/Window.Assign as step s8
May 05, 2022 12:32:51 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/MakeUnionTable0 as step s9
May 05, 2022 12:32:51 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/MakeUnionTable1 as step s10
May 05, 2022 12:32:51 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/Flatten as step s11
May 05, 2022 12:32:51 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/GBK as step s12
May 05, 2022 12:32:51 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/ConstructCoGbkResultFn as step s13
May 05, 2022 12:32:51 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Ungroup and reiterate as step s14
May 05, 2022 12:32:51 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect total bytes as step s15
May 05, 2022 12:32:51 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect end time metrics as step s16
May 05, 2022 12:32:51 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Dataflow SDK version: 2.40.0-SNAPSHOT
May 05, 2022 12:32:52 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To access the Dataflow monitoring console, please navigate to https://console.cloud.google.com/dataflow/jobs/us-central1/2022-05-05_05_32_52-6914658191205046925?project=apache-beam-testing
May 05, 2022 12:32:52 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Submitted job: 2022-05-05_05_32_52-6914658191205046925
May 05, 2022 12:32:52 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To cancel the job using the 'gcloud' tool, run:
> gcloud dataflow jobs --project=apache-beam-testing cancel --region=us-central1 2022-05-05_05_32_52-6914658191205046925
May 05, 2022 12:33:07 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
WARNING: 2022-05-05T12:33:04.755Z: The workflow name is not a valid Cloud Label. Labels applied to Cloud resources (such as GCE Instances) for monitoring will be labeled with this modified job name: load0tests0java170dataflow0v20streaming0cogbk01-jenkins-05-839r. For the best monitoring experience, please name your job with a valid Cloud Label. For details, see: https://cloud.google.com/compute/docs/labeling-resources#restrictions
May 05, 2022 12:33:13 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-05T12:33:12.969Z: Worker configuration: e2-standard-2 in us-central1-b.
May 05, 2022 12:33:16 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-05T12:33:13.766Z: Expanding SplittableParDo operations into optimizable parts.
May 05, 2022 12:33:16 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-05T12:33:13.798Z: Expanding CollectionToSingleton operations into optimizable parts.
May 05, 2022 12:33:16 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-05T12:33:13.898Z: Expanding CoGroupByKey operations into optimizable parts.
May 05, 2022 12:33:16 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-05T12:33:13.959Z: Expanding SplittableProcessKeyed operations into optimizable parts.
May 05, 2022 12:33:16 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-05T12:33:13.987Z: Expanding GroupByKey operations into streaming Read/Write steps
May 05, 2022 12:33:16 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-05T12:33:14.068Z: Lifting ValueCombiningMappingFns into MergeBucketsMappingFns
May 05, 2022 12:33:16 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-05T12:33:14.146Z: Fusing adjacent ParDo, Read, Write, and Flatten operations
May 05, 2022 12:33:16 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-05T12:33:14.178Z: Unzipping flatten CoGroupByKey-Flatten for input CoGroupByKey-MakeUnionTable0-ParMultiDo-ConstructUnionTable-.output
May 05, 2022 12:33:16 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-05T12:33:14.208Z: Fusing unzipped copy of CoGroupByKey/GBK/WriteStream, through flatten CoGroupByKey/Flatten, into producer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable)
May 05, 2022 12:33:16 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-05T12:33:14.232Z: Fusing consumer CoGroupByKey/GBK/WriteStream into CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable)
May 05, 2022 12:33:16 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-05T12:33:14.252Z: Fusing consumer Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read input/Impulse
May 05, 2022 12:33:16 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-05T12:33:14.281Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
May 05, 2022 12:33:16 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-05T12:33:14.316Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
May 05, 2022 12:33:16 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-05T12:33:14.346Z: Fusing consumer Read input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
May 05, 2022 12:33:16 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-05T12:33:14.380Z: Fusing consumer Collect start time metrics (input)/ParMultiDo(TimeMonitor) into Read input/ParDo(StripIds)/ParMultiDo(StripIds)
May 05, 2022 12:33:16 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-05T12:33:14.404Z: Fusing consumer Window.Into()/Window.Assign into Collect start time metrics (input)/ParMultiDo(TimeMonitor)
May 05, 2022 12:33:16 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-05T12:33:14.461Z: Fusing consumer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable) into Window.Into()/Window.Assign
May 05, 2022 12:33:16 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-05T12:33:14.500Z: Fusing consumer Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read co-input/Impulse
May 05, 2022 12:33:16 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-05T12:33:14.531Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
May 05, 2022 12:33:16 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-05T12:33:14.563Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
May 05, 2022 12:33:16 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-05T12:33:14.596Z: Fusing consumer Read co-input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
May 05, 2022 12:33:16 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-05T12:33:14.630Z: Fusing consumer Collect start time metrics (co-input)/ParMultiDo(TimeMonitor) into Read co-input/ParDo(StripIds)/ParMultiDo(StripIds)
May 05, 2022 12:33:16 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-05T12:33:14.665Z: Fusing consumer Window.Into()2/Window.Assign into Collect start time metrics (co-input)/ParMultiDo(TimeMonitor)
May 05, 2022 12:33:16 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-05T12:33:14.697Z: Fusing consumer CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable) into Window.Into()2/Window.Assign
May 05, 2022 12:33:16 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-05T12:33:14.718Z: Fusing consumer CoGroupByKey/GBK/MergeBuckets into CoGroupByKey/GBK/ReadStream
May 05, 2022 12:33:16 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-05T12:33:14.744Z: Fusing consumer CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult) into CoGroupByKey/GBK/MergeBuckets
May 05, 2022 12:33:16 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-05T12:33:14.781Z: Fusing consumer Ungroup and reiterate/ParMultiDo(UngroupAndReiterate) into CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult)
May 05, 2022 12:33:16 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-05T12:33:14.848Z: Fusing consumer Collect total bytes/ParMultiDo(ByteMonitor) into Ungroup and reiterate/ParMultiDo(UngroupAndReiterate)
May 05, 2022 12:33:16 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-05T12:33:14.877Z: Fusing consumer Collect end time metrics/ParMultiDo(TimeMonitor) into Collect total bytes/ParMultiDo(ByteMonitor)
May 05, 2022 12:33:16 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-05T12:33:15.025Z: Running job using Streaming Engine
May 05, 2022 12:33:16 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-05T12:33:15.265Z: Starting 5 ****s in us-central1-b...
May 05, 2022 12:33:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-05T12:33:41.370Z: Autoscaling: Raised the number of ****s to 4 so that the pipeline can catch up with its backlog and keep up with its input rate.
May 05, 2022 12:33:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-05T12:33:41.398Z: Resized **** pool to 4, though goal was 5.  This could be a quota issue.
May 05, 2022 12:33:48 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-05T12:33:46.902Z: Your project already contains 100 Dataflow-created metric descriptors, so new user metrics of the form custom.googleapis.com/* will not be created. However, all user metrics are also available in the metric dataflow.googleapis.com/job/user_counter. If you rely on the custom metrics, you can delete old / unused metric descriptors. See https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.list and https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.delete
May 05, 2022 12:33:52 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-05T12:33:51.587Z: Autoscaling: Raised the number of ****s to 5 so that the pipeline can catch up with its backlog and keep up with its input rate.
May 05, 2022 12:34:51 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-05T12:34:50.485Z: Workers have started successfully.
May 05, 2022 3:05:20 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-05T15:05:19.624Z: Your project already contains 100 Dataflow-created metric descriptors, so new user metrics of the form custom.googleapis.com/* will not be created. However, all user metrics are also available in the metric dataflow.googleapis.com/job/user_counter. If you rely on the custom metrics, you can delete old / unused metric descriptors. See https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.list and https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.delete
May 05, 2022 3:05:25 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-05T15:05:24.302Z: Worker configuration: e2-standard-2 in us-central1-b.
May 05, 2022 4:00:50 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-05T16:00:49.572Z: Cancel request is committed for workflow job: 2022-05-05_05_32_52-6914658191205046925.
May 05, 2022 4:00:50 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-05T16:00:49.600Z: Cleaning up.
May 05, 2022 4:00:50 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-05T16:00:49.808Z: Stopping **** pool...
May 05, 2022 4:00:50 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-05T16:00:49.897Z: Stopping **** pool...
May 05, 2022 4:01:28 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-05T16:01:26.536Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
May 05, 2022 4:01:28 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-05T16:01:26.569Z: Worker pool stopped.
May 05, 2022 4:01:39 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-05-05_05_32_52-6914658191205046925 finished with status CANCELLED.
Load test results for test (ID): 3e4bb4f5-730b-4386-9ddb-8239732ef044 and timestamp: 2022-05-05T12:32:46.775000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                 12316.972
dataflow_v2_java17_total_bytes_count             3.41388437E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220505123041
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:131e4451715f7df49079705cc0d0df461d263e958b7b9a5e9d379950b22ba9a0
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220505123041]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:131e4451715f7df49079705cc0d0df461d263e958b7b9a5e9d379950b22ba9a0]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220505123041] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:131e4451715f7df49079705cc0d0df461d263e958b7b9a5e9d379950b22ba9a0])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:131e4451715f7df49079705cc0d0df461d263e958b7b9a5e9d379950b22ba9a0
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:131e4451715f7df49079705cc0d0df461d263e958b7b9a5e9d379950b22ba9a0
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:131e4451715f7df49079705cc0d0df461d263e958b7b9a5e9d379950b22ba9a0].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.4/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 31m 20s
110 actionable tasks: 76 executed, 30 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/vqaloxpdhzprm

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #135

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/135/display/redirect?page=changes>

Changes:

[noreply] fix: JDBC config schema fields order

[Brian Hulette] Revert "Merge pull request #17255 from kileys/test-revert"

[Brian Hulette] BEAM-14231: bypass schema cache for

[noreply] [BEAM-13657] Follow up update version warning in __init__ (#17493)

[noreply] Merge pull request #17431 from [BEAM-14273] Add integration tests for BQ

[noreply] Merge pull request #17205 from [BEAM-14145] [Website] add carousel to

[noreply] [BEAM-14064] fix es io windowing (#17112)

[noreply] [BEAM-13670] Upgraded ipython from v7 to v8 (#17529)

[noreply] [BEAM-11104] Enable ProcessContinuation return values, add unit test

[Robert Bradshaw] [BEAM-14403] Allow Prime to be used with legacy workers.

[noreply] [BEAM-11106] Support drain in Go SDK (#17432)

[noreply] add __Init__ to inference. (#17514)


------------------------------------------
[...truncated 50.32 KB...]
3bc383470c05: Preparing
e93827457889: Preparing
08fa02ce37eb: Preparing
a037458de4e0: Preparing
bafdbe68e4ae: Preparing
a13c519c6361: Preparing
3bc70159237a: Waiting
a6e4d1804105: Waiting
2d543f6e6e21: Waiting
0307cef3be55: Waiting
a037458de4e0: Waiting
08fa02ce37eb: Waiting
8b92e4958fd1: Waiting
bafdbe68e4ae: Waiting
9969bcd1e3fd: Waiting
e93827457889: Waiting
3bc383470c05: Waiting
2520eb2e8567: Waiting
72bfc7510f27: Waiting
a4ad5e0bd783: Waiting
f116eab367a0: Pushed
331c9a0b820a: Pushed
cb7330ae032c: Pushed
25a6908e0e13: Pushed
b7bbab1ffc45: Pushed
a4ad5e0bd783: Pushed
3bc70159237a: Pushed
2520eb2e8567: Pushed
9745019c1f1b: Pushed
2d543f6e6e21: Pushed
72bfc7510f27: Pushed
0307cef3be55: Pushed
3bc383470c05: Layer already exists
e93827457889: Layer already exists
08fa02ce37eb: Layer already exists
a037458de4e0: Layer already exists
bafdbe68e4ae: Layer already exists
a13c519c6361: Layer already exists
8b92e4958fd1: Pushed
9969bcd1e3fd: Pushed
a6e4d1804105: Pushed
20220504123041: digest: sha256:b607c702456435e49fbe29707a5e8a335df87c310c8eae2069a24e6102daeb1f size: 4729

> Task :sdks:java:testing:load-tests:run
May 04, 2022 12:32:36 PM org.apache.beam.runners.dataflow.options.DataflowPipelineOptions$StagingLocationFactory create
INFO: No stagingLocation provided, falling back to gcpTempLocation
May 04, 2022 12:32:37 PM org.apache.beam.runners.dataflow.DataflowRunner fromOptions
INFO: PipelineOptions.filesToStage was not specified. Defaulting to files from the classpath: will stage 221 files. Enable logging at DEBUG level to see which files will be staged.
May 04, 2022 12:32:38 PM org.apache.beam.sdk.Pipeline validate
WARNING: The following transforms do not have stable unique names: Window.Into()
May 04, 2022 12:32:38 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Executing pipeline on the Dataflow Service, which will have billing implications related to Google Compute Engine usage and other Google Cloud Services.
May 04, 2022 12:32:40 PM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Uploading 221 files from PipelineOptions.filesToStage to staging location to prepare for execution.
May 04, 2022 12:32:41 PM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Staging files complete: 221 files cached, 0 files newly uploaded in 0 seconds
May 04, 2022 12:32:41 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Staging portable pipeline proto to gs://temp-storage-for-perf-tests/loadtests/staging/
May 04, 2022 12:32:41 PM org.apache.beam.runners.dataflow.util.PackageUtil tryStagePackage
INFO: Uploading <119943 bytes, hash a19c96ecdc20c9355bed3a3f2f46e859fa201abbc15dd86e5cf410cc6d045525> to gs://temp-storage-for-perf-tests/loadtests/staging/pipeline-oZyW7NwgyTVb7To_L0boWfogGrvBXdhuXPQQzG0EVSU.pb
May 04, 2022 12:32:43 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read input/DataflowRunner.StreamingUnboundedRead.ReadWithIds as step s1
May 04, 2022 12:32:43 PM org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource split
INFO: Split into 20 bundles of sizes: [org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@659feb22, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3468ee6e, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2f4b98f6, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@421def93, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@58c1da09, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2b2954e1, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@58d6e55a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@751ae8a4, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@235d659c, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4232b34a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2da16263, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@f5ce0bb, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@47e51549, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@101a461c, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@360e9c06, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5ebffb44, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@311ff287, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7377781e, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@31db34da, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@109f8c7e]
May 04, 2022 12:32:43 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read input/StripIds as step s2
May 04, 2022 12:32:43 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect start time metrics (input) as step s3
May 04, 2022 12:32:43 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Window.Into()/Window.Assign as step s4
May 04, 2022 12:32:43 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read co-input/DataflowRunner.StreamingUnboundedRead.ReadWithIds as step s5
May 04, 2022 12:32:43 PM org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource split
INFO: Split into 20 bundles of sizes: [org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3bf54172, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2c9a6717, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7b3cde6f, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6d091cad, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7c663eaf, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3bb5ceb, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1e692555, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3ba0ae41, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@76fe6cdc, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2ffb3aec, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@786ff1cb, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@46039a21, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@431e86b1, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@35c4e864, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@32a2a6be, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@682af059, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5f36c8e3, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4da39ca9, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6a9344f5, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5584d9c6]
May 04, 2022 12:32:43 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read co-input/StripIds as step s6
May 04, 2022 12:32:43 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect start time metrics (co-input) as step s7
May 04, 2022 12:32:43 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Window.Into()2/Window.Assign as step s8
May 04, 2022 12:32:43 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/MakeUnionTable0 as step s9
May 04, 2022 12:32:43 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/MakeUnionTable1 as step s10
May 04, 2022 12:32:43 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/Flatten as step s11
May 04, 2022 12:32:43 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/GBK as step s12
May 04, 2022 12:32:43 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/ConstructCoGbkResultFn as step s13
May 04, 2022 12:32:43 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Ungroup and reiterate as step s14
May 04, 2022 12:32:43 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect total bytes as step s15
May 04, 2022 12:32:43 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect end time metrics as step s16
May 04, 2022 12:32:43 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Dataflow SDK version: 2.39.0-SNAPSHOT
May 04, 2022 12:32:44 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To access the Dataflow monitoring console, please navigate to https://console.cloud.google.com/dataflow/jobs/us-central1/2022-05-04_05_32_43-7390128921444977935?project=apache-beam-testing
May 04, 2022 12:32:44 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Submitted job: 2022-05-04_05_32_43-7390128921444977935
May 04, 2022 12:32:44 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To cancel the job using the 'gcloud' tool, run:
> gcloud dataflow jobs --project=apache-beam-testing cancel --region=us-central1 2022-05-04_05_32_43-7390128921444977935
May 04, 2022 12:32:48 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
WARNING: 2022-05-04T12:32:47.612Z: The workflow name is not a valid Cloud Label. Labels applied to Cloud resources (such as GCE Instances) for monitoring will be labeled with this modified job name: load0tests0java170dataflow0v20streaming0cogbk01-jenkins-05-q8mb. For the best monitoring experience, please name your job with a valid Cloud Label. For details, see: https://cloud.google.com/compute/docs/labeling-resources#restrictions
May 04, 2022 12:32:53 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-04T12:32:52.883Z: Worker configuration: e2-standard-2 in us-central1-b.
May 04, 2022 12:32:53 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-04T12:32:53.521Z: Expanding SplittableParDo operations into optimizable parts.
May 04, 2022 12:32:53 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-04T12:32:53.565Z: Expanding CollectionToSingleton operations into optimizable parts.
May 04, 2022 12:32:53 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-04T12:32:53.692Z: Expanding CoGroupByKey operations into optimizable parts.
May 04, 2022 12:32:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-04T12:32:53.790Z: Expanding SplittableProcessKeyed operations into optimizable parts.
May 04, 2022 12:32:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-04T12:32:53.843Z: Expanding GroupByKey operations into streaming Read/Write steps
May 04, 2022 12:32:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-04T12:32:53.962Z: Lifting ValueCombiningMappingFns into MergeBucketsMappingFns
May 04, 2022 12:32:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-04T12:32:54.105Z: Fusing adjacent ParDo, Read, Write, and Flatten operations
May 04, 2022 12:32:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-04T12:32:54.144Z: Unzipping flatten CoGroupByKey-Flatten for input CoGroupByKey-MakeUnionTable0-ParMultiDo-ConstructUnionTable-.output
May 04, 2022 12:32:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-04T12:32:54.191Z: Fusing unzipped copy of CoGroupByKey/GBK/WriteStream, through flatten CoGroupByKey/Flatten, into producer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable)
May 04, 2022 12:32:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-04T12:32:54.237Z: Fusing consumer CoGroupByKey/GBK/WriteStream into CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable)
May 04, 2022 12:32:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-04T12:32:54.311Z: Fusing consumer Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read input/Impulse
May 04, 2022 12:32:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-04T12:32:54.350Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
May 04, 2022 12:32:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-04T12:32:54.403Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
May 04, 2022 12:32:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-04T12:32:54.441Z: Fusing consumer Read input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
May 04, 2022 12:32:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-04T12:32:54.483Z: Fusing consumer Collect start time metrics (input)/ParMultiDo(TimeMonitor) into Read input/ParDo(StripIds)/ParMultiDo(StripIds)
May 04, 2022 12:32:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-04T12:32:54.527Z: Fusing consumer Window.Into()/Window.Assign into Collect start time metrics (input)/ParMultiDo(TimeMonitor)
May 04, 2022 12:32:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-04T12:32:54.581Z: Fusing consumer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable) into Window.Into()/Window.Assign
May 04, 2022 12:32:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-04T12:32:54.620Z: Fusing consumer Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read co-input/Impulse
May 04, 2022 12:32:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-04T12:32:54.659Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
May 04, 2022 12:32:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-04T12:32:54.705Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
May 04, 2022 12:32:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-04T12:32:54.742Z: Fusing consumer Read co-input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
May 04, 2022 12:32:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-04T12:32:54.779Z: Fusing consumer Collect start time metrics (co-input)/ParMultiDo(TimeMonitor) into Read co-input/ParDo(StripIds)/ParMultiDo(StripIds)
May 04, 2022 12:32:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-04T12:32:54.819Z: Fusing consumer Window.Into()2/Window.Assign into Collect start time metrics (co-input)/ParMultiDo(TimeMonitor)
May 04, 2022 12:32:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-04T12:32:54.858Z: Fusing consumer CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable) into Window.Into()2/Window.Assign
May 04, 2022 12:32:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-04T12:32:54.925Z: Fusing consumer CoGroupByKey/GBK/MergeBuckets into CoGroupByKey/GBK/ReadStream
May 04, 2022 12:32:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-04T12:32:54.966Z: Fusing consumer CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult) into CoGroupByKey/GBK/MergeBuckets
May 04, 2022 12:32:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-04T12:32:55.004Z: Fusing consumer Ungroup and reiterate/ParMultiDo(UngroupAndReiterate) into CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult)
May 04, 2022 12:32:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-04T12:32:55.043Z: Fusing consumer Collect total bytes/ParMultiDo(ByteMonitor) into Ungroup and reiterate/ParMultiDo(UngroupAndReiterate)
May 04, 2022 12:32:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-04T12:32:55.123Z: Fusing consumer Collect end time metrics/ParMultiDo(TimeMonitor) into Collect total bytes/ParMultiDo(ByteMonitor)
May 04, 2022 12:32:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-04T12:32:55.382Z: Running job using Streaming Engine
May 04, 2022 12:32:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-04T12:32:55.912Z: Starting 5 ****s in us-central1-b...
May 04, 2022 12:33:19 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-04T12:33:17.504Z: Your project already contains 100 Dataflow-created metric descriptors, so new user metrics of the form custom.googleapis.com/* will not be created. However, all user metrics are also available in the metric dataflow.googleapis.com/job/user_counter. If you rely on the custom metrics, you can delete old / unused metric descriptors. See https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.list and https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.delete
May 04, 2022 12:33:22 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-04T12:33:20.346Z: Autoscaling: Raised the number of ****s to 4 so that the pipeline can catch up with its backlog and keep up with its input rate.
May 04, 2022 12:33:22 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-04T12:33:20.377Z: Resized **** pool to 4, though goal was 5.  This could be a quota issue.
May 04, 2022 12:33:31 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-04T12:33:30.625Z: Autoscaling: Raised the number of ****s to 5 so that the pipeline can catch up with its backlog and keep up with its input rate.
May 04, 2022 12:34:23 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-04T12:34:22.731Z: Workers have started successfully.
May 04, 2022 4:01:00 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-04T16:00:59.996Z: Cancel request is committed for workflow job: 2022-05-04_05_32_43-7390128921444977935.
May 04, 2022 4:01:02 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-04T16:01:00.111Z: Cleaning up.
May 04, 2022 4:01:02 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-04T16:01:00.314Z: Stopping **** pool...
May 04, 2022 4:01:02 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-04T16:01:00.383Z: Stopping **** pool...
May 04, 2022 4:01:37 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-04T16:01:35.915Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
May 04, 2022 4:01:37 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-04T16:01:35.987Z: Worker pool stopped.
May 04, 2022 4:01:44 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-05-04_05_32_43-7390128921444977935 finished with status CANCELLED.
Load test results for test (ID): 1fb5cbe1-adcf-4c95-8641-b4766f1d23bb and timestamp: 2022-05-04T12:32:37.757000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                 12363.386
dataflow_v2_java17_total_bytes_count             4.12943456E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220504123041
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:b607c702456435e49fbe29707a5e8a335df87c310c8eae2069a24e6102daeb1f
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220504123041]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:b607c702456435e49fbe29707a5e8a335df87c310c8eae2069a24e6102daeb1f]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220504123041] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:b607c702456435e49fbe29707a5e8a335df87c310c8eae2069a24e6102daeb1f])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:b607c702456435e49fbe29707a5e8a335df87c310c8eae2069a24e6102daeb1f
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:b607c702456435e49fbe29707a5e8a335df87c310c8eae2069a24e6102daeb1f
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:b607c702456435e49fbe29707a5e8a335df87c310c8eae2069a24e6102daeb1f].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.4/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 31m 26s
110 actionable tasks: 73 executed, 33 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/d7u3qhbfqkica

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #134

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/134/display/redirect?page=changes>

Changes:

[yathu] [BEAM-14375] Fix Java Wordcount Dataflow postcommit

[Robert Bradshaw] Allow arithmetic between deferred scalars.

[noreply] [BEAM-14390] Set user-agent when pulling licenses to avoid 403s (#17521)

[noreply] [BEAM-8688] Upgrade GCSIO to 2.2.6 (#17486)

[noreply] [BEAM-14253] patch SubscriptionPartitionLoader to work around a dataflow

[noreply] Add website link log to notify user of pre-build workflow. (#17498)

[noreply] [BEAM-11105] Add timestamp observing watermark estimation (#17476)

[noreply] Merge pull request #17487 from Adding user-agent to GCS client in Python

[noreply] [BEAM-10265] Display error message if trying to infer recursive schema

[noreply] [BEAM-12575] Upgraded ipykernel from v5 to v6 (#17526)

[noreply] [BEAM-11105] Add docs + CHANGES.md entry for Go Watermark Estimation

[noreply] Merge pull request #17380 from [BEAM-14314][BEAM-9532] Add last_updated


------------------------------------------
[...truncated 45.93 KB...]
> Task :release:go-licenses:java:dockerRun
+ go-licenses save github.com/apache/beam/sdks/java/container --save_path=/output/licenses
+ go-licenses csv github.com/apache/beam/sdks/java/container
+ tee /output/licenses/list.csv
+ chmod -R a+w /output/licenses

> Task :release:go-licenses:java:createLicenses
> Task :sdks:java:container:java17:copyGolangLicenses
> Task :sdks:java:container:java17:dockerPrepare
> Task :sdks:java:container:java17:docker

> Task :runners:google-cloud-dataflow-java:buildAndPushDockerJavaContainer
WARNING: `gcloud docker` will not be supported for Docker client versions above 18.03.

As an alternative, use `gcloud auth configure-docker` to configure `docker` to
use `gcloud` as a credential helper, then use `docker` as you would for non-GCR
registries, e.g. `docker pull gcr.io/project-id/my-image`. Add
`--verbosity=error` to silence this warning: `gcloud docker
--verbosity=error -- pull gcr.io/project-id/my-image`.

See: https://cloud.google.com/container-registry/docs/support/deprecation-notices#gcloud-docker

The push refers to repository [us.gcr.io/apache-beam-testing/java-postcommit-it/java]
52eb74fe55c7: Preparing
a0047307b411: Preparing
22f5c719ac2a: Preparing
571a60fbcd54: Preparing
5cafb396ea19: Preparing
b9bb22d55e3e: Preparing
cf1017aad468: Preparing
d825907d653d: Preparing
8db3937692b0: Preparing
bac86cd751e4: Preparing
9fd422015e17: Preparing
5e79d87fff88: Preparing
51a379cd90f3: Preparing
ea4b0956d47c: Preparing
75795add637a: Preparing
3bc383470c05: Preparing
e93827457889: Preparing
08fa02ce37eb: Preparing
a037458de4e0: Preparing
bafdbe68e4ae: Preparing
a13c519c6361: Preparing
ea4b0956d47c: Waiting
75795add637a: Waiting
8db3937692b0: Waiting
a037458de4e0: Waiting
b9bb22d55e3e: Waiting
bac86cd751e4: Waiting
bafdbe68e4ae: Waiting
3bc383470c05: Waiting
a13c519c6361: Waiting
9fd422015e17: Waiting
51a379cd90f3: Waiting
e93827457889: Waiting
08fa02ce37eb: Waiting
d825907d653d: Waiting
cf1017aad468: Waiting
5cafb396ea19: Pushed
22f5c719ac2a: Pushed
a0047307b411: Pushed
571a60fbcd54: Pushed
52eb74fe55c7: Pushed
cf1017aad468: Pushed
d825907d653d: Pushed
bac86cd751e4: Pushed
8db3937692b0: Pushed
5e79d87fff88: Pushed
b9bb22d55e3e: Pushed
3bc383470c05: Layer already exists
e93827457889: Layer already exists
9fd422015e17: Pushed
08fa02ce37eb: Layer already exists
a037458de4e0: Layer already exists
bafdbe68e4ae: Layer already exists
a13c519c6361: Layer already exists
ea4b0956d47c: Pushed
75795add637a: Pushed
51a379cd90f3: Pushed
20220503123036: digest: sha256:d3e45bd5042d6ca189215f85538763fdb51a453a71b9f9ce4dd097d09e46a5dd size: 4729

> Task :sdks:java:testing:load-tests:run
May 03, 2022 12:32:36 PM org.apache.beam.runners.dataflow.options.DataflowPipelineOptions$StagingLocationFactory create
INFO: No stagingLocation provided, falling back to gcpTempLocation
May 03, 2022 12:32:36 PM org.apache.beam.runners.dataflow.DataflowRunner fromOptions
INFO: PipelineOptions.filesToStage was not specified. Defaulting to files from the classpath: will stage 221 files. Enable logging at DEBUG level to see which files will be staged.
May 03, 2022 12:32:37 PM org.apache.beam.sdk.Pipeline validate
WARNING: The following transforms do not have stable unique names: Window.Into()
May 03, 2022 12:32:37 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Executing pipeline on the Dataflow Service, which will have billing implications related to Google Compute Engine usage and other Google Cloud Services.
May 03, 2022 12:32:41 PM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Uploading 221 files from PipelineOptions.filesToStage to staging location to prepare for execution.
May 03, 2022 12:32:43 PM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Staging files complete: 221 files cached, 0 files newly uploaded in 2 seconds
May 03, 2022 12:32:43 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Staging portable pipeline proto to gs://temp-storage-for-perf-tests/loadtests/staging/
May 03, 2022 12:32:44 PM org.apache.beam.runners.dataflow.util.PackageUtil tryStagePackage
INFO: Uploading <119943 bytes, hash 59f916c5e19489bd88ce5811761090b62724c332f6f3411ef2c783946ee7ef48> to gs://temp-storage-for-perf-tests/loadtests/staging/pipeline-WfkWxeGUib2IzlgRdhCQtickwzL280Ee8seDlG7n70g.pb
May 03, 2022 12:32:45 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read input/DataflowRunner.StreamingUnboundedRead.ReadWithIds as step s1
May 03, 2022 12:32:45 PM org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource split
INFO: Split into 20 bundles of sizes: [org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@463561c5, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@659feb22, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3468ee6e, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2f4b98f6, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@421def93, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@58c1da09, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2b2954e1, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@58d6e55a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@751ae8a4, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@235d659c, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4232b34a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2da16263, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@f5ce0bb, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@47e51549, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@101a461c, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@360e9c06, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5ebffb44, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@311ff287, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7377781e, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@31db34da]
May 03, 2022 12:32:45 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read input/StripIds as step s2
May 03, 2022 12:32:45 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect start time metrics (input) as step s3
May 03, 2022 12:32:45 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Window.Into()/Window.Assign as step s4
May 03, 2022 12:32:45 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read co-input/DataflowRunner.StreamingUnboundedRead.ReadWithIds as step s5
May 03, 2022 12:32:45 PM org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource split
INFO: Split into 20 bundles of sizes: [org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3a2e9f5b, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7da34b26, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@10c47c79, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3b2553d9, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@753fd7a1, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@12a2585b, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@21b6c9c2, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@309cedb6, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3b95a6db, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3bf54172, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2c9a6717, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7b3cde6f, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6d091cad, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7c663eaf, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3bb5ceb, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1e692555, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3ba0ae41, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@76fe6cdc, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2ffb3aec, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@786ff1cb]
May 03, 2022 12:32:45 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read co-input/StripIds as step s6
May 03, 2022 12:32:45 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect start time metrics (co-input) as step s7
May 03, 2022 12:32:45 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Window.Into()2/Window.Assign as step s8
May 03, 2022 12:32:45 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/MakeUnionTable0 as step s9
May 03, 2022 12:32:45 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/MakeUnionTable1 as step s10
May 03, 2022 12:32:45 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/Flatten as step s11
May 03, 2022 12:32:45 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/GBK as step s12
May 03, 2022 12:32:45 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/ConstructCoGbkResultFn as step s13
May 03, 2022 12:32:45 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Ungroup and reiterate as step s14
May 03, 2022 12:32:45 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect total bytes as step s15
May 03, 2022 12:32:45 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect end time metrics as step s16
May 03, 2022 12:32:45 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Dataflow SDK version: 2.39.0-SNAPSHOT
May 03, 2022 12:32:46 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To access the Dataflow monitoring console, please navigate to https://console.cloud.google.com/dataflow/jobs/us-central1/2022-05-03_05_32_46-11901399501120612900?project=apache-beam-testing
May 03, 2022 12:32:46 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Submitted job: 2022-05-03_05_32_46-11901399501120612900
May 03, 2022 12:32:46 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To cancel the job using the 'gcloud' tool, run:
> gcloud dataflow jobs --project=apache-beam-testing cancel --region=us-central1 2022-05-03_05_32_46-11901399501120612900
May 03, 2022 12:32:53 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
WARNING: 2022-05-03T12:32:50.653Z: The workflow name is not a valid Cloud Label. Labels applied to Cloud resources (such as GCE Instances) for monitoring will be labeled with this modified job name: load0tests0java170dataflow0v20streaming0cogbk01-jenkins-05-trww. For the best monitoring experience, please name your job with a valid Cloud Label. For details, see: https://cloud.google.com/compute/docs/labeling-resources#restrictions
May 03, 2022 12:32:58 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-03T12:32:58.016Z: Worker configuration: e2-standard-2 in us-central1-b.
May 03, 2022 12:33:00 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-03T12:32:58.605Z: Expanding SplittableParDo operations into optimizable parts.
May 03, 2022 12:33:00 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-03T12:32:58.682Z: Expanding CollectionToSingleton operations into optimizable parts.
May 03, 2022 12:33:00 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-03T12:32:59.174Z: Expanding CoGroupByKey operations into optimizable parts.
May 03, 2022 12:33:00 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-03T12:32:59.417Z: Expanding SplittableProcessKeyed operations into optimizable parts.
May 03, 2022 12:33:00 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-03T12:32:59.458Z: Expanding GroupByKey operations into streaming Read/Write steps
May 03, 2022 12:33:00 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-03T12:32:59.533Z: Lifting ValueCombiningMappingFns into MergeBucketsMappingFns
May 03, 2022 12:33:00 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-03T12:32:59.645Z: Fusing adjacent ParDo, Read, Write, and Flatten operations
May 03, 2022 12:33:00 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-03T12:32:59.687Z: Unzipping flatten CoGroupByKey-Flatten for input CoGroupByKey-MakeUnionTable0-ParMultiDo-ConstructUnionTable-.output
May 03, 2022 12:33:00 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-03T12:32:59.721Z: Fusing unzipped copy of CoGroupByKey/GBK/WriteStream, through flatten CoGroupByKey/Flatten, into producer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable)
May 03, 2022 12:33:00 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-03T12:32:59.767Z: Fusing consumer CoGroupByKey/GBK/WriteStream into CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable)
May 03, 2022 12:33:00 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-03T12:32:59.798Z: Fusing consumer Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read input/Impulse
May 03, 2022 12:33:00 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-03T12:32:59.831Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
May 03, 2022 12:33:00 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-03T12:32:59.865Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
May 03, 2022 12:33:00 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-03T12:32:59.897Z: Fusing consumer Read input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
May 03, 2022 12:33:00 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-03T12:32:59.929Z: Fusing consumer Collect start time metrics (input)/ParMultiDo(TimeMonitor) into Read input/ParDo(StripIds)/ParMultiDo(StripIds)
May 03, 2022 12:33:00 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-03T12:32:59.964Z: Fusing consumer Window.Into()/Window.Assign into Collect start time metrics (input)/ParMultiDo(TimeMonitor)
May 03, 2022 12:33:00 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-03T12:32:59.995Z: Fusing consumer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable) into Window.Into()/Window.Assign
May 03, 2022 12:33:00 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-03T12:33:00.049Z: Fusing consumer Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read co-input/Impulse
May 03, 2022 12:33:00 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-03T12:33:00.097Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
May 03, 2022 12:33:00 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-03T12:33:00.227Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
May 03, 2022 12:33:00 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-03T12:33:00.316Z: Fusing consumer Read co-input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
May 03, 2022 12:33:03 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-03T12:33:00.528Z: Fusing consumer Collect start time metrics (co-input)/ParMultiDo(TimeMonitor) into Read co-input/ParDo(StripIds)/ParMultiDo(StripIds)
May 03, 2022 12:33:03 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-03T12:33:00.660Z: Fusing consumer Window.Into()2/Window.Assign into Collect start time metrics (co-input)/ParMultiDo(TimeMonitor)
May 03, 2022 12:33:03 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-03T12:33:00.811Z: Fusing consumer CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable) into Window.Into()2/Window.Assign
May 03, 2022 12:33:03 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-03T12:33:00.895Z: Fusing consumer CoGroupByKey/GBK/MergeBuckets into CoGroupByKey/GBK/ReadStream
May 03, 2022 12:33:03 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-03T12:33:00.945Z: Fusing consumer CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult) into CoGroupByKey/GBK/MergeBuckets
May 03, 2022 12:33:03 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-03T12:33:00.987Z: Fusing consumer Ungroup and reiterate/ParMultiDo(UngroupAndReiterate) into CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult)
May 03, 2022 12:33:03 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-03T12:33:01.020Z: Fusing consumer Collect total bytes/ParMultiDo(ByteMonitor) into Ungroup and reiterate/ParMultiDo(UngroupAndReiterate)
May 03, 2022 12:33:03 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-03T12:33:01.061Z: Fusing consumer Collect end time metrics/ParMultiDo(TimeMonitor) into Collect total bytes/ParMultiDo(ByteMonitor)
May 03, 2022 12:33:03 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-03T12:33:01.275Z: Running job using Streaming Engine
May 03, 2022 12:33:03 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-03T12:33:01.528Z: Starting 5 ****s in us-central1-b...
May 03, 2022 12:33:19 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-03T12:33:18.790Z: Your project already contains 100 Dataflow-created metric descriptors, so new user metrics of the form custom.googleapis.com/* will not be created. However, all user metrics are also available in the metric dataflow.googleapis.com/job/user_counter. If you rely on the custom metrics, you can delete old / unused metric descriptors. See https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.list and https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.delete
May 03, 2022 12:33:25 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-03T12:33:24.290Z: Autoscaling: Raised the number of ****s to 5 so that the pipeline can catch up with its backlog and keep up with its input rate.
May 03, 2022 12:34:37 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-05-03T12:34:36.212Z: Workers have started successfully.
FATAL: command execution failed
java.io.IOException: Backing channel 'apache-beam-jenkins-10' is disconnected.
	at hudson.remoting.RemoteInvocationHandler.channelOrFail(RemoteInvocationHandler.java:216)
	at hudson.remoting.RemoteInvocationHandler.invoke(RemoteInvocationHandler.java:286)
	at com.sun.proxy.$Proxy132.isAlive(Unknown Source)
	at hudson.Launcher$RemoteLauncher$ProcImpl.isAlive(Launcher.java:1213)
	at hudson.Launcher$RemoteLauncher$ProcImpl.join(Launcher.java:1205)
	at hudson.Launcher$ProcStarter.join(Launcher.java:522)
	at hudson.plugins.gradle.Gradle.perform(Gradle.java:317)
	at hudson.tasks.BuildStepMonitor$1.perform(BuildStepMonitor.java:20)
	at hudson.model.AbstractBuild$AbstractBuildExecution.perform(AbstractBuild.java:806)
	at hudson.model.Build$BuildExecution.build(Build.java:198)
	at hudson.model.Build$BuildExecution.doRun(Build.java:163)
	at hudson.model.AbstractBuild$AbstractBuildExecution.run(AbstractBuild.java:514)
	at hudson.model.Run.execute(Run.java:1888)
	at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:43)
	at hudson.model.ResourceController.execute(ResourceController.java:99)
	at hudson.model.Executor.run(Executor.java:432)
Caused by: hudson.remoting.Channel$OrderlyShutdown: Command Close created at
	at hudson.remoting.Channel$CloseCommand.execute(Channel.java:1320)
	at hudson.remoting.Channel$1.handle(Channel.java:607)
	at hudson.remoting.SynchronousCommandTransport$ReaderThread.run(SynchronousCommandTransport.java:81)
Caused by: Command Close created at
	at hudson.remoting.Command.<init>(Command.java:70)
	at hudson.remoting.Channel$CloseCommand.<init>(Channel.java:1313)
	at hudson.remoting.Channel$CloseCommand.<init>(Channel.java:1311)
	at hudson.remoting.Channel.close(Channel.java:1487)
	at hudson.remoting.Channel.close(Channel.java:1454)
	at hudson.remoting.Channel$CloseCommand.execute(Channel.java:1319)
	... 2 more
Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure
ERROR: apache-beam-jenkins-10 is offline; cannot locate jdk_1.8_latest

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #133

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/133/display/redirect?page=changes>

Changes:

[noreply] [BEAM-11105] Add manual watermark estimation (#17475)


------------------------------------------
[...truncated 388.69 KB...]
    return opener.open(url, data, timeout)
  File "/usr/lib/python3.8/urllib/request.py", line 531, in open
    response = meth(req, response)
  File "/usr/lib/python3.8/urllib/request.py", line 640, in http_response
    response = self.parent.error(
  File "/usr/lib/python3.8/urllib/request.py", line 569, in error
    return self._call_chain(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 502, in _call_chain
    result = func(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 649, in http_error_default
    raise HTTPError(req.full_url, code, msg, hdrs, fp)
urllib.error.HTTPError: HTTP Error 403: Forbidden
ERROR:root:Invalid url for slf4j-api-1.7.30: http://www.opensource.org/licenses/mit-license.php after 9 retries.
Traceback (most recent call last):
  File "<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/license_scripts/pull_licenses_java.py",> line 58, in pull_from_url
    url_read = urlopen(url)
  File "/usr/lib/python3.8/urllib/request.py", line 222, in urlopen
    return opener.open(url, data, timeout)
  File "/usr/lib/python3.8/urllib/request.py", line 531, in open
    response = meth(req, response)
  File "/usr/lib/python3.8/urllib/request.py", line 640, in http_response
    response = self.parent.error(
  File "/usr/lib/python3.8/urllib/request.py", line 569, in error
    return self._call_chain(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 502, in _call_chain
    result = func(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 649, in http_error_default
    raise HTTPError(req.full_url, code, msg, hdrs, fp)
urllib.error.HTTPError: HTTP Error 403: Forbidden
ERROR:root:Invalid url for jdbc-1.16.3: http://opensource.org/licenses/MIT after 9 retries.
Traceback (most recent call last):
  File "<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/license_scripts/pull_licenses_java.py",> line 58, in pull_from_url
    url_read = urlopen(url)
  File "/usr/lib/python3.8/urllib/request.py", line 222, in urlopen
    return opener.open(url, data, timeout)
  File "/usr/lib/python3.8/urllib/request.py", line 531, in open
    response = meth(req, response)
  File "/usr/lib/python3.8/urllib/request.py", line 640, in http_response
    response = self.parent.error(
  File "/usr/lib/python3.8/urllib/request.py", line 569, in error
    return self._call_chain(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 502, in _call_chain
    result = func(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 649, in http_error_default
    raise HTTPError(req.full_url, code, msg, hdrs, fp)
urllib.error.HTTPError: HTTP Error 403: Forbidden
ERROR:root:Invalid url for kafka-1.16.3: http://opensource.org/licenses/MIT. Retrying...
Traceback (most recent call last):
  File "<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/license_scripts/pull_licenses_java.py",> line 58, in pull_from_url
    url_read = urlopen(url)
  File "/usr/lib/python3.8/urllib/request.py", line 222, in urlopen
    return opener.open(url, data, timeout)
  File "/usr/lib/python3.8/urllib/request.py", line 531, in open
    response = meth(req, response)
  File "/usr/lib/python3.8/urllib/request.py", line 640, in http_response
    response = self.parent.error(
  File "/usr/lib/python3.8/urllib/request.py", line 569, in error
    return self._call_chain(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 502, in _call_chain
    result = func(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 649, in http_error_default
    raise HTTPError(req.full_url, code, msg, hdrs, fp)
urllib.error.HTTPError: HTTP Error 403: Forbidden
ERROR:root:Invalid url for postgresql-1.16.3: http://opensource.org/licenses/MIT after 9 retries.
Traceback (most recent call last):
  File "<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/license_scripts/pull_licenses_java.py",> line 58, in pull_from_url
    url_read = urlopen(url)
  File "/usr/lib/python3.8/urllib/request.py", line 222, in urlopen
    return opener.open(url, data, timeout)
  File "/usr/lib/python3.8/urllib/request.py", line 531, in open
    response = meth(req, response)
  File "/usr/lib/python3.8/urllib/request.py", line 640, in http_response
    response = self.parent.error(
  File "/usr/lib/python3.8/urllib/request.py", line 569, in error
    return self._call_chain(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 502, in _call_chain
    result = func(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 649, in http_error_default
    raise HTTPError(req.full_url, code, msg, hdrs, fp)
urllib.error.HTTPError: HTTP Error 403: Forbidden
ERROR:root:Invalid url for kafka-1.16.3: http://opensource.org/licenses/MIT. Retrying...
Traceback (most recent call last):
  File "<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/license_scripts/pull_licenses_java.py",> line 58, in pull_from_url
    url_read = urlopen(url)
  File "/usr/lib/python3.8/urllib/request.py", line 222, in urlopen
    return opener.open(url, data, timeout)
  File "/usr/lib/python3.8/urllib/request.py", line 531, in open
    response = meth(req, response)
  File "/usr/lib/python3.8/urllib/request.py", line 640, in http_response
    response = self.parent.error(
  File "/usr/lib/python3.8/urllib/request.py", line 569, in error
    return self._call_chain(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 502, in _call_chain
    result = func(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 649, in http_error_default
    raise HTTPError(req.full_url, code, msg, hdrs, fp)
urllib.error.HTTPError: HTTP Error 403: Forbidden
ERROR:root:Invalid url for kafka-1.16.3: http://opensource.org/licenses/MIT. Retrying...
Traceback (most recent call last):
  File "<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/license_scripts/pull_licenses_java.py",> line 58, in pull_from_url
    url_read = urlopen(url)
  File "/usr/lib/python3.8/urllib/request.py", line 222, in urlopen
    return opener.open(url, data, timeout)
  File "/usr/lib/python3.8/urllib/request.py", line 531, in open
    response = meth(req, response)
  File "/usr/lib/python3.8/urllib/request.py", line 640, in http_response
    response = self.parent.error(
  File "/usr/lib/python3.8/urllib/request.py", line 569, in error
    return self._call_chain(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 502, in _call_chain
    result = func(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 649, in http_error_default
    raise HTTPError(req.full_url, code, msg, hdrs, fp)
urllib.error.HTTPError: HTTP Error 403: Forbidden
ERROR:root:Invalid url for kafka-1.16.3: http://opensource.org/licenses/MIT. Retrying...
Traceback (most recent call last):
  File "<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/license_scripts/pull_licenses_java.py",> line 58, in pull_from_url
    url_read = urlopen(url)
  File "/usr/lib/python3.8/urllib/request.py", line 222, in urlopen
    return opener.open(url, data, timeout)
  File "/usr/lib/python3.8/urllib/request.py", line 531, in open
    response = meth(req, response)
  File "/usr/lib/python3.8/urllib/request.py", line 640, in http_response
    response = self.parent.error(
  File "/usr/lib/python3.8/urllib/request.py", line 569, in error
    return self._call_chain(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 502, in _call_chain
    result = func(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 649, in http_error_default
    raise HTTPError(req.full_url, code, msg, hdrs, fp)
urllib.error.HTTPError: HTTP Error 403: Forbidden
ERROR:root:Invalid url for kafka-1.16.3: http://opensource.org/licenses/MIT. Retrying...
Traceback (most recent call last):
  File "<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/license_scripts/pull_licenses_java.py",> line 58, in pull_from_url
    url_read = urlopen(url)
  File "/usr/lib/python3.8/urllib/request.py", line 222, in urlopen
    return opener.open(url, data, timeout)
  File "/usr/lib/python3.8/urllib/request.py", line 531, in open
    response = meth(req, response)
  File "/usr/lib/python3.8/urllib/request.py", line 640, in http_response
    response = self.parent.error(
  File "/usr/lib/python3.8/urllib/request.py", line 569, in error
    return self._call_chain(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 502, in _call_chain
    result = func(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 649, in http_error_default
    raise HTTPError(req.full_url, code, msg, hdrs, fp)
urllib.error.HTTPError: HTTP Error 403: Forbidden
ERROR:root:Invalid url for kafka-1.16.3: http://opensource.org/licenses/MIT. Retrying...
Traceback (most recent call last):
  File "<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/license_scripts/pull_licenses_java.py",> line 58, in pull_from_url
    url_read = urlopen(url)
  File "/usr/lib/python3.8/urllib/request.py", line 222, in urlopen
    return opener.open(url, data, timeout)
  File "/usr/lib/python3.8/urllib/request.py", line 531, in open
    response = meth(req, response)
  File "/usr/lib/python3.8/urllib/request.py", line 640, in http_response
    response = self.parent.error(
  File "/usr/lib/python3.8/urllib/request.py", line 569, in error
    return self._call_chain(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 502, in _call_chain
    result = func(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 649, in http_error_default
    raise HTTPError(req.full_url, code, msg, hdrs, fp)
urllib.error.HTTPError: HTTP Error 403: Forbidden
ERROR:root:Invalid url for kafka-1.16.3: http://opensource.org/licenses/MIT. Retrying...
Traceback (most recent call last):
  File "<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/license_scripts/pull_licenses_java.py",> line 58, in pull_from_url
    url_read = urlopen(url)
  File "/usr/lib/python3.8/urllib/request.py", line 222, in urlopen
    return opener.open(url, data, timeout)
  File "/usr/lib/python3.8/urllib/request.py", line 531, in open
    response = meth(req, response)
  File "/usr/lib/python3.8/urllib/request.py", line 640, in http_response
    response = self.parent.error(
  File "/usr/lib/python3.8/urllib/request.py", line 569, in error
    return self._call_chain(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 502, in _call_chain
    result = func(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 649, in http_error_default
    raise HTTPError(req.full_url, code, msg, hdrs, fp)
urllib.error.HTTPError: HTTP Error 403: Forbidden
ERROR:root:Invalid url for kafka-1.16.3: http://opensource.org/licenses/MIT. Retrying...
Traceback (most recent call last):
  File "<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/license_scripts/pull_licenses_java.py",> line 58, in pull_from_url
    url_read = urlopen(url)
  File "/usr/lib/python3.8/urllib/request.py", line 222, in urlopen
    return opener.open(url, data, timeout)
  File "/usr/lib/python3.8/urllib/request.py", line 531, in open
    response = meth(req, response)
  File "/usr/lib/python3.8/urllib/request.py", line 640, in http_response
    response = self.parent.error(
  File "/usr/lib/python3.8/urllib/request.py", line 569, in error
    return self._call_chain(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 502, in _call_chain
    result = func(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 649, in http_error_default
    raise HTTPError(req.full_url, code, msg, hdrs, fp)
urllib.error.HTTPError: HTTP Error 403: Forbidden
ERROR:root:Invalid url for kafka-1.16.3: http://opensource.org/licenses/MIT after 9 retries.
ERROR:root:['system-rules-1.19.0', 'google-auth-library-oauth2-http-1.4.0', 'reflectasm-1.07', 'software-and-algorithms-1.0', 'zstd-jni-1.4.3-1', 'kryo-2.21', 'junit-quickcheck-generators-0.8', 'google-auth-library-credentials-1.4.0', 'protobuf-java-util-3.19.3', 'protobuf-java-3.19.3', 'classgraph-4.8.104', 'grpc-context-1.44.0', 'grpc-alts-1.44.0', 'grpc-protobuf-1.44.0', 'perfmark-api-0.23.0', 'junit-dep-4.11', 'minlog-1.2', 'zstd-jni-1.5.2-1', 'junit-quickcheck-core-0.8', 'checker-compat-qual-2.5.3', 'grpc-core-1.44.0', 'grpc-api-1.44.0', 'grpc-protobuf-lite-1.44.0', 'pcollections-2.1.2', 'hamcrest-2.1', 'slf4j-jdk14-1.7.30', 'duct-tape-1.0.8', 'database-commons-1.16.3', 'mysql-1.16.3', 'checker-compat-qual-2.5.5', 'grpc-grpclb-1.44.0', 'grpc-auth-1.44.0', 'grpc-stub-1.44.0', 'slf4j-simple-1.7.30', 'slf4j-api-1.7.30', 'jdbc-1.16.3', 'postgresql-1.16.3', 'kafka-1.16.3']
ERROR:root:**************************************** Licenses were not able to be pulled automatically for some dependencies. Please search source code of the dependencies on the internet and add "license" and "notice" (if available) field to <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/license_scripts/dep_urls_java.yaml> for each missing license. Dependency List: [checker-compat-qual-2.5.3,checker-compat-qual-2.5.5,classgraph-4.8.104,database-commons-1.16.3,duct-tape-1.0.8,google-auth-library-credentials-1.4.0,google-auth-library-oauth2-http-1.4.0,grpc-alts-1.44.0,grpc-api-1.44.0,grpc-auth-1.44.0,grpc-context-1.44.0,grpc-core-1.44.0,grpc-grpclb-1.44.0,grpc-protobuf-1.44.0,grpc-protobuf-lite-1.44.0,grpc-stub-1.44.0,hamcrest-2.1,jdbc-1.16.3,junit-dep-4.11,junit-quickcheck-core-0.8,junit-quickcheck-generators-0.8,kafka-1.16.3,kryo-2.21,minlog-1.2,mysql-1.16.3,pcollections-2.1.2,perfmark-api-0.23.0,postgresql-1.16.3,protobuf-java-3.19.3,protobuf-java-util-3.19.3,reflectasm-1.07,slf4j-api-1.7.30,slf4j-jdk14-1.7.30,slf4j-simple-1.7.30,software-and-algorithms-1.0,system-rules-1.19.0,zstd-jni-1.4.3-1,zstd-jni-1.5.2-1]
INFO:root:pull_licenses_java.py failed. It took 162.359743 seconds with 16 threads.
Traceback (most recent call last):
  File "<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/license_scripts/pull_licenses_java.py",> line 321, in <module>
    raise RuntimeError('{n} error(s) occurred.'.format(n=len(error_msg)),
RuntimeError: ('1 error(s) occurred.', ['**************************************** Licenses were not able to be pulled automatically for some dependencies. Please search source code of the dependencies on the internet and add "license" and "notice" (if available) field to <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/license_scripts/dep_urls_java.yaml> for each missing license. Dependency List: [checker-compat-qual-2.5.3,checker-compat-qual-2.5.5,classgraph-4.8.104,database-commons-1.16.3,duct-tape-1.0.8,google-auth-library-credentials-1.4.0,google-auth-library-oauth2-http-1.4.0,grpc-alts-1.44.0,grpc-api-1.44.0,grpc-auth-1.44.0,grpc-context-1.44.0,grpc-core-1.44.0,grpc-grpclb-1.44.0,grpc-protobuf-1.44.0,grpc-protobuf-lite-1.44.0,grpc-stub-1.44.0,hamcrest-2.1,jdbc-1.16.3,junit-dep-4.11,junit-quickcheck-core-0.8,junit-quickcheck-generators-0.8,kafka-1.16.3,kryo-2.21,minlog-1.2,mysql-1.16.3,pcollections-2.1.2,perfmark-api-0.23.0,postgresql-1.16.3,protobuf-java-3.19.3,protobuf-java-util-3.19.3,reflectasm-1.07,slf4j-api-1.7.30,slf4j-jdk14-1.7.30,slf4j-simple-1.7.30,software-and-algorithms-1.0,system-rules-1.19.0,zstd-jni-1.4.3-1,zstd-jni-1.5.2-1]'])

> Task :sdks:java:container:pullLicenses FAILED
> Task :sdks:java:container:goPrepare UP-TO-DATE

> Task :sdks:java:container:goBuild
/home/jenkins/go/bin/go1.16.12 build -o ./build/target/linux_amd64/boot boot.go

> Task :sdks:java:container:java17:copySdkHarnessLauncher
Execution optimizations have been disabled for task ':sdks:java:container:java17:copySdkHarnessLauncher' to ensure correctness due to the following reasons:
  - Gradle detected a problem with the following location: '<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/build/target'.> Reason: Task ':sdks:java:container:java17:copySdkHarnessLauncher' uses this output of task ':sdks:java:container:downloadCloudProfilerAgent' without declaring an explicit or implicit dependency. This can lead to incorrect results being produced, depending on what order the tasks are executed. Please refer to https://docs.gradle.org/7.4/userguide/validation_problems.html#implicit_dependency for more details about this problem.
  - Gradle detected a problem with the following location: '<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/build/target'.> Reason: Task ':sdks:java:container:java17:copySdkHarnessLauncher' uses this output of task ':sdks:java:container:pullLicenses' without declaring an explicit or implicit dependency. This can lead to incorrect results being produced, depending on what order the tasks are executed. Please refer to https://docs.gradle.org/7.4/userguide/validation_problems.html#implicit_dependency for more details about this problem.

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:container:pullLicenses'.
> Process 'command './license_scripts/license_script.sh'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.4/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3m 15s
104 actionable tasks: 67 executed, 33 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/om7opgkgr7qm2

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #132

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/132/display/redirect?page=changes>

Changes:

[noreply] Revert "Improvement to Seed job configuration to launch against PRs

[ilion.beyst] Minor: fix typo

[noreply] Merge pull request #17422 from [BEAM-14344]: remove tracing from


------------------------------------------
[...truncated 384.85 KB...]
    return opener.open(url, data, timeout)
  File "/usr/lib/python3.8/urllib/request.py", line 531, in open
    response = meth(req, response)
  File "/usr/lib/python3.8/urllib/request.py", line 640, in http_response
    response = self.parent.error(
  File "/usr/lib/python3.8/urllib/request.py", line 569, in error
    return self._call_chain(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 502, in _call_chain
    result = func(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 649, in http_error_default
    raise HTTPError(req.full_url, code, msg, hdrs, fp)
urllib.error.HTTPError: HTTP Error 403: Forbidden
ERROR:root:Invalid url for grpc-stub-1.44.0: https://opensource.org/licenses/Apache-2.0 after 9 retries.
Traceback (most recent call last):
  File "<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/license_scripts/pull_licenses_java.py",> line 58, in pull_from_url
    url_read = urlopen(url)
  File "/usr/lib/python3.8/urllib/request.py", line 222, in urlopen
    return opener.open(url, data, timeout)
  File "/usr/lib/python3.8/urllib/request.py", line 531, in open
    response = meth(req, response)
  File "/usr/lib/python3.8/urllib/request.py", line 640, in http_response
    response = self.parent.error(
  File "/usr/lib/python3.8/urllib/request.py", line 569, in error
    return self._call_chain(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 502, in _call_chain
    result = func(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 649, in http_error_default
    raise HTTPError(req.full_url, code, msg, hdrs, fp)
urllib.error.HTTPError: HTTP Error 403: Forbidden
ERROR:root:Invalid url for slf4j-simple-1.7.30: http://www.opensource.org/licenses/mit-license.php after 9 retries.
Traceback (most recent call last):
  File "<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/license_scripts/pull_licenses_java.py",> line 58, in pull_from_url
    url_read = urlopen(url)
  File "/usr/lib/python3.8/urllib/request.py", line 222, in urlopen
    return opener.open(url, data, timeout)
  File "/usr/lib/python3.8/urllib/request.py", line 531, in open
    response = meth(req, response)
  File "/usr/lib/python3.8/urllib/request.py", line 640, in http_response
    response = self.parent.error(
  File "/usr/lib/python3.8/urllib/request.py", line 569, in error
    return self._call_chain(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 502, in _call_chain
    result = func(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 649, in http_error_default
    raise HTTPError(req.full_url, code, msg, hdrs, fp)
urllib.error.HTTPError: HTTP Error 403: Forbidden
ERROR:root:Invalid url for kafka-1.16.3: http://opensource.org/licenses/MIT. Retrying...
Traceback (most recent call last):
  File "<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/license_scripts/pull_licenses_java.py",> line 58, in pull_from_url
    url_read = urlopen(url)
  File "/usr/lib/python3.8/urllib/request.py", line 222, in urlopen
    return opener.open(url, data, timeout)
  File "/usr/lib/python3.8/urllib/request.py", line 531, in open
    response = meth(req, response)
  File "/usr/lib/python3.8/urllib/request.py", line 640, in http_response
    response = self.parent.error(
  File "/usr/lib/python3.8/urllib/request.py", line 569, in error
    return self._call_chain(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 502, in _call_chain
    result = func(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 649, in http_error_default
    raise HTTPError(req.full_url, code, msg, hdrs, fp)
urllib.error.HTTPError: HTTP Error 403: Forbidden
ERROR:root:Invalid url for postgresql-1.16.3: http://opensource.org/licenses/MIT after 9 retries.
Traceback (most recent call last):
  File "<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/license_scripts/pull_licenses_java.py",> line 58, in pull_from_url
    url_read = urlopen(url)
  File "/usr/lib/python3.8/urllib/request.py", line 222, in urlopen
    return opener.open(url, data, timeout)
  File "/usr/lib/python3.8/urllib/request.py", line 531, in open
    response = meth(req, response)
  File "/usr/lib/python3.8/urllib/request.py", line 640, in http_response
    response = self.parent.error(
  File "/usr/lib/python3.8/urllib/request.py", line 569, in error
    return self._call_chain(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 502, in _call_chain
    result = func(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 649, in http_error_default
    raise HTTPError(req.full_url, code, msg, hdrs, fp)
urllib.error.HTTPError: HTTP Error 403: Forbidden
ERROR:root:Invalid url for kafka-1.16.3: http://opensource.org/licenses/MIT. Retrying...
Traceback (most recent call last):
  File "<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/license_scripts/pull_licenses_java.py",> line 58, in pull_from_url
    url_read = urlopen(url)
  File "/usr/lib/python3.8/urllib/request.py", line 222, in urlopen
    return opener.open(url, data, timeout)
  File "/usr/lib/python3.8/urllib/request.py", line 531, in open
    response = meth(req, response)
  File "/usr/lib/python3.8/urllib/request.py", line 640, in http_response
    response = self.parent.error(
  File "/usr/lib/python3.8/urllib/request.py", line 569, in error
    return self._call_chain(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 502, in _call_chain
    result = func(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 649, in http_error_default
    raise HTTPError(req.full_url, code, msg, hdrs, fp)
urllib.error.HTTPError: HTTP Error 403: Forbidden
ERROR:root:Invalid url for kafka-1.16.3: http://opensource.org/licenses/MIT. Retrying...
Traceback (most recent call last):
  File "<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/license_scripts/pull_licenses_java.py",> line 58, in pull_from_url
    url_read = urlopen(url)
  File "/usr/lib/python3.8/urllib/request.py", line 222, in urlopen
    return opener.open(url, data, timeout)
  File "/usr/lib/python3.8/urllib/request.py", line 531, in open
    response = meth(req, response)
  File "/usr/lib/python3.8/urllib/request.py", line 640, in http_response
    response = self.parent.error(
  File "/usr/lib/python3.8/urllib/request.py", line 569, in error
    return self._call_chain(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 502, in _call_chain
    result = func(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 649, in http_error_default
    raise HTTPError(req.full_url, code, msg, hdrs, fp)
urllib.error.HTTPError: HTTP Error 403: Forbidden
ERROR:root:Invalid url for kafka-1.16.3: http://opensource.org/licenses/MIT. Retrying...
Traceback (most recent call last):
  File "<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/license_scripts/pull_licenses_java.py",> line 58, in pull_from_url
    url_read = urlopen(url)
  File "/usr/lib/python3.8/urllib/request.py", line 222, in urlopen
    return opener.open(url, data, timeout)
  File "/usr/lib/python3.8/urllib/request.py", line 531, in open
    response = meth(req, response)
  File "/usr/lib/python3.8/urllib/request.py", line 640, in http_response
    response = self.parent.error(
  File "/usr/lib/python3.8/urllib/request.py", line 569, in error
    return self._call_chain(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 502, in _call_chain
    result = func(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 649, in http_error_default
    raise HTTPError(req.full_url, code, msg, hdrs, fp)
urllib.error.HTTPError: HTTP Error 403: Forbidden
ERROR:root:Invalid url for kafka-1.16.3: http://opensource.org/licenses/MIT. Retrying...
Traceback (most recent call last):
  File "<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/license_scripts/pull_licenses_java.py",> line 58, in pull_from_url
    url_read = urlopen(url)
  File "/usr/lib/python3.8/urllib/request.py", line 222, in urlopen
    return opener.open(url, data, timeout)
  File "/usr/lib/python3.8/urllib/request.py", line 531, in open
    response = meth(req, response)
  File "/usr/lib/python3.8/urllib/request.py", line 640, in http_response
    response = self.parent.error(
  File "/usr/lib/python3.8/urllib/request.py", line 569, in error
    return self._call_chain(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 502, in _call_chain
    result = func(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 649, in http_error_default
    raise HTTPError(req.full_url, code, msg, hdrs, fp)
urllib.error.HTTPError: HTTP Error 403: Forbidden
ERROR:root:Invalid url for kafka-1.16.3: http://opensource.org/licenses/MIT. Retrying...
Traceback (most recent call last):
  File "<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/license_scripts/pull_licenses_java.py",> line 58, in pull_from_url
    url_read = urlopen(url)
  File "/usr/lib/python3.8/urllib/request.py", line 222, in urlopen
    return opener.open(url, data, timeout)
  File "/usr/lib/python3.8/urllib/request.py", line 531, in open
    response = meth(req, response)
  File "/usr/lib/python3.8/urllib/request.py", line 640, in http_response
    response = self.parent.error(
  File "/usr/lib/python3.8/urllib/request.py", line 569, in error
    return self._call_chain(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 502, in _call_chain
    result = func(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 649, in http_error_default
    raise HTTPError(req.full_url, code, msg, hdrs, fp)
urllib.error.HTTPError: HTTP Error 403: Forbidden
ERROR:root:Invalid url for kafka-1.16.3: http://opensource.org/licenses/MIT. Retrying...
Traceback (most recent call last):
  File "<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/license_scripts/pull_licenses_java.py",> line 58, in pull_from_url
    url_read = urlopen(url)
  File "/usr/lib/python3.8/urllib/request.py", line 222, in urlopen
    return opener.open(url, data, timeout)
  File "/usr/lib/python3.8/urllib/request.py", line 531, in open
    response = meth(req, response)
  File "/usr/lib/python3.8/urllib/request.py", line 640, in http_response
    response = self.parent.error(
  File "/usr/lib/python3.8/urllib/request.py", line 569, in error
    return self._call_chain(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 502, in _call_chain
    result = func(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 649, in http_error_default
    raise HTTPError(req.full_url, code, msg, hdrs, fp)
urllib.error.HTTPError: HTTP Error 403: Forbidden
ERROR:root:Invalid url for kafka-1.16.3: http://opensource.org/licenses/MIT. Retrying...
Traceback (most recent call last):
  File "<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/license_scripts/pull_licenses_java.py",> line 58, in pull_from_url
    url_read = urlopen(url)
  File "/usr/lib/python3.8/urllib/request.py", line 222, in urlopen
    return opener.open(url, data, timeout)
  File "/usr/lib/python3.8/urllib/request.py", line 531, in open
    response = meth(req, response)
  File "/usr/lib/python3.8/urllib/request.py", line 640, in http_response
    response = self.parent.error(
  File "/usr/lib/python3.8/urllib/request.py", line 569, in error
    return self._call_chain(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 502, in _call_chain
    result = func(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 649, in http_error_default
    raise HTTPError(req.full_url, code, msg, hdrs, fp)
urllib.error.HTTPError: HTTP Error 403: Forbidden
ERROR:root:Invalid url for kafka-1.16.3: http://opensource.org/licenses/MIT after 9 retries.
ERROR:root:['google-auth-library-oauth2-http-1.4.0', 'reflectasm-1.07', 'system-rules-1.19.0', 'zstd-jni-1.4.3-1', 'software-and-algorithms-1.0', 'google-auth-library-credentials-1.4.0', 'kryo-2.21', 'junit-quickcheck-generators-0.8', 'protobuf-java-util-3.19.3', 'classgraph-4.8.104', 'protobuf-java-3.19.3', 'grpc-context-1.44.0', 'grpc-alts-1.44.0', 'grpc-protobuf-1.44.0', 'perfmark-api-0.23.0', 'junit-dep-4.11', 'minlog-1.2', 'zstd-jni-1.5.2-1', 'junit-quickcheck-core-0.8', 'checker-compat-qual-2.5.3', 'grpc-core-1.44.0', 'grpc-api-1.44.0', 'grpc-protobuf-lite-1.44.0', 'pcollections-2.1.2', 'duct-tape-1.0.8', 'hamcrest-2.1', 'database-commons-1.16.3', 'slf4j-jdk14-1.7.30', 'mysql-1.16.3', 'checker-compat-qual-2.5.5', 'grpc-grpclb-1.44.0', 'slf4j-api-1.7.30', 'grpc-auth-1.44.0', 'jdbc-1.16.3', 'grpc-stub-1.44.0', 'slf4j-simple-1.7.30', 'postgresql-1.16.3', 'kafka-1.16.3']
ERROR:root:**************************************** Licenses were not able to be pulled automatically for some dependencies. Please search source code of the dependencies on the internet and add "license" and "notice" (if available) field to <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/license_scripts/dep_urls_java.yaml> for each missing license. Dependency List: [checker-compat-qual-2.5.3,checker-compat-qual-2.5.5,classgraph-4.8.104,database-commons-1.16.3,duct-tape-1.0.8,google-auth-library-credentials-1.4.0,google-auth-library-oauth2-http-1.4.0,grpc-alts-1.44.0,grpc-api-1.44.0,grpc-auth-1.44.0,grpc-context-1.44.0,grpc-core-1.44.0,grpc-grpclb-1.44.0,grpc-protobuf-1.44.0,grpc-protobuf-lite-1.44.0,grpc-stub-1.44.0,hamcrest-2.1,jdbc-1.16.3,junit-dep-4.11,junit-quickcheck-core-0.8,junit-quickcheck-generators-0.8,kafka-1.16.3,kryo-2.21,minlog-1.2,mysql-1.16.3,pcollections-2.1.2,perfmark-api-0.23.0,postgresql-1.16.3,protobuf-java-3.19.3,protobuf-java-util-3.19.3,reflectasm-1.07,slf4j-api-1.7.30,slf4j-jdk14-1.7.30,slf4j-simple-1.7.30,software-and-algorithms-1.0,system-rules-1.19.0,zstd-jni-1.4.3-1,zstd-jni-1.5.2-1]
INFO:root:pull_licenses_java.py failed. It took 162.258068 seconds with 16 threads.
Traceback (most recent call last):
  File "<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/license_scripts/pull_licenses_java.py",> line 321, in <module>
    raise RuntimeError('{n} error(s) occurred.'.format(n=len(error_msg)),
RuntimeError: ('1 error(s) occurred.', ['**************************************** Licenses were not able to be pulled automatically for some dependencies. Please search source code of the dependencies on the internet and add "license" and "notice" (if available) field to <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/license_scripts/dep_urls_java.yaml> for each missing license. Dependency List: [checker-compat-qual-2.5.3,checker-compat-qual-2.5.5,classgraph-4.8.104,database-commons-1.16.3,duct-tape-1.0.8,google-auth-library-credentials-1.4.0,google-auth-library-oauth2-http-1.4.0,grpc-alts-1.44.0,grpc-api-1.44.0,grpc-auth-1.44.0,grpc-context-1.44.0,grpc-core-1.44.0,grpc-grpclb-1.44.0,grpc-protobuf-1.44.0,grpc-protobuf-lite-1.44.0,grpc-stub-1.44.0,hamcrest-2.1,jdbc-1.16.3,junit-dep-4.11,junit-quickcheck-core-0.8,junit-quickcheck-generators-0.8,kafka-1.16.3,kryo-2.21,minlog-1.2,mysql-1.16.3,pcollections-2.1.2,perfmark-api-0.23.0,postgresql-1.16.3,protobuf-java-3.19.3,protobuf-java-util-3.19.3,reflectasm-1.07,slf4j-api-1.7.30,slf4j-jdk14-1.7.30,slf4j-simple-1.7.30,software-and-algorithms-1.0,system-rules-1.19.0,zstd-jni-1.4.3-1,zstd-jni-1.5.2-1]'])

> Task :sdks:java:container:pullLicenses FAILED
> Task :sdks:java:container:goPrepare UP-TO-DATE

> Task :sdks:java:container:goBuild
/home/jenkins/go/bin/go1.16.12 build -o ./build/target/linux_amd64/boot boot.go

> Task :sdks:java:container:java17:copySdkHarnessLauncher
Execution optimizations have been disabled for task ':sdks:java:container:java17:copySdkHarnessLauncher' to ensure correctness due to the following reasons:
  - Gradle detected a problem with the following location: '<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/build/target'.> Reason: Task ':sdks:java:container:java17:copySdkHarnessLauncher' uses this output of task ':sdks:java:container:downloadCloudProfilerAgent' without declaring an explicit or implicit dependency. This can lead to incorrect results being produced, depending on what order the tasks are executed. Please refer to https://docs.gradle.org/7.4/userguide/validation_problems.html#implicit_dependency for more details about this problem.
  - Gradle detected a problem with the following location: '<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/build/target'.> Reason: Task ':sdks:java:container:java17:copySdkHarnessLauncher' uses this output of task ':sdks:java:container:pullLicenses' without declaring an explicit or implicit dependency. This can lead to incorrect results being produced, depending on what order the tasks are executed. Please refer to https://docs.gradle.org/7.4/userguide/validation_problems.html#implicit_dependency for more details about this problem.

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:container:pullLicenses'.
> Process 'command './license_scripts/license_script.sh'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.4/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3m 15s
104 actionable tasks: 68 executed, 32 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/fsmqg7svzemh4

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #131

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/131/display/redirect?page=changes>

Changes:

[Robert Bradshaw] Add element weighting parameter to BatchElements.

[Robert Bradshaw] Clearer test.

[noreply] Revert "Merge pull request #17260 from [BEAM-13229] [Website] bug side

[noreply] [BEAM-14001] Add missing test cases to existing suites in exec package

[noreply] [BEAM-14243] Add staticcheck to Github Actions Precommits (#17479)

[noreply] [BEAM-14368][BEAM-13984]Change model loading from constructor to

[noreply] [BEAM-13983] changed file name from sklearn_loader to sklearn_inference

[noreply] Add SQL in Notebooks blog post (#17481)

[noreply] Merge pull request #17404: [BEAM-13990] support date and timestamp


------------------------------------------
[...truncated 48.53 KB...]
b66a2701ea27: Preparing
3bc383470c05: Preparing
69c1338cdef4: Waiting
e93827457889: Preparing
3810379bedde: Waiting
2b0de254a7c2: Waiting
08fa02ce37eb: Preparing
a037458de4e0: Preparing
b1d971149db0: Waiting
b66a2701ea27: Waiting
cba591b85723: Waiting
bafdbe68e4ae: Preparing
a13c519c6361: Preparing
3bc383470c05: Waiting
a13c519c6361: Waiting
bb84462952c7: Waiting
52766a5bccea: Waiting
da40c4782e7f: Waiting
bafdbe68e4ae: Waiting
a037458de4e0: Waiting
872a42936a09: Pushed
c32361a1a860: Pushed
e7c6c95e5083: Pushed
e9f9a1f648fa: Pushed
65541ee2f3b3: Pushed
69c1338cdef4: Pushed
69042358dd56: Pushed
3810379bedde: Pushed
52766a5bccea: Pushed
2b0de254a7c2: Pushed
cba591b85723: Pushed
b1d971149db0: Pushed
3bc383470c05: Layer already exists
e93827457889: Layer already exists
08fa02ce37eb: Layer already exists
a037458de4e0: Layer already exists
bafdbe68e4ae: Layer already exists
a13c519c6361: Layer already exists
bb84462952c7: Pushed
b66a2701ea27: Pushed
da40c4782e7f: Pushed
20220430123036: digest: sha256:0e47d17ca7537760a8c2dbd9549153d80ea0d48c42799e5647dcbcee5d1d52fc size: 4729

> Task :sdks:java:testing:load-tests:run
Apr 30, 2022 12:32:34 PM org.apache.beam.runners.dataflow.options.DataflowPipelineOptions$StagingLocationFactory create
INFO: No stagingLocation provided, falling back to gcpTempLocation
Apr 30, 2022 12:32:35 PM org.apache.beam.runners.dataflow.DataflowRunner fromOptions
INFO: PipelineOptions.filesToStage was not specified. Defaulting to files from the classpath: will stage 210 files. Enable logging at DEBUG level to see which files will be staged.
Apr 30, 2022 12:32:36 PM org.apache.beam.sdk.Pipeline validate
WARNING: The following transforms do not have stable unique names: Window.Into()
Apr 30, 2022 12:32:36 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Executing pipeline on the Dataflow Service, which will have billing implications related to Google Compute Engine usage and other Google Cloud Services.
Apr 30, 2022 12:32:38 PM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Uploading 210 files from PipelineOptions.filesToStage to staging location to prepare for execution.
Apr 30, 2022 12:32:39 PM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Staging files complete: 210 files cached, 0 files newly uploaded in 0 seconds
Apr 30, 2022 12:32:39 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Staging portable pipeline proto to gs://temp-storage-for-perf-tests/loadtests/staging/
Apr 30, 2022 12:32:39 PM org.apache.beam.runners.dataflow.util.PackageUtil tryStagePackage
INFO: Uploading <116094 bytes, hash 4491a2b601e674059ec8006d689c509ec298af1a0439951d10b800e6dd010454> to gs://temp-storage-for-perf-tests/loadtests/staging/pipeline-RJGitgHmdAWeyABtaJxQnsKYrxoEOZUdELgA5t0BBFQ.pb
Apr 30, 2022 12:32:40 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read input/DataflowRunner.StreamingUnboundedRead.ReadWithIds as step s1
Apr 30, 2022 12:32:40 PM org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource split
INFO: Split into 20 bundles of sizes: [org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@d5af0a5, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5981f4a6, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@63dfada0, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6f231ced, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@35a60674, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@63d4f0a2, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7d78f3d5, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@a4b5ce3, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@f5b6e78, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4b4eced1, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@71926a36, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@216e9ca3, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@75120e58, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@48976e6d, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2a367e93, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7f6874f2, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1a6dc589, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@697a34af, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@70211df5, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4c5228e7]
Apr 30, 2022 12:32:40 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read input/StripIds as step s2
Apr 30, 2022 12:32:40 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect start time metrics (input) as step s3
Apr 30, 2022 12:32:40 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Window.Into()/Window.Assign as step s4
Apr 30, 2022 12:32:40 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read co-input/DataflowRunner.StreamingUnboundedRead.ReadWithIds as step s5
Apr 30, 2022 12:32:40 PM org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource split
INFO: Split into 20 bundles of sizes: [org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7b81616b, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@15d42ccb, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@279dd959, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@46383a78, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@36c281ed, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@244418a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4b5a078a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4c361f63, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6ed922e1, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4eb166a1, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@554c4eaa, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@29fd8e67, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@e146f93, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4bd5849e, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7cdbaa50, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@39909d1a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1455154c, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7343922c, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@526b2f3e, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6f2e1024]
Apr 30, 2022 12:32:40 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read co-input/StripIds as step s6
Apr 30, 2022 12:32:40 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect start time metrics (co-input) as step s7
Apr 30, 2022 12:32:40 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Window.Into()2/Window.Assign as step s8
Apr 30, 2022 12:32:40 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/MakeUnionTable0 as step s9
Apr 30, 2022 12:32:40 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/MakeUnionTable1 as step s10
Apr 30, 2022 12:32:40 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/Flatten as step s11
Apr 30, 2022 12:32:40 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/GBK as step s12
Apr 30, 2022 12:32:40 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/ConstructCoGbkResultFn as step s13
Apr 30, 2022 12:32:40 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Ungroup and reiterate as step s14
Apr 30, 2022 12:32:40 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect total bytes as step s15
Apr 30, 2022 12:32:40 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect end time metrics as step s16
Apr 30, 2022 12:32:41 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Dataflow SDK version: 2.39.0-SNAPSHOT
Apr 30, 2022 12:32:41 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To access the Dataflow monitoring console, please navigate to https://console.cloud.google.com/dataflow/jobs/us-central1/2022-04-30_05_32_41-17059695585751445620?project=apache-beam-testing
Apr 30, 2022 12:32:41 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Submitted job: 2022-04-30_05_32_41-17059695585751445620
Apr 30, 2022 12:32:41 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To cancel the job using the 'gcloud' tool, run:
> gcloud dataflow jobs --project=apache-beam-testing cancel --region=us-central1 2022-04-30_05_32_41-17059695585751445620
Apr 30, 2022 12:32:53 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
WARNING: 2022-04-30T12:32:52.695Z: The workflow name is not a valid Cloud Label. Labels applied to Cloud resources (such as GCE Instances) for monitoring will be labeled with this modified job name: load0tests0java170dataflow0v20streaming0cogbk01-jenkins-04-ltel. For the best monitoring experience, please name your job with a valid Cloud Label. For details, see: https://cloud.google.com/compute/docs/labeling-resources#restrictions
Apr 30, 2022 12:32:59 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-30T12:32:59.118Z: Worker configuration: e2-standard-2 in us-central1-b.
Apr 30, 2022 12:33:01 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-30T12:33:00.085Z: Expanding SplittableParDo operations into optimizable parts.
Apr 30, 2022 12:33:01 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-30T12:33:00.133Z: Expanding CollectionToSingleton operations into optimizable parts.
Apr 30, 2022 12:33:01 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-30T12:33:00.242Z: Expanding CoGroupByKey operations into optimizable parts.
Apr 30, 2022 12:33:01 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-30T12:33:00.391Z: Expanding SplittableProcessKeyed operations into optimizable parts.
Apr 30, 2022 12:33:01 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-30T12:33:00.462Z: Expanding GroupByKey operations into streaming Read/Write steps
Apr 30, 2022 12:33:01 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-30T12:33:00.610Z: Lifting ValueCombiningMappingFns into MergeBucketsMappingFns
Apr 30, 2022 12:33:01 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-30T12:33:00.820Z: Fusing adjacent ParDo, Read, Write, and Flatten operations
Apr 30, 2022 12:33:01 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-30T12:33:00.866Z: Unzipping flatten CoGroupByKey-Flatten for input CoGroupByKey-MakeUnionTable0-ParMultiDo-ConstructUnionTable-.output
Apr 30, 2022 12:33:01 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-30T12:33:00.931Z: Fusing unzipped copy of CoGroupByKey/GBK/WriteStream, through flatten CoGroupByKey/Flatten, into producer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable)
Apr 30, 2022 12:33:01 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-30T12:33:00.975Z: Fusing consumer CoGroupByKey/GBK/WriteStream into CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable)
Apr 30, 2022 12:33:01 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-30T12:33:01.008Z: Fusing consumer Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read input/Impulse
Apr 30, 2022 12:33:01 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-30T12:33:01.075Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
Apr 30, 2022 12:33:01 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-30T12:33:01.139Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
Apr 30, 2022 12:33:01 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-30T12:33:01.179Z: Fusing consumer Read input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
Apr 30, 2022 12:33:01 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-30T12:33:01.240Z: Fusing consumer Collect start time metrics (input)/ParMultiDo(TimeMonitor) into Read input/ParDo(StripIds)/ParMultiDo(StripIds)
Apr 30, 2022 12:33:01 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-30T12:33:01.283Z: Fusing consumer Window.Into()/Window.Assign into Collect start time metrics (input)/ParMultiDo(TimeMonitor)
Apr 30, 2022 12:33:01 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-30T12:33:01.310Z: Fusing consumer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable) into Window.Into()/Window.Assign
Apr 30, 2022 12:33:01 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-30T12:33:01.353Z: Fusing consumer Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read co-input/Impulse
Apr 30, 2022 12:33:01 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-30T12:33:01.400Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
Apr 30, 2022 12:33:01 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-30T12:33:01.461Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
Apr 30, 2022 12:33:01 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-30T12:33:01.498Z: Fusing consumer Read co-input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
Apr 30, 2022 12:33:01 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-30T12:33:01.544Z: Fusing consumer Collect start time metrics (co-input)/ParMultiDo(TimeMonitor) into Read co-input/ParDo(StripIds)/ParMultiDo(StripIds)
Apr 30, 2022 12:33:01 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-30T12:33:01.580Z: Fusing consumer Window.Into()2/Window.Assign into Collect start time metrics (co-input)/ParMultiDo(TimeMonitor)
Apr 30, 2022 12:33:01 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-30T12:33:01.649Z: Fusing consumer CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable) into Window.Into()2/Window.Assign
Apr 30, 2022 12:33:01 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-30T12:33:01.701Z: Fusing consumer CoGroupByKey/GBK/MergeBuckets into CoGroupByKey/GBK/ReadStream
Apr 30, 2022 12:33:01 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-30T12:33:01.764Z: Fusing consumer CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult) into CoGroupByKey/GBK/MergeBuckets
Apr 30, 2022 12:33:03 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-30T12:33:01.804Z: Fusing consumer Ungroup and reiterate/ParMultiDo(UngroupAndReiterate) into CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult)
Apr 30, 2022 12:33:03 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-30T12:33:01.860Z: Fusing consumer Collect total bytes/ParMultiDo(ByteMonitor) into Ungroup and reiterate/ParMultiDo(UngroupAndReiterate)
Apr 30, 2022 12:33:03 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-30T12:33:01.897Z: Fusing consumer Collect end time metrics/ParMultiDo(TimeMonitor) into Collect total bytes/ParMultiDo(ByteMonitor)
Apr 30, 2022 12:33:03 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-30T12:33:02.138Z: Running job using Streaming Engine
Apr 30, 2022 12:33:03 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-30T12:33:02.452Z: Starting 5 ****s in us-central1-b...
Apr 30, 2022 12:33:25 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-30T12:33:25.346Z: Your project already contains 100 Dataflow-created metric descriptors, so new user metrics of the form custom.googleapis.com/* will not be created. However, all user metrics are also available in the metric dataflow.googleapis.com/job/user_counter. If you rely on the custom metrics, you can delete old / unused metric descriptors. See https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.list and https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.delete
Apr 30, 2022 12:33:30 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-30T12:33:29.128Z: Autoscaling: Raised the number of ****s to 4 so that the pipeline can catch up with its backlog and keep up with its input rate.
Apr 30, 2022 12:33:30 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-30T12:33:29.181Z: Resized **** pool to 4, though goal was 5.  This could be a quota issue.
Apr 30, 2022 12:33:40 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-30T12:33:39.423Z: Autoscaling: Raised the number of ****s to 5 so that the pipeline can catch up with its backlog and keep up with its input rate.
Apr 30, 2022 12:34:38 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-30T12:34:37.924Z: Workers have started successfully.
Apr 30, 2022 4:01:00 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-30T16:00:58.753Z: Cancel request is committed for workflow job: 2022-04-30_05_32_41-17059695585751445620.
Apr 30, 2022 4:01:00 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-30T16:00:58.828Z: Cleaning up.
Apr 30, 2022 4:01:00 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-30T16:00:58.936Z: Stopping **** pool...
Apr 30, 2022 4:01:00 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-30T16:00:58.989Z: Stopping **** pool...
Apr 30, 2022 4:01:35 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-30T16:01:34.198Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Apr 30, 2022 4:01:35 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-30T16:01:34.235Z: Worker pool stopped.
Apr 30, 2022 4:01:41 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-04-30_05_32_41-17059695585751445620 finished with status CANCELLED.
Load test results for test (ID): 772bdddf-7839-4d9d-b5d6-a0269f657b3b and timestamp: 2022-04-30T12:32:35.683000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                 12358.036
dataflow_v2_java17_total_bytes_count             3.44843553E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220430123036
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:0e47d17ca7537760a8c2dbd9549153d80ea0d48c42799e5647dcbcee5d1d52fc
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220430123036]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:0e47d17ca7537760a8c2dbd9549153d80ea0d48c42799e5647dcbcee5d1d52fc]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220430123036] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:0e47d17ca7537760a8c2dbd9549153d80ea0d48c42799e5647dcbcee5d1d52fc])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:0e47d17ca7537760a8c2dbd9549153d80ea0d48c42799e5647dcbcee5d1d52fc
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:0e47d17ca7537760a8c2dbd9549153d80ea0d48c42799e5647dcbcee5d1d52fc
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:0e47d17ca7537760a8c2dbd9549153d80ea0d48c42799e5647dcbcee5d1d52fc].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.4/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 31m 25s
110 actionable tasks: 74 executed, 32 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/ey4p62jcnji2q

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #130

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/130/display/redirect?page=changes>

Changes:

[ihr] Update Java katas to Beam 2.38

[noreply] [BEAM-14369] Fix "target/options: no such file or directory" error while

[noreply] [BEAM-14297] Enable nullable key and value arrays for xlang kafka io

[noreply] Merge pull request #17444 from [BEAM-14310] [Website] bug home

[noreply] Merge pull request #17388 from [BEAM-14311] [Website] Home Page

[noreply] [BEAM-14376] Typo in method description doc

[noreply] Add default classpath when not present (#17491)

[thiagotnunes] fix: update javadocs for ChangeStreamMetrics

[noreply] Merge pull request #17443 from [BEAM-12164]: use the end timestamp for

[noreply] Merge pull request #17260 from [BEAM-13229] [Website] bug side nav

[noreply] [BEAM-14351] Fix the template and move the announcement to the next


------------------------------------------
[...truncated 50.05 KB...]
1a6c66a7e504: Preparing
9f2fba01df7f: Preparing
939d4a0c2578: Preparing
10bfe1d713b4: Preparing
3bc383470c05: Preparing
e93827457889: Preparing
08fa02ce37eb: Preparing
a037458de4e0: Preparing
bafdbe68e4ae: Preparing
a13c519c6361: Preparing
e93827457889: Waiting
3bc383470c05: Waiting
08fa02ce37eb: Waiting
1a6c66a7e504: Waiting
a037458de4e0: Waiting
9f2fba01df7f: Waiting
bafdbe68e4ae: Waiting
939d4a0c2578: Waiting
a13c519c6361: Waiting
53913aed19d2: Waiting
10bfe1d713b4: Waiting
7a763751f65a: Waiting
f30932317082: Waiting
c3f9929bdbbf: Waiting
89905dd1bd7f: Pushed
4e895253da30: Pushed
df122f2a33c4: Pushed
ad33fb1004d8: Pushed
f7689c509955: Pushed
53913aed19d2: Pushed
f30932317082: Pushed
e2ecd49ff6ec: Pushed
1a6c66a7e504: Pushed
86aa7f9706f2: Pushed
7a763751f65a: Pushed
3bc383470c05: Layer already exists
c3f9929bdbbf: Pushed
e93827457889: Layer already exists
08fa02ce37eb: Layer already exists
a037458de4e0: Layer already exists
bafdbe68e4ae: Layer already exists
a13c519c6361: Layer already exists
939d4a0c2578: Pushed
10bfe1d713b4: Pushed
9f2fba01df7f: Pushed
20220429123038: digest: sha256:0eff4209fd3bba2527bb7ad5bf3d3235bf96dd181813188181b9303e3ea58608 size: 4729

> Task :sdks:java:testing:load-tests:run
Apr 29, 2022 12:32:38 PM org.apache.beam.runners.dataflow.options.DataflowPipelineOptions$StagingLocationFactory create
INFO: No stagingLocation provided, falling back to gcpTempLocation
Apr 29, 2022 12:32:39 PM org.apache.beam.runners.dataflow.DataflowRunner fromOptions
INFO: PipelineOptions.filesToStage was not specified. Defaulting to files from the classpath: will stage 208 files. Enable logging at DEBUG level to see which files will be staged.
Apr 29, 2022 12:32:39 PM org.apache.beam.sdk.Pipeline validate
WARNING: The following transforms do not have stable unique names: Window.Into()
Apr 29, 2022 12:32:39 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Executing pipeline on the Dataflow Service, which will have billing implications related to Google Compute Engine usage and other Google Cloud Services.
Apr 29, 2022 12:32:42 PM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Uploading 208 files from PipelineOptions.filesToStage to staging location to prepare for execution.
Apr 29, 2022 12:32:42 PM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Staging files complete: 208 files cached, 0 files newly uploaded in 0 seconds
Apr 29, 2022 12:32:42 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Staging portable pipeline proto to gs://temp-storage-for-perf-tests/loadtests/staging/
Apr 29, 2022 12:32:42 PM org.apache.beam.runners.dataflow.util.PackageUtil tryStagePackage
INFO: Uploading <115405 bytes, hash b28ad4e8e1c0fbba6f24af826606c4176dfe00c843d23c387855f37925ad9698> to gs://temp-storage-for-perf-tests/loadtests/staging/pipeline-sorU6OHA-7pvJK-CZgbEF23-AMhD0jw4eFXzeSWtlpg.pb
Apr 29, 2022 12:32:44 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read input/DataflowRunner.StreamingUnboundedRead.ReadWithIds as step s1
Apr 29, 2022 12:32:44 PM org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource split
INFO: Split into 20 bundles of sizes: [org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@57272109, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@59696551, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@648d0e6d, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@79e66b2f, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@17273273, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5f69e2b, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@984169e, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@43f1bb92, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6d6bbd35, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5c5d6175, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7544ac86, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3b27b497, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@b1534d3, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3c74aa0d, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6c841199, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6a818392, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@489091bd, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@512d6e60, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1de9b505, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7b122839]
Apr 29, 2022 12:32:44 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read input/StripIds as step s2
Apr 29, 2022 12:32:44 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect start time metrics (input) as step s3
Apr 29, 2022 12:32:44 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Window.Into()/Window.Assign as step s4
Apr 29, 2022 12:32:44 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read co-input/DataflowRunner.StreamingUnboundedRead.ReadWithIds as step s5
Apr 29, 2022 12:32:44 PM org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource split
INFO: Split into 20 bundles of sizes: [org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6a6f6c7e, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6c5ddccd, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1dbd580, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6c101cc1, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7d0d91a1, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7fb48179, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@201c3cda, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4c86da0c, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5d97caa4, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6732726, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@474821de, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3d64c581, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5ec5ea63, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4190bc8a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@47d023b7, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5c83ae01, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2d64c100, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@69d45cca, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2fdf17dc, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6e6d4780]
Apr 29, 2022 12:32:44 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read co-input/StripIds as step s6
Apr 29, 2022 12:32:44 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect start time metrics (co-input) as step s7
Apr 29, 2022 12:32:44 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Window.Into()2/Window.Assign as step s8
Apr 29, 2022 12:32:44 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/MakeUnionTable0 as step s9
Apr 29, 2022 12:32:44 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/MakeUnionTable1 as step s10
Apr 29, 2022 12:32:44 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/Flatten as step s11
Apr 29, 2022 12:32:44 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/GBK as step s12
Apr 29, 2022 12:32:44 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/ConstructCoGbkResultFn as step s13
Apr 29, 2022 12:32:44 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Ungroup and reiterate as step s14
Apr 29, 2022 12:32:44 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect total bytes as step s15
Apr 29, 2022 12:32:44 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect end time metrics as step s16
Apr 29, 2022 12:32:44 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Dataflow SDK version: 2.39.0-SNAPSHOT
Apr 29, 2022 12:32:45 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To access the Dataflow monitoring console, please navigate to https://console.cloud.google.com/dataflow/jobs/us-central1/2022-04-29_05_32_44-9382013507233506701?project=apache-beam-testing
Apr 29, 2022 12:32:45 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Submitted job: 2022-04-29_05_32_44-9382013507233506701
Apr 29, 2022 12:32:45 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To cancel the job using the 'gcloud' tool, run:
> gcloud dataflow jobs --project=apache-beam-testing cancel --region=us-central1 2022-04-29_05_32_44-9382013507233506701
Apr 29, 2022 12:32:53 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
WARNING: 2022-04-29T12:32:50.482Z: The workflow name is not a valid Cloud Label. Labels applied to Cloud resources (such as GCE Instances) for monitoring will be labeled with this modified job name: load0tests0java170dataflow0v20streaming0cogbk01-jenkins-04-3d59. For the best monitoring experience, please name your job with a valid Cloud Label. For details, see: https://cloud.google.com/compute/docs/labeling-resources#restrictions
Apr 29, 2022 12:32:57 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-29T12:32:57.076Z: Worker configuration: e2-standard-2 in us-central1-b.
Apr 29, 2022 12:32:57 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-29T12:32:57.618Z: Expanding SplittableParDo operations into optimizable parts.
Apr 29, 2022 12:32:57 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-29T12:32:57.643Z: Expanding CollectionToSingleton operations into optimizable parts.
Apr 29, 2022 12:32:57 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-29T12:32:57.715Z: Expanding CoGroupByKey operations into optimizable parts.
Apr 29, 2022 12:32:59 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-29T12:32:57.776Z: Expanding SplittableProcessKeyed operations into optimizable parts.
Apr 29, 2022 12:32:59 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-29T12:32:57.804Z: Expanding GroupByKey operations into streaming Read/Write steps
Apr 29, 2022 12:32:59 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-29T12:32:57.852Z: Lifting ValueCombiningMappingFns into MergeBucketsMappingFns
Apr 29, 2022 12:32:59 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-29T12:32:57.973Z: Fusing adjacent ParDo, Read, Write, and Flatten operations
Apr 29, 2022 12:32:59 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-29T12:32:58.027Z: Unzipping flatten CoGroupByKey-Flatten for input CoGroupByKey-MakeUnionTable0-ParMultiDo-ConstructUnionTable-.output
Apr 29, 2022 12:32:59 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-29T12:32:58.056Z: Fusing unzipped copy of CoGroupByKey/GBK/WriteStream, through flatten CoGroupByKey/Flatten, into producer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable)
Apr 29, 2022 12:32:59 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-29T12:32:58.085Z: Fusing consumer CoGroupByKey/GBK/WriteStream into CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable)
Apr 29, 2022 12:32:59 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-29T12:32:58.109Z: Fusing consumer Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read input/Impulse
Apr 29, 2022 12:32:59 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-29T12:32:58.135Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
Apr 29, 2022 12:32:59 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-29T12:32:58.169Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
Apr 29, 2022 12:32:59 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-29T12:32:58.192Z: Fusing consumer Read input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
Apr 29, 2022 12:32:59 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-29T12:32:58.215Z: Fusing consumer Collect start time metrics (input)/ParMultiDo(TimeMonitor) into Read input/ParDo(StripIds)/ParMultiDo(StripIds)
Apr 29, 2022 12:32:59 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-29T12:32:58.237Z: Fusing consumer Window.Into()/Window.Assign into Collect start time metrics (input)/ParMultiDo(TimeMonitor)
Apr 29, 2022 12:32:59 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-29T12:32:58.261Z: Fusing consumer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable) into Window.Into()/Window.Assign
Apr 29, 2022 12:32:59 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-29T12:32:58.304Z: Fusing consumer Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read co-input/Impulse
Apr 29, 2022 12:32:59 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-29T12:32:58.336Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
Apr 29, 2022 12:32:59 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-29T12:32:58.360Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
Apr 29, 2022 12:32:59 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-29T12:32:58.388Z: Fusing consumer Read co-input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
Apr 29, 2022 12:32:59 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-29T12:32:58.424Z: Fusing consumer Collect start time metrics (co-input)/ParMultiDo(TimeMonitor) into Read co-input/ParDo(StripIds)/ParMultiDo(StripIds)
Apr 29, 2022 12:32:59 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-29T12:32:58.447Z: Fusing consumer Window.Into()2/Window.Assign into Collect start time metrics (co-input)/ParMultiDo(TimeMonitor)
Apr 29, 2022 12:32:59 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-29T12:32:58.481Z: Fusing consumer CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable) into Window.Into()2/Window.Assign
Apr 29, 2022 12:32:59 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-29T12:32:58.507Z: Fusing consumer CoGroupByKey/GBK/MergeBuckets into CoGroupByKey/GBK/ReadStream
Apr 29, 2022 12:32:59 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-29T12:32:58.549Z: Fusing consumer CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult) into CoGroupByKey/GBK/MergeBuckets
Apr 29, 2022 12:32:59 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-29T12:32:58.578Z: Fusing consumer Ungroup and reiterate/ParMultiDo(UngroupAndReiterate) into CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult)
Apr 29, 2022 12:32:59 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-29T12:32:58.634Z: Fusing consumer Collect total bytes/ParMultiDo(ByteMonitor) into Ungroup and reiterate/ParMultiDo(UngroupAndReiterate)
Apr 29, 2022 12:32:59 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-29T12:32:58.664Z: Fusing consumer Collect end time metrics/ParMultiDo(TimeMonitor) into Collect total bytes/ParMultiDo(ByteMonitor)
Apr 29, 2022 12:32:59 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-29T12:32:58.834Z: Running job using Streaming Engine
Apr 29, 2022 12:32:59 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-29T12:32:59.055Z: Starting 5 ****s in us-central1-b...
Apr 29, 2022 12:33:08 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-29T12:33:07.776Z: Your project already contains 100 Dataflow-created metric descriptors, so new user metrics of the form custom.googleapis.com/* will not be created. However, all user metrics are also available in the metric dataflow.googleapis.com/job/user_counter. If you rely on the custom metrics, you can delete old / unused metric descriptors. See https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.list and https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.delete
Apr 29, 2022 12:33:34 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-29T12:33:31.754Z: Autoscaling: Raised the number of ****s to 5 so that the pipeline can catch up with its backlog and keep up with its input rate.
Apr 29, 2022 12:34:37 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-29T12:34:35.673Z: Workers have started successfully.
Apr 29, 2022 4:01:04 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-29T16:01:02.934Z: Cancel request is committed for workflow job: 2022-04-29_05_32_44-9382013507233506701.
Apr 29, 2022 4:01:04 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-29T16:01:03.019Z: Cleaning up.
Apr 29, 2022 4:01:04 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-29T16:01:03.114Z: Stopping **** pool...
Apr 29, 2022 4:01:04 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-29T16:01:03.157Z: Stopping **** pool...
Apr 29, 2022 4:01:35 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-29T16:01:35.807Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Apr 29, 2022 4:01:35 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-29T16:01:35.840Z: Worker pool stopped.
Apr 29, 2022 4:01:45 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-04-29_05_32_44-9382013507233506701 finished with status CANCELLED.
Load test results for test (ID): 47ad33e9-9c96-4a7d-a563-f53a45509e9b and timestamp: 2022-04-29T12:32:39.443000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                  12346.36
dataflow_v2_java17_total_bytes_count              3.8917995E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220429123038
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:0eff4209fd3bba2527bb7ad5bf3d3235bf96dd181813188181b9303e3ea58608
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220429123038]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:0eff4209fd3bba2527bb7ad5bf3d3235bf96dd181813188181b9303e3ea58608]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220429123038] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:0eff4209fd3bba2527bb7ad5bf3d3235bf96dd181813188181b9303e3ea58608])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:0eff4209fd3bba2527bb7ad5bf3d3235bf96dd181813188181b9303e3ea58608
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:0eff4209fd3bba2527bb7ad5bf3d3235bf96dd181813188181b9303e3ea58608
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:0eff4209fd3bba2527bb7ad5bf3d3235bf96dd181813188181b9303e3ea58608].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.4/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 32m 11s
110 actionable tasks: 73 executed, 33 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/j7urcfn26q5km

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #129

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/129/display/redirect?page=changes>

Changes:

[noreply] [BEAM-11104] Add Checkpointing split to Go SDK (#17386)

[noreply] Merge pull request #17226 from [BEAM-14204] [Playground] Tests for

[noreply] [BEAM-13015, BEAM-14184] Address unbounded number of messages being

[noreply] Improvement to Seed job configuration to launch against PRs (#17468)

[noreply] [BEAM-13983] Small changes to sklearn runinference (#17459)

[chamikaramj] Renames ExternalPythonTransform to PythonExternalTransform

[noreply] [BEAM-14351] Inherit from Coder. (#17437)


------------------------------------------
[...truncated 46.64 KB...]
022c02dc48eb: Pushed
d1fb25c57111: Pushed
286332b2b2d6: Pushed
0ed3aae1b811: Pushed
61b45dd0fa4a: Pushed
2e38b96b808c: Pushed
7aaa05e8eee5: Pushed
8d69c2f825d4: Pushed
3bc383470c05: Layer already exists
e93827457889: Layer already exists
e394883f3a23: Pushed
08fa02ce37eb: Layer already exists
a037458de4e0: Layer already exists
e3c4e9c86913: Pushed
bafdbe68e4ae: Layer already exists
a13c519c6361: Layer already exists
c2e2f9b1b251: Pushed
1d53b7ecd1e9: Pushed
20220428123039: digest: sha256:e150bab5afd54e7eb0a9e2e9d2379ba06eab09426d968d7a5ad2e2b44cfcd34c size: 4729

> Task :sdks:java:testing:load-tests:run
Apr 28, 2022 12:32:26 PM org.apache.beam.runners.dataflow.options.DataflowPipelineOptions$StagingLocationFactory create
INFO: No stagingLocation provided, falling back to gcpTempLocation
Apr 28, 2022 12:32:26 PM org.apache.beam.runners.dataflow.DataflowRunner fromOptions
INFO: PipelineOptions.filesToStage was not specified. Defaulting to files from the classpath: will stage 208 files. Enable logging at DEBUG level to see which files will be staged.
Apr 28, 2022 12:32:27 PM org.apache.beam.sdk.Pipeline validate
WARNING: The following transforms do not have stable unique names: Window.Into()
Apr 28, 2022 12:32:27 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Executing pipeline on the Dataflow Service, which will have billing implications related to Google Compute Engine usage and other Google Cloud Services.
Apr 28, 2022 12:32:29 PM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Uploading 208 files from PipelineOptions.filesToStage to staging location to prepare for execution.
Apr 28, 2022 12:32:30 PM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Staging files complete: 208 files cached, 0 files newly uploaded in 0 seconds
Apr 28, 2022 12:32:30 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Staging portable pipeline proto to gs://temp-storage-for-perf-tests/loadtests/staging/
Apr 28, 2022 12:32:30 PM org.apache.beam.runners.dataflow.util.PackageUtil tryStagePackage
INFO: Uploading <115405 bytes, hash 447bac8dad8fd3432a03289c1dcc4ee41d1147d664d6de8d48db944d17c2003b> to gs://temp-storage-for-perf-tests/loadtests/staging/pipeline-RHusja2P00MqAyicHcxO5B0RR9Zk1t6NSNuUTRfCADs.pb
Apr 28, 2022 12:32:32 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read input/DataflowRunner.StreamingUnboundedRead.ReadWithIds as step s1
Apr 28, 2022 12:32:32 PM org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource split
INFO: Split into 20 bundles of sizes: [org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@57272109, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@59696551, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@648d0e6d, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@79e66b2f, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@17273273, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5f69e2b, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@984169e, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@43f1bb92, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6d6bbd35, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5c5d6175, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7544ac86, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3b27b497, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@b1534d3, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3c74aa0d, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6c841199, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6a818392, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@489091bd, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@512d6e60, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1de9b505, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7b122839]
Apr 28, 2022 12:32:32 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read input/StripIds as step s2
Apr 28, 2022 12:32:32 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect start time metrics (input) as step s3
Apr 28, 2022 12:32:32 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Window.Into()/Window.Assign as step s4
Apr 28, 2022 12:32:32 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read co-input/DataflowRunner.StreamingUnboundedRead.ReadWithIds as step s5
Apr 28, 2022 12:32:32 PM org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource split
INFO: Split into 20 bundles of sizes: [org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6a6f6c7e, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6c5ddccd, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1dbd580, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6c101cc1, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7d0d91a1, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7fb48179, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@201c3cda, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4c86da0c, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5d97caa4, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6732726, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@474821de, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3d64c581, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5ec5ea63, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4190bc8a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@47d023b7, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5c83ae01, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2d64c100, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@69d45cca, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2fdf17dc, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6e6d4780]
Apr 28, 2022 12:32:32 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read co-input/StripIds as step s6
Apr 28, 2022 12:32:32 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect start time metrics (co-input) as step s7
Apr 28, 2022 12:32:32 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Window.Into()2/Window.Assign as step s8
Apr 28, 2022 12:32:32 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/MakeUnionTable0 as step s9
Apr 28, 2022 12:32:32 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/MakeUnionTable1 as step s10
Apr 28, 2022 12:32:32 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/Flatten as step s11
Apr 28, 2022 12:32:32 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/GBK as step s12
Apr 28, 2022 12:32:32 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/ConstructCoGbkResultFn as step s13
Apr 28, 2022 12:32:32 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Ungroup and reiterate as step s14
Apr 28, 2022 12:32:32 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect total bytes as step s15
Apr 28, 2022 12:32:32 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect end time metrics as step s16
Apr 28, 2022 12:32:32 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Dataflow SDK version: 2.39.0-SNAPSHOT
Apr 28, 2022 12:32:32 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To access the Dataflow monitoring console, please navigate to https://console.cloud.google.com/dataflow/jobs/us-central1/2022-04-28_05_32_32-10235830869270665189?project=apache-beam-testing
Apr 28, 2022 12:32:32 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Submitted job: 2022-04-28_05_32_32-10235830869270665189
Apr 28, 2022 12:32:32 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To cancel the job using the 'gcloud' tool, run:
> gcloud dataflow jobs --project=apache-beam-testing cancel --region=us-central1 2022-04-28_05_32_32-10235830869270665189
Apr 28, 2022 12:32:38 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
WARNING: 2022-04-28T12:32:38.272Z: The workflow name is not a valid Cloud Label. Labels applied to Cloud resources (such as GCE Instances) for monitoring will be labeled with this modified job name: load0tests0java170dataflow0v20streaming0cogbk01-jenkins-04-t8zh. For the best monitoring experience, please name your job with a valid Cloud Label. For details, see: https://cloud.google.com/compute/docs/labeling-resources#restrictions
Apr 28, 2022 12:32:48 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-28T12:32:46.137Z: Worker configuration: e2-standard-2 in us-central1-f.
Apr 28, 2022 12:32:48 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-28T12:32:46.706Z: Expanding SplittableParDo operations into optimizable parts.
Apr 28, 2022 12:32:48 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-28T12:32:46.743Z: Expanding CollectionToSingleton operations into optimizable parts.
Apr 28, 2022 12:32:48 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-28T12:32:46.837Z: Expanding CoGroupByKey operations into optimizable parts.
Apr 28, 2022 12:32:48 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-28T12:32:46.922Z: Expanding SplittableProcessKeyed operations into optimizable parts.
Apr 28, 2022 12:32:48 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-28T12:32:46.958Z: Expanding GroupByKey operations into streaming Read/Write steps
Apr 28, 2022 12:32:48 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-28T12:32:47.022Z: Lifting ValueCombiningMappingFns into MergeBucketsMappingFns
Apr 28, 2022 12:32:48 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-28T12:32:47.130Z: Fusing adjacent ParDo, Read, Write, and Flatten operations
Apr 28, 2022 12:32:48 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-28T12:32:47.155Z: Unzipping flatten CoGroupByKey-Flatten for input CoGroupByKey-MakeUnionTable0-ParMultiDo-ConstructUnionTable-.output
Apr 28, 2022 12:32:48 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-28T12:32:47.190Z: Fusing unzipped copy of CoGroupByKey/GBK/WriteStream, through flatten CoGroupByKey/Flatten, into producer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable)
Apr 28, 2022 12:32:48 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-28T12:32:47.211Z: Fusing consumer CoGroupByKey/GBK/WriteStream into CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable)
Apr 28, 2022 12:32:48 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-28T12:32:47.233Z: Fusing consumer Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read input/Impulse
Apr 28, 2022 12:32:48 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-28T12:32:47.268Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
Apr 28, 2022 12:32:48 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-28T12:32:47.300Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
Apr 28, 2022 12:32:48 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-28T12:32:47.334Z: Fusing consumer Read input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
Apr 28, 2022 12:32:48 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-28T12:32:47.366Z: Fusing consumer Collect start time metrics (input)/ParMultiDo(TimeMonitor) into Read input/ParDo(StripIds)/ParMultiDo(StripIds)
Apr 28, 2022 12:32:48 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-28T12:32:47.400Z: Fusing consumer Window.Into()/Window.Assign into Collect start time metrics (input)/ParMultiDo(TimeMonitor)
Apr 28, 2022 12:32:48 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-28T12:32:47.440Z: Fusing consumer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable) into Window.Into()/Window.Assign
Apr 28, 2022 12:32:48 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-28T12:32:47.468Z: Fusing consumer Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read co-input/Impulse
Apr 28, 2022 12:32:48 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-28T12:32:47.506Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
Apr 28, 2022 12:32:48 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-28T12:32:47.538Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
Apr 28, 2022 12:32:48 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-28T12:32:47.568Z: Fusing consumer Read co-input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
Apr 28, 2022 12:32:48 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-28T12:32:47.600Z: Fusing consumer Collect start time metrics (co-input)/ParMultiDo(TimeMonitor) into Read co-input/ParDo(StripIds)/ParMultiDo(StripIds)
Apr 28, 2022 12:32:48 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-28T12:32:47.632Z: Fusing consumer Window.Into()2/Window.Assign into Collect start time metrics (co-input)/ParMultiDo(TimeMonitor)
Apr 28, 2022 12:32:48 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-28T12:32:47.669Z: Fusing consumer CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable) into Window.Into()2/Window.Assign
Apr 28, 2022 12:32:48 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-28T12:32:47.701Z: Fusing consumer CoGroupByKey/GBK/MergeBuckets into CoGroupByKey/GBK/ReadStream
Apr 28, 2022 12:32:48 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-28T12:32:47.723Z: Fusing consumer CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult) into CoGroupByKey/GBK/MergeBuckets
Apr 28, 2022 12:32:48 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-28T12:32:47.755Z: Fusing consumer Ungroup and reiterate/ParMultiDo(UngroupAndReiterate) into CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult)
Apr 28, 2022 12:32:48 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-28T12:32:47.777Z: Fusing consumer Collect total bytes/ParMultiDo(ByteMonitor) into Ungroup and reiterate/ParMultiDo(UngroupAndReiterate)
Apr 28, 2022 12:32:48 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-28T12:32:47.824Z: Fusing consumer Collect end time metrics/ParMultiDo(TimeMonitor) into Collect total bytes/ParMultiDo(ByteMonitor)
Apr 28, 2022 12:32:48 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-28T12:32:48.025Z: Running job using Streaming Engine
Apr 28, 2022 12:32:48 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-28T12:32:48.320Z: Starting 5 ****s in us-central1-f...
Apr 28, 2022 12:33:10 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-28T12:33:09.513Z: Your project already contains 100 Dataflow-created metric descriptors, so new user metrics of the form custom.googleapis.com/* will not be created. However, all user metrics are also available in the metric dataflow.googleapis.com/job/user_counter. If you rely on the custom metrics, you can delete old / unused metric descriptors. See https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.list and https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.delete
Apr 28, 2022 12:33:11 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-28T12:33:10.851Z: Autoscaling: Raised the number of ****s to 5 so that the pipeline can catch up with its backlog and keep up with its input rate.
Apr 28, 2022 12:34:19 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-28T12:34:18.153Z: Workers have started successfully.
Apr 28, 2022 1:17:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
SEVERE: 2022-04-28T13:17:50.263Z: Staged package gax-grpc-2.8.1-Oaqma5KQUeKekzhWvuQ9BCsF3Uf1NF_DCF0Um3fxw3s.jar at location 'gs://temp-storage-for-perf-tests/loadtests/staging/gax-grpc-2.8.1-Oaqma5KQUeKekzhWvuQ9BCsF3Uf1NF_DCF0Um3fxw3s.jar' is inaccessible.
Apr 28, 2022 1:17:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
WARNING: 2022-04-28T13:17:56.765Z: One or more access checks for temp location or staged files failed. Please refer to other error messages for details. For more information on security and permissions, please see https://cloud.google.com/dataflow/security-and-permissions.
Apr 28, 2022 1:20:55 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
WARNING: 2022-04-28T13:20:52.937Z: One or more access checks for temp location or staged files failed. Please refer to other error messages for details. For more information on security and permissions, please see https://cloud.google.com/dataflow/security-and-permissions.
Apr 28, 2022 4:01:05 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-28T16:01:05.061Z: Cancel request is committed for workflow job: 2022-04-28_05_32_32-10235830869270665189.
Apr 28, 2022 4:01:05 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-28T16:01:05.121Z: Cleaning up.
Apr 28, 2022 4:01:05 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-28T16:01:05.211Z: Stopping **** pool...
Apr 28, 2022 4:01:05 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-28T16:01:05.279Z: Stopping **** pool...
Apr 28, 2022 4:01:39 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-28T16:01:38.538Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Apr 28, 2022 4:01:39 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-28T16:01:38.609Z: Worker pool stopped.
Apr 28, 2022 4:01:56 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-04-28_05_32_32-10235830869270665189 finished with status CANCELLED.
Load test results for test (ID): 902d45ce-f2c6-48bd-aa04-223c5fdbd611 and timestamp: 2022-04-28T12:32:27.273000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                 12356.592
dataflow_v2_java17_total_bytes_count              4.5381291E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220428123039
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:e150bab5afd54e7eb0a9e2e9d2379ba06eab09426d968d7a5ad2e2b44cfcd34c
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220428123039]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:e150bab5afd54e7eb0a9e2e9d2379ba06eab09426d968d7a5ad2e2b44cfcd34c]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220428123039] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:e150bab5afd54e7eb0a9e2e9d2379ba06eab09426d968d7a5ad2e2b44cfcd34c])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:1fc998e42139727d2c3786f1de49786ca539cf13c81d83afa97ab6cf29387608
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:1fc998e42139727d2c3786f1de49786ca539cf13c81d83afa97ab6cf29387608
ERROR: (gcloud.container.images.delete) Not found: response: {'docker-distribution-api-version': 'registry/2.0', 'content-type': 'application/json', 'date': 'Thu, 28 Apr 2022 16:02:03 GMT', 'server': 'Docker Registry', 'cache-control': 'private', 'x-xss-protection': '0', 'x-frame-options': 'SAMEORIGIN', 'transfer-encoding': 'chunked', 'status': '404', 'content-length': '168', '-content-encoding': 'gzip'}
Failed to compute blob liveness for manifest: 'sha256:1fc998e42139727d2c3786f1de49786ca539cf13c81d83afa97ab6cf29387608': None

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages FAILED

FAILURE: Build completed with 2 failures.

1: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.
==============================================================================

2: Task failed with an exception.
-----------
* Where:
Build file '<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/runners/google-cloud-dataflow-java/build.gradle'> line: 297

* What went wrong:
Execution failed for task ':runners:google-cloud-dataflow-java:cleanUpDockerJavaImages'.
> Process 'command './scripts/cleanup_untagged_gcr_images.sh'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.
==============================================================================

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.4/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 31m 38s
110 actionable tasks: 73 executed, 33 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/23fok3zo25qsc

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #128

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/128/display/redirect?page=changes>

Changes:

[msbukal] FhirIO: use .search() or .searchType instead of .setResourceType()

[nick.caballero] [BEAM-14363] Fixes WatermarkParameters builder for Kinesis

[noreply] Remove unnecessary decorator from RunInference interface (#17463)

[noreply] [BEAM-13590] Minor deprecated warning fix (#17453)

[noreply] [BEAM-12164]: fix the negative throughput issue (#17461)

[noreply] Updated goldens for the screen diff integration tests (#17467)

[noreply] fixes copy by value error for bytes.Buffer in Error (#17469)

[noreply] Merge pull request #17354 from [BEAM-14170] - Create a test that runs

[noreply] Merge pull request #17447 from [BEAM-14357] Fix

[noreply] [BEAM-14324, BEAM-14325] Staticcheck cleanup in test files (#17393)

[noreply] BEAM-14187 Fix NPE (#17454)

[noreply] [BEAM-11105] Stateful watermark estimation (#17374)

[noreply] [BEAM-14304] implement parquetio to read/write parquet files (#17347)


------------------------------------------
[...truncated 49.72 KB...]
08fa02ce37eb: Preparing
a037458de4e0: Preparing
bafdbe68e4ae: Preparing
a13c519c6361: Preparing
7b9a4c020e68: Waiting
964154449ba6: Waiting
3fe9d00d75d7: Waiting
7a1c1c9d079b: Waiting
44c9846e824b: Waiting
5e5f3000d6ff: Waiting
784049eb671f: Waiting
12e1940c6229: Waiting
08fa02ce37eb: Waiting
a13c519c6361: Waiting
34461a8d10d3: Waiting
bafdbe68e4ae: Waiting
a037458de4e0: Waiting
e93827457889: Waiting
3bc383470c05: Waiting
4a6eae1d28c6: Waiting
d89f02fa017d: Pushed
98c5441bf5ed: Pushed
48aaa9aaf2dd: Pushed
c2b2d9642fa7: Pushed
9cb62a2ef04b: Pushed
964154449ba6: Pushed
3fe9d00d75d7: Pushed
12e1940c6229: Pushed
7b9a4c020e68: Pushed
44c9846e824b: Pushed
5e5f3000d6ff: Pushed
3bc383470c05: Layer already exists
7a1c1c9d079b: Pushed
e93827457889: Layer already exists
08fa02ce37eb: Layer already exists
a037458de4e0: Layer already exists
a13c519c6361: Layer already exists
bafdbe68e4ae: Layer already exists
34461a8d10d3: Pushed
4a6eae1d28c6: Pushed
784049eb671f: Pushed
20220427123044: digest: sha256:663e4583e395c5e749b4eb8afd1c9c87ebcf9081e05a13c047a84b1c639c52fd size: 4729

> Task :sdks:java:testing:load-tests:run
Apr 27, 2022 12:32:46 PM org.apache.beam.runners.dataflow.options.DataflowPipelineOptions$StagingLocationFactory create
INFO: No stagingLocation provided, falling back to gcpTempLocation
Apr 27, 2022 12:32:46 PM org.apache.beam.runners.dataflow.DataflowRunner fromOptions
INFO: PipelineOptions.filesToStage was not specified. Defaulting to files from the classpath: will stage 208 files. Enable logging at DEBUG level to see which files will be staged.
Apr 27, 2022 12:32:47 PM org.apache.beam.sdk.Pipeline validate
WARNING: The following transforms do not have stable unique names: Window.Into()
Apr 27, 2022 12:32:47 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Executing pipeline on the Dataflow Service, which will have billing implications related to Google Compute Engine usage and other Google Cloud Services.
Apr 27, 2022 12:32:49 PM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Uploading 208 files from PipelineOptions.filesToStage to staging location to prepare for execution.
Apr 27, 2022 12:32:50 PM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Staging files complete: 208 files cached, 0 files newly uploaded in 0 seconds
Apr 27, 2022 12:32:50 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Staging portable pipeline proto to gs://temp-storage-for-perf-tests/loadtests/staging/
Apr 27, 2022 12:32:50 PM org.apache.beam.runners.dataflow.util.PackageUtil tryStagePackage
INFO: Uploading <115405 bytes, hash a52582fcb6e4e90b59c0310f4221259ea839fc486655d0e5411dd50e9750d274> to gs://temp-storage-for-perf-tests/loadtests/staging/pipeline-pSWC_Lbk6QtZwDEPQiElnqg5_EhmVdDlQR3VDpdQ0nQ.pb
Apr 27, 2022 12:32:51 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read input/DataflowRunner.StreamingUnboundedRead.ReadWithIds as step s1
Apr 27, 2022 12:32:51 PM org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource split
INFO: Split into 20 bundles of sizes: [org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@57272109, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@59696551, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@648d0e6d, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@79e66b2f, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@17273273, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5f69e2b, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@984169e, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@43f1bb92, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6d6bbd35, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5c5d6175, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7544ac86, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3b27b497, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@b1534d3, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3c74aa0d, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6c841199, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6a818392, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@489091bd, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@512d6e60, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1de9b505, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7b122839]
Apr 27, 2022 12:32:52 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read input/StripIds as step s2
Apr 27, 2022 12:32:52 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect start time metrics (input) as step s3
Apr 27, 2022 12:32:52 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Window.Into()/Window.Assign as step s4
Apr 27, 2022 12:32:52 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read co-input/DataflowRunner.StreamingUnboundedRead.ReadWithIds as step s5
Apr 27, 2022 12:32:52 PM org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource split
INFO: Split into 20 bundles of sizes: [org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6a6f6c7e, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6c5ddccd, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1dbd580, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6c101cc1, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7d0d91a1, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7fb48179, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@201c3cda, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4c86da0c, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5d97caa4, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6732726, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@474821de, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3d64c581, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5ec5ea63, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4190bc8a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@47d023b7, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5c83ae01, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2d64c100, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@69d45cca, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2fdf17dc, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6e6d4780]
Apr 27, 2022 12:32:52 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read co-input/StripIds as step s6
Apr 27, 2022 12:32:52 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect start time metrics (co-input) as step s7
Apr 27, 2022 12:32:52 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Window.Into()2/Window.Assign as step s8
Apr 27, 2022 12:32:52 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/MakeUnionTable0 as step s9
Apr 27, 2022 12:32:52 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/MakeUnionTable1 as step s10
Apr 27, 2022 12:32:52 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/Flatten as step s11
Apr 27, 2022 12:32:52 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/GBK as step s12
Apr 27, 2022 12:32:52 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/ConstructCoGbkResultFn as step s13
Apr 27, 2022 12:32:52 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Ungroup and reiterate as step s14
Apr 27, 2022 12:32:52 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect total bytes as step s15
Apr 27, 2022 12:32:52 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect end time metrics as step s16
Apr 27, 2022 12:32:52 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Dataflow SDK version: 2.39.0-SNAPSHOT
Apr 27, 2022 12:32:55 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To access the Dataflow monitoring console, please navigate to https://console.cloud.google.com/dataflow/jobs/us-central1/2022-04-27_05_32_52-13783208352401644293?project=apache-beam-testing
Apr 27, 2022 12:32:55 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Submitted job: 2022-04-27_05_32_52-13783208352401644293
Apr 27, 2022 12:32:55 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To cancel the job using the 'gcloud' tool, run:
> gcloud dataflow jobs --project=apache-beam-testing cancel --region=us-central1 2022-04-27_05_32_52-13783208352401644293
Apr 27, 2022 12:33:01 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
WARNING: 2022-04-27T12:33:00.309Z: The workflow name is not a valid Cloud Label. Labels applied to Cloud resources (such as GCE Instances) for monitoring will be labeled with this modified job name: load0tests0java170dataflow0v20streaming0cogbk01-jenkins-04-oxw8. For the best monitoring experience, please name your job with a valid Cloud Label. For details, see: https://cloud.google.com/compute/docs/labeling-resources#restrictions
Apr 27, 2022 12:33:06 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-27T12:33:04.785Z: Worker configuration: e2-standard-2 in us-central1-f.
Apr 27, 2022 12:33:06 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-27T12:33:05.467Z: Expanding SplittableParDo operations into optimizable parts.
Apr 27, 2022 12:33:06 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-27T12:33:05.508Z: Expanding CollectionToSingleton operations into optimizable parts.
Apr 27, 2022 12:33:06 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-27T12:33:05.573Z: Expanding CoGroupByKey operations into optimizable parts.
Apr 27, 2022 12:33:06 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-27T12:33:05.643Z: Expanding SplittableProcessKeyed operations into optimizable parts.
Apr 27, 2022 12:33:06 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-27T12:33:05.688Z: Expanding GroupByKey operations into streaming Read/Write steps
Apr 27, 2022 12:33:06 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-27T12:33:05.742Z: Lifting ValueCombiningMappingFns into MergeBucketsMappingFns
Apr 27, 2022 12:33:06 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-27T12:33:05.818Z: Fusing adjacent ParDo, Read, Write, and Flatten operations
Apr 27, 2022 12:33:06 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-27T12:33:05.842Z: Unzipping flatten CoGroupByKey-Flatten for input CoGroupByKey-MakeUnionTable0-ParMultiDo-ConstructUnionTable-.output
Apr 27, 2022 12:33:06 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-27T12:33:05.876Z: Fusing unzipped copy of CoGroupByKey/GBK/WriteStream, through flatten CoGroupByKey/Flatten, into producer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable)
Apr 27, 2022 12:33:06 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-27T12:33:05.911Z: Fusing consumer CoGroupByKey/GBK/WriteStream into CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable)
Apr 27, 2022 12:33:06 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-27T12:33:05.942Z: Fusing consumer Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read input/Impulse
Apr 27, 2022 12:33:06 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-27T12:33:06.005Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
Apr 27, 2022 12:33:06 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-27T12:33:06.037Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
Apr 27, 2022 12:33:06 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-27T12:33:06.069Z: Fusing consumer Read input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
Apr 27, 2022 12:33:06 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-27T12:33:06.103Z: Fusing consumer Collect start time metrics (input)/ParMultiDo(TimeMonitor) into Read input/ParDo(StripIds)/ParMultiDo(StripIds)
Apr 27, 2022 12:33:06 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-27T12:33:06.136Z: Fusing consumer Window.Into()/Window.Assign into Collect start time metrics (input)/ParMultiDo(TimeMonitor)
Apr 27, 2022 12:33:06 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-27T12:33:06.174Z: Fusing consumer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable) into Window.Into()/Window.Assign
Apr 27, 2022 12:33:06 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-27T12:33:06.229Z: Fusing consumer Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read co-input/Impulse
Apr 27, 2022 12:33:06 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-27T12:33:06.251Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
Apr 27, 2022 12:33:06 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-27T12:33:06.277Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
Apr 27, 2022 12:33:06 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-27T12:33:06.312Z: Fusing consumer Read co-input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
Apr 27, 2022 12:33:06 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-27T12:33:06.346Z: Fusing consumer Collect start time metrics (co-input)/ParMultiDo(TimeMonitor) into Read co-input/ParDo(StripIds)/ParMultiDo(StripIds)
Apr 27, 2022 12:33:06 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-27T12:33:06.378Z: Fusing consumer Window.Into()2/Window.Assign into Collect start time metrics (co-input)/ParMultiDo(TimeMonitor)
Apr 27, 2022 12:33:06 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-27T12:33:06.414Z: Fusing consumer CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable) into Window.Into()2/Window.Assign
Apr 27, 2022 12:33:06 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-27T12:33:06.459Z: Fusing consumer CoGroupByKey/GBK/MergeBuckets into CoGroupByKey/GBK/ReadStream
Apr 27, 2022 12:33:06 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-27T12:33:06.491Z: Fusing consumer CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult) into CoGroupByKey/GBK/MergeBuckets
Apr 27, 2022 12:33:06 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-27T12:33:06.522Z: Fusing consumer Ungroup and reiterate/ParMultiDo(UngroupAndReiterate) into CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult)
Apr 27, 2022 12:33:06 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-27T12:33:06.555Z: Fusing consumer Collect total bytes/ParMultiDo(ByteMonitor) into Ungroup and reiterate/ParMultiDo(UngroupAndReiterate)
Apr 27, 2022 12:33:06 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-27T12:33:06.588Z: Fusing consumer Collect end time metrics/ParMultiDo(TimeMonitor) into Collect total bytes/ParMultiDo(ByteMonitor)
Apr 27, 2022 12:33:06 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-27T12:33:06.790Z: Running job using Streaming Engine
Apr 27, 2022 12:33:09 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-27T12:33:07.056Z: Starting 5 ****s in us-central1-f...
Apr 27, 2022 12:33:16 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-27T12:33:14.872Z: Your project already contains 100 Dataflow-created metric descriptors, so new user metrics of the form custom.googleapis.com/* will not be created. However, all user metrics are also available in the metric dataflow.googleapis.com/job/user_counter. If you rely on the custom metrics, you can delete old / unused metric descriptors. See https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.list and https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.delete
Apr 27, 2022 12:33:30 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-27T12:33:29.714Z: Autoscaling: Raised the number of ****s to 2 so that the pipeline can catch up with its backlog and keep up with its input rate.
Apr 27, 2022 12:33:30 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-27T12:33:29.748Z: Resized **** pool to 2, though goal was 5.  This could be a quota issue.
Apr 27, 2022 12:33:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-27T12:33:39.988Z: Autoscaling: Raised the number of ****s to 5 so that the pipeline can catch up with its backlog and keep up with its input rate.
Apr 27, 2022 12:34:33 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-27T12:34:32.448Z: Workers have started successfully.
Apr 27, 2022 4:01:07 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-27T16:01:05.631Z: Cancel request is committed for workflow job: 2022-04-27_05_32_52-13783208352401644293.
Apr 27, 2022 4:01:07 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-27T16:01:05.715Z: Cleaning up.
Apr 27, 2022 4:01:07 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-27T16:01:05.797Z: Stopping **** pool...
Apr 27, 2022 4:01:07 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-27T16:01:05.851Z: Stopping **** pool...
Apr 27, 2022 4:01:42 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-27T16:01:39.762Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Apr 27, 2022 4:01:42 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-27T16:01:39.810Z: Worker pool stopped.
Apr 27, 2022 4:01:47 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-04-27_05_32_52-13783208352401644293 finished with status CANCELLED.
Load test results for test (ID): f0c9bf3d-1dfb-4478-b631-3345fde1f7d2 and timestamp: 2022-04-27T12:32:46.828000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                 12344.836
dataflow_v2_java17_total_bytes_count              3.6730648E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220427123044
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:663e4583e395c5e749b4eb8afd1c9c87ebcf9081e05a13c047a84b1c639c52fd
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220427123044]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:663e4583e395c5e749b4eb8afd1c9c87ebcf9081e05a13c047a84b1c639c52fd]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220427123044] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:663e4583e395c5e749b4eb8afd1c9c87ebcf9081e05a13c047a84b1c639c52fd])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:663e4583e395c5e749b4eb8afd1c9c87ebcf9081e05a13c047a84b1c639c52fd
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:663e4583e395c5e749b4eb8afd1c9c87ebcf9081e05a13c047a84b1c639c52fd
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:663e4583e395c5e749b4eb8afd1c9c87ebcf9081e05a13c047a84b1c639c52fd].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.4/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 31m 28s
110 actionable tasks: 74 executed, 32 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/qx42rbv36bzjg

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #127

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/127/display/redirect?page=changes>

Changes:

[Heejong Lee] [BEAM-14343] Allow expansion service override in ExternalPythonTransform

[Heejong Lee] update

[Heejong Lee] allows remote host

[Heejong Lee] improve compatibility with python rowcoder

[ahmedabualsaud] added tempLocation to test pipeline options

[ahmedabualsaud] using tempRoot for temp bucket location

[ahmedabualsaud] small fixes

[noreply] [BEAM-14320] Update programming-guide w/Java GroupByKey example (#17369)

[noreply] Minor: Fix release script for `current` symlinks (#17457)

[noreply] Minor: fix typo (#17452)

[noreply] Change return type for PytorchInferenceRunner (#17460)

[noreply] [BEAM-13608] JmsIO dynamic topics feature (#17163)

[Heejong Lee] add test


------------------------------------------
[...truncated 266.00 KB...]
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmi
Apr 26, 2022 12:41:39 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
SEVERE: 2022-04-26T12:41:38.717Z: generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
Apr 26, 2022 4:01:09 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-26T16:01:06.136Z: Cancel request is committed for workflow job: 2022-04-26_05_32_36-10559723092899471543.
Apr 26, 2022 4:01:09 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-26T16:01:06.236Z: Cleaning up.
Apr 26, 2022 4:01:09 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-26T16:01:06.348Z: Stopping **** pool...
Apr 26, 2022 4:01:09 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-26T16:01:06.402Z: Stopping **** pool...
Apr 26, 2022 4:01:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-26T16:01:41.609Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Apr 26, 2022 4:01:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-26T16:01:41.698Z: Worker pool stopped.
Apr 26, 2022 4:01:49 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-04-26_05_32_36-10559723092899471543 finished with status CANCELLED.
Load test results for test (ID): 4d63233f-5eeb-4551-9e29-9f8f1a886f3e and timestamp: 2022-04-26T12:32:31.049000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                 12381.962
dataflow_v2_java17_total_bytes_count             3.83593313E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220426123035
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:dda484922e902119919c5fb891b4ba16ad2be7bb5db8e969575b5cef50f11c50
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220426123035]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:dda484922e902119919c5fb891b4ba16ad2be7bb5db8e969575b5cef50f11c50]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220426123035] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:dda484922e902119919c5fb891b4ba16ad2be7bb5db8e969575b5cef50f11c50])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:dda484922e902119919c5fb891b4ba16ad2be7bb5db8e969575b5cef50f11c50
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:dda484922e902119919c5fb891b4ba16ad2be7bb5db8e969575b5cef50f11c50
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:dda484922e902119919c5fb891b4ba16ad2be7bb5db8e969575b5cef50f11c50].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.4/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 31m 35s
110 actionable tasks: 73 executed, 33 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/ck5n63lkyh5d6

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #126

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/126/display/redirect>

Changes:


------------------------------------------
[...truncated 52.75 KB...]
cc9d3b2e0c73: Preparing
06b058a4b9a7: Preparing
3bc383470c05: Preparing
e93827457889: Preparing
08fa02ce37eb: Preparing
a037458de4e0: Preparing
bafdbe68e4ae: Preparing
a13c519c6361: Preparing
2ae9e2f61925: Waiting
a037458de4e0: Waiting
bafdbe68e4ae: Waiting
08fa02ce37eb: Waiting
cc9d3b2e0c73: Waiting
a13c519c6361: Waiting
e93827457889: Waiting
06b058a4b9a7: Waiting
3bc383470c05: Waiting
712d27e00e03: Waiting
47c9d7fb5d05: Waiting
9cc4406b0a68: Waiting
a82745df9946: Waiting
e9d68c658358: Waiting
f0134388448b: Waiting
45ff9c2b079e: Waiting
34000863c977: Pushed
c9e867e365e1: Pushed
4e79872db04a: Pushed
d9073639ad20: Pushed
47cf847cdf23: Pushed
9cc4406b0a68: Pushed
a82745df9946: Pushed
e9d68c658358: Pushed
47c9d7fb5d05: Pushed
712d27e00e03: Pushed
45ff9c2b079e: Pushed
3bc383470c05: Layer already exists
e93827457889: Layer already exists
08fa02ce37eb: Layer already exists
a037458de4e0: Layer already exists
f0134388448b: Pushed
bafdbe68e4ae: Layer already exists
a13c519c6361: Layer already exists
cc9d3b2e0c73: Pushed
06b058a4b9a7: Pushed
2ae9e2f61925: Pushed
20220425123042: digest: sha256:c8d7b6367a6d100d98235579c4752deeafcba963def5ca559ea35987a5fb17ad size: 4729

> Task :sdks:java:testing:load-tests:run
Apr 25, 2022 12:33:22 PM org.apache.beam.runners.dataflow.options.DataflowPipelineOptions$StagingLocationFactory create
INFO: No stagingLocation provided, falling back to gcpTempLocation
Apr 25, 2022 12:33:22 PM org.apache.beam.runners.dataflow.DataflowRunner fromOptions
INFO: PipelineOptions.filesToStage was not specified. Defaulting to files from the classpath: will stage 208 files. Enable logging at DEBUG level to see which files will be staged.
Apr 25, 2022 12:33:23 PM org.apache.beam.sdk.Pipeline validate
WARNING: The following transforms do not have stable unique names: Window.Into()
Apr 25, 2022 12:33:23 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Executing pipeline on the Dataflow Service, which will have billing implications related to Google Compute Engine usage and other Google Cloud Services.
Apr 25, 2022 12:33:26 PM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Uploading 208 files from PipelineOptions.filesToStage to staging location to prepare for execution.
Apr 25, 2022 12:33:28 PM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Staging files complete: 208 files cached, 0 files newly uploaded in 1 seconds
Apr 25, 2022 12:33:28 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Staging portable pipeline proto to gs://temp-storage-for-perf-tests/loadtests/staging/
Apr 25, 2022 12:33:28 PM org.apache.beam.runners.dataflow.util.PackageUtil tryStagePackage
INFO: Uploading <115405 bytes, hash b7aa88cc699ec5780882b5423e514ec2096d7b4ba193a228de2f0cdd1a28c105> to gs://temp-storage-for-perf-tests/loadtests/staging/pipeline-t6qIzGmexXgIgrVCPlFOwglte0uhk6Io3i8M3RoowQU.pb
Apr 25, 2022 12:33:29 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read input/DataflowRunner.StreamingUnboundedRead.ReadWithIds as step s1
Apr 25, 2022 12:33:30 PM org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource split
INFO: Split into 20 bundles of sizes: [org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@57272109, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@59696551, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@648d0e6d, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@79e66b2f, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@17273273, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5f69e2b, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@984169e, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@43f1bb92, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6d6bbd35, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5c5d6175, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7544ac86, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3b27b497, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@b1534d3, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3c74aa0d, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6c841199, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6a818392, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@489091bd, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@512d6e60, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1de9b505, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7b122839]
Apr 25, 2022 12:33:30 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read input/StripIds as step s2
Apr 25, 2022 12:33:30 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect start time metrics (input) as step s3
Apr 25, 2022 12:33:30 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Window.Into()/Window.Assign as step s4
Apr 25, 2022 12:33:30 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read co-input/DataflowRunner.StreamingUnboundedRead.ReadWithIds as step s5
Apr 25, 2022 12:33:30 PM org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource split
INFO: Split into 20 bundles of sizes: [org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6a6f6c7e, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6c5ddccd, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1dbd580, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6c101cc1, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7d0d91a1, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7fb48179, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@201c3cda, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4c86da0c, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5d97caa4, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6732726, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@474821de, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3d64c581, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5ec5ea63, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4190bc8a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@47d023b7, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5c83ae01, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2d64c100, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@69d45cca, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2fdf17dc, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6e6d4780]
Apr 25, 2022 12:33:30 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read co-input/StripIds as step s6
Apr 25, 2022 12:33:30 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect start time metrics (co-input) as step s7
Apr 25, 2022 12:33:30 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Window.Into()2/Window.Assign as step s8
Apr 25, 2022 12:33:30 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/MakeUnionTable0 as step s9
Apr 25, 2022 12:33:30 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/MakeUnionTable1 as step s10
Apr 25, 2022 12:33:30 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/Flatten as step s11
Apr 25, 2022 12:33:30 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/GBK as step s12
Apr 25, 2022 12:33:30 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/ConstructCoGbkResultFn as step s13
Apr 25, 2022 12:33:30 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Ungroup and reiterate as step s14
Apr 25, 2022 12:33:30 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect total bytes as step s15
Apr 25, 2022 12:33:30 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect end time metrics as step s16
Apr 25, 2022 12:33:30 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Dataflow SDK version: 2.39.0-SNAPSHOT
Apr 25, 2022 12:33:31 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To access the Dataflow monitoring console, please navigate to https://console.cloud.google.com/dataflow/jobs/us-central1/2022-04-25_05_33_30-12043335608045540008?project=apache-beam-testing
Apr 25, 2022 12:33:31 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Submitted job: 2022-04-25_05_33_30-12043335608045540008
Apr 25, 2022 12:33:31 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To cancel the job using the 'gcloud' tool, run:
> gcloud dataflow jobs --project=apache-beam-testing cancel --region=us-central1 2022-04-25_05_33_30-12043335608045540008
Apr 25, 2022 12:33:37 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
WARNING: 2022-04-25T12:33:36.436Z: The workflow name is not a valid Cloud Label. Labels applied to Cloud resources (such as GCE Instances) for monitoring will be labeled with this modified job name: load0tests0java170dataflow0v20streaming0cogbk01-jenkins-04-l9tt. For the best monitoring experience, please name your job with a valid Cloud Label. For details, see: https://cloud.google.com/compute/docs/labeling-resources#restrictions
Apr 25, 2022 12:33:42 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-25T12:33:41.206Z: Worker configuration: e2-standard-2 in us-central1-f.
Apr 25, 2022 12:33:42 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-25T12:33:42.026Z: Expanding SplittableParDo operations into optimizable parts.
Apr 25, 2022 12:33:42 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-25T12:33:42.061Z: Expanding CollectionToSingleton operations into optimizable parts.
Apr 25, 2022 12:33:42 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-25T12:33:42.119Z: Expanding CoGroupByKey operations into optimizable parts.
Apr 25, 2022 12:33:42 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-25T12:33:42.182Z: Expanding SplittableProcessKeyed operations into optimizable parts.
Apr 25, 2022 12:33:42 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-25T12:33:42.205Z: Expanding GroupByKey operations into streaming Read/Write steps
Apr 25, 2022 12:33:42 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-25T12:33:42.273Z: Lifting ValueCombiningMappingFns into MergeBucketsMappingFns
Apr 25, 2022 12:33:45 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-25T12:33:42.372Z: Fusing adjacent ParDo, Read, Write, and Flatten operations
Apr 25, 2022 12:33:45 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-25T12:33:42.439Z: Unzipping flatten CoGroupByKey-Flatten for input CoGroupByKey-MakeUnionTable0-ParMultiDo-ConstructUnionTable-.output
Apr 25, 2022 12:33:45 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-25T12:33:42.469Z: Fusing unzipped copy of CoGroupByKey/GBK/WriteStream, through flatten CoGroupByKey/Flatten, into producer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable)
Apr 25, 2022 12:33:45 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-25T12:33:42.522Z: Fusing consumer CoGroupByKey/GBK/WriteStream into CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable)
Apr 25, 2022 12:33:45 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-25T12:33:42.569Z: Fusing consumer Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read input/Impulse
Apr 25, 2022 12:33:45 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-25T12:33:42.591Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
Apr 25, 2022 12:33:45 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-25T12:33:42.622Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
Apr 25, 2022 12:33:45 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-25T12:33:42.646Z: Fusing consumer Read input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
Apr 25, 2022 12:33:45 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-25T12:33:42.677Z: Fusing consumer Collect start time metrics (input)/ParMultiDo(TimeMonitor) into Read input/ParDo(StripIds)/ParMultiDo(StripIds)
Apr 25, 2022 12:33:45 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-25T12:33:42.712Z: Fusing consumer Window.Into()/Window.Assign into Collect start time metrics (input)/ParMultiDo(TimeMonitor)
Apr 25, 2022 12:33:45 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-25T12:33:42.778Z: Fusing consumer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable) into Window.Into()/Window.Assign
Apr 25, 2022 12:33:45 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-25T12:33:42.800Z: Fusing consumer Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read co-input/Impulse
Apr 25, 2022 12:33:45 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-25T12:33:42.833Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
Apr 25, 2022 12:33:45 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-25T12:33:42.865Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
Apr 25, 2022 12:33:45 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-25T12:33:42.888Z: Fusing consumer Read co-input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
Apr 25, 2022 12:33:45 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-25T12:33:42.922Z: Fusing consumer Collect start time metrics (co-input)/ParMultiDo(TimeMonitor) into Read co-input/ParDo(StripIds)/ParMultiDo(StripIds)
Apr 25, 2022 12:33:45 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-25T12:33:42.956Z: Fusing consumer Window.Into()2/Window.Assign into Collect start time metrics (co-input)/ParMultiDo(TimeMonitor)
Apr 25, 2022 12:33:45 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-25T12:33:42.977Z: Fusing consumer CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable) into Window.Into()2/Window.Assign
Apr 25, 2022 12:33:45 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-25T12:33:42.999Z: Fusing consumer CoGroupByKey/GBK/MergeBuckets into CoGroupByKey/GBK/ReadStream
Apr 25, 2022 12:33:45 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-25T12:33:43.025Z: Fusing consumer CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult) into CoGroupByKey/GBK/MergeBuckets
Apr 25, 2022 12:33:45 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-25T12:33:43.048Z: Fusing consumer Ungroup and reiterate/ParMultiDo(UngroupAndReiterate) into CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult)
Apr 25, 2022 12:33:45 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-25T12:33:43.082Z: Fusing consumer Collect total bytes/ParMultiDo(ByteMonitor) into Ungroup and reiterate/ParMultiDo(UngroupAndReiterate)
Apr 25, 2022 12:33:45 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-25T12:33:43.114Z: Fusing consumer Collect end time metrics/ParMultiDo(TimeMonitor) into Collect total bytes/ParMultiDo(ByteMonitor)
Apr 25, 2022 12:33:45 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-25T12:33:43.273Z: Running job using Streaming Engine
Apr 25, 2022 12:33:45 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-25T12:33:43.560Z: Starting 5 ****s in us-central1-f...
Apr 25, 2022 12:34:06 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-25T12:34:05.733Z: Autoscaling: Raised the number of ****s to 5 so that the pipeline can catch up with its backlog and keep up with its input rate.
Apr 25, 2022 12:34:15 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-25T12:34:15.369Z: Your project already contains 100 Dataflow-created metric descriptors, so new user metrics of the form custom.googleapis.com/* will not be created. However, all user metrics are also available in the metric dataflow.googleapis.com/job/user_counter. If you rely on the custom metrics, you can delete old / unused metric descriptors. See https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.list and https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.delete
Apr 25, 2022 12:35:15 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-25T12:35:13.639Z: Workers have started successfully.
Apr 25, 2022 4:00:51 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-25T16:00:50.504Z: Cancel request is committed for workflow job: 2022-04-25_05_33_30-12043335608045540008.
Apr 25, 2022 4:00:55 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-25T16:00:55.276Z: Cleaning up.
Apr 25, 2022 4:00:55 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-25T16:00:55.348Z: Stopping **** pool...
Apr 25, 2022 4:00:55 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-25T16:00:55.393Z: Stopping **** pool...
Apr 25, 2022 4:01:33 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-25T16:01:32.367Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Apr 25, 2022 4:01:33 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-25T16:01:32.412Z: Worker pool stopped.
Apr 25, 2022 4:01:42 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-04-25_05_33_30-12043335608045540008 finished with status CANCELLED.
Load test results for test (ID): 1b9e34f8-3d87-47e9-a001-6cf2ce2c8668 and timestamp: 2022-04-25T12:33:23.163000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                 12314.251
dataflow_v2_java17_total_bytes_count             1.86143849E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220425123042
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:c8d7b6367a6d100d98235579c4752deeafcba963def5ca559ea35987a5fb17ad
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220425123042]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:c8d7b6367a6d100d98235579c4752deeafcba963def5ca559ea35987a5fb17ad]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220425123042] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:c8d7b6367a6d100d98235579c4752deeafcba963def5ca559ea35987a5fb17ad])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:c8d7b6367a6d100d98235579c4752deeafcba963def5ca559ea35987a5fb17ad
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:c8d7b6367a6d100d98235579c4752deeafcba963def5ca559ea35987a5fb17ad
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:c8d7b6367a6d100d98235579c4752deeafcba963def5ca559ea35987a5fb17ad].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.4/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 31m 22s
110 actionable tasks: 73 executed, 33 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/nhzlzapq2wnzq

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #125

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/125/display/redirect?page=changes>

Changes:

[noreply] [BEAM-13953] added documentation for BQ Storage Write API (#17391)


------------------------------------------
[...truncated 556.30 KB...]
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmi
Apr 24, 2022 3:29:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
SEVERE: 2022-04-24T15:29:54.472Z: Staged package error_prone_annotations-2.10.0-oknU0l37htQebIL8M131gBifDJ_uq9xTIz_B5QYHJKE.jar at location 'gs://temp-storage-for-perf-tests/loadtests/staging/error_prone_annotations-2.10.0-oknU0l37htQebIL8M131gBifDJ_uq9xTIz_B5QYHJKE.jar' is inaccessible.
Apr 24, 2022 3:29:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
SEVERE: 2022-04-24T15:29:54.736Z: Staged package gax-grpc-2.8.1-testlib-nH-xuZAVjRReb1TzhwGz0aBrSNkuxID3L6Ws8zqaZD8.jar at location 'gs://temp-storage-for-perf-tests/loadtests/staging/gax-grpc-2.8.1-testlib-nH-xuZAVjRReb1TzhwGz0aBrSNkuxID3L6Ws8zqaZD8.jar' is inaccessible.
Apr 24, 2022 3:29:58 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
WARNING: 2022-04-24T15:29:57.352Z: One or more access checks for temp location or staged files failed. Please refer to other error messages for details. For more information on security and permissions, please see https://cloud.google.com/dataflow/security-and-permissions.
Apr 24, 2022 4:00:58 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-24T16:00:57.071Z: Cancel request is committed for workflow job: 2022-04-24_05_32_40-2758132320092160384.
Apr 24, 2022 4:00:58 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-24T16:00:57.116Z: Cleaning up.
Apr 24, 2022 4:00:58 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-24T16:00:57.190Z: Stopping **** pool...
Apr 24, 2022 4:00:58 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-24T16:00:57.279Z: Stopping **** pool...
Apr 24, 2022 4:01:32 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-24T16:01:30.830Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Apr 24, 2022 4:01:32 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-24T16:01:30.880Z: Worker pool stopped.
Apr 24, 2022 4:01:36 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-04-24_05_32_40-2758132320092160384 finished with status CANCELLED.
Load test results for test (ID): b4faf98d-4252-476d-97ed-e3c0021847bf and timestamp: 2022-04-24T12:32:34.311000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                 12240.272
dataflow_v2_java17_total_bytes_count             3.88454972E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220424123036
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:e61bdcb503f1696876b5e7165dad0472c06f854503531b3a55567435696a11c6
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220424123036]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:e61bdcb503f1696876b5e7165dad0472c06f854503531b3a55567435696a11c6]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220424123036] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:e61bdcb503f1696876b5e7165dad0472c06f854503531b3a55567435696a11c6])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:e61bdcb503f1696876b5e7165dad0472c06f854503531b3a55567435696a11c6
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:e61bdcb503f1696876b5e7165dad0472c06f854503531b3a55567435696a11c6
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:e61bdcb503f1696876b5e7165dad0472c06f854503531b3a55567435696a11c6].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.4/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 31m 23s
110 actionable tasks: 73 executed, 33 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/e5x46eyj3ggzi

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 - Build # 124 - Aborted!

Posted by Apache Jenkins Server <je...@builds.apache.org>.
beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 - Build # 124 - Aborted:

Check console output at https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/124/ to view the results.

Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #123

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/123/display/redirect?page=changes>

Changes:

[mmack] [BEAM-14335] Spotless Spark sources

[mmack] [BEAM-14345] Force paranamer 2.8 for Spark Hadoop version tests to avoid

[kamil.bregula] Revert "[BEAM-14300] Fix Java precommit failure"

[kamil.bregula] Revert "Merge pull request #17223 from [BEAM-14215] Improve argument

[noreply] [BEAM-13657] Sunset python 3.6 (#17252)

[noreply] Removes unsupported Python 3.6 from the release validation script

[noreply] [BEAM-13984] Implement RunInference for PyTorch (#17196)

[noreply] [BEAM-13945] add json type support for java bigquery connector (#17209)

[noreply] [BEAM-14346] Fix incorrect error case index in ret2() (#17425)

[noreply] [BEAM-14342] Fix wrong default buffer type in fn_runner (#17420)

[noreply] Updates opencensus-api dependency to the latest version - 0.31.0

[noreply] [BEAM-14306] Add unit testing to pane coder (#17370)

[noreply] Updated the dep and golden for screen diff integration tests (#17442)


------------------------------------------
[...truncated 73.50 KB...]
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
Apr 22, 2022 4:01:08 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-22T16:01:06.907Z: Cancel request is committed for workflow job: 2022-04-22_05_32_45-12150457720570078207.
Apr 22, 2022 4:01:08 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-22T16:01:07.360Z: Cleaning up.
Apr 22, 2022 4:01:08 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-22T16:01:07.483Z: Stopping **** pool...
Apr 22, 2022 4:01:08 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-22T16:01:07.527Z: Stopping **** pool...
Apr 22, 2022 4:01:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-22T16:01:39.793Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Apr 22, 2022 4:01:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-22T16:01:39.919Z: Worker pool stopped.
Apr 22, 2022 4:01:46 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-04-22_05_32_45-12150457720570078207 finished with status CANCELLED.
Load test results for test (ID): 8efaf76e-f982-4e87-a1c5-795f0aefdff6 and timestamp: 2022-04-22T12:32:39.059000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                 12367.843
dataflow_v2_java17_total_bytes_count             2.30870961E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220422123041
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:5a40815573ef05e81b82b884f5f2067779fb56ac6eb374813ef4a350e4a2f56b
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220422123041]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:5a40815573ef05e81b82b884f5f2067779fb56ac6eb374813ef4a350e4a2f56b]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220422123041] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:5a40815573ef05e81b82b884f5f2067779fb56ac6eb374813ef4a350e4a2f56b])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:5a40815573ef05e81b82b884f5f2067779fb56ac6eb374813ef4a350e4a2f56b
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:5a40815573ef05e81b82b884f5f2067779fb56ac6eb374813ef4a350e4a2f56b
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:5a40815573ef05e81b82b884f5f2067779fb56ac6eb374813ef4a350e4a2f56b].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.4/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 31m 26s
110 actionable tasks: 74 executed, 32 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/vqmcwxbnxikfe

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #122

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/122/display/redirect?page=changes>

Changes:

[vachan] Annotating Read API tests.

[bulat.safiullin] [BEAM-14247] [Website] add image

[bulat.safiullin] [BEAM-14247] [Website] center image

[mattcasters] BEAM-1857 : CHANGES.md entry for 2.38.0

[mmack] [BEAM-14323] Improve IDE integration of Spark cross version builds

[noreply] [BEAM-14112] Fixed ReadFromBigQuery with Interactive Beam (#17306)

[noreply] Update .asf.yaml (#17409)

[noreply] [BEAM-14336] Sickbay flight delays test - dataset seems to be missing

[noreply] [BEAM-14338] Update watermark unit tests to use time.Time.Equals()

[noreply] [BEAM-14328] Tweaks to "Differences from pandas" page (#17413)

[Andrew Pilloud] [BEAM-14253] Disable broken test pending Dataflow fix

[yiru] fix: BigQuery Storage Connector trace id population missing bracket

[noreply] [BEAM-14330] Temporarily disable the clusters auto-cleanup (#17400)

[noreply] Update Beam website to release 2.38.0 (#17378)

[noreply] [BEAM-14213] Add API and construction time validation for Batched DoFns

[noreply] Minor: Update release guide regarding archive.apache.org (#17419)

[noreply] [BEAM-14017] beam_PreCommit_CommunityMetrics_Cron test failing (#17396)

[noreply] BEAM-13582 Fixing broken links in the documentation (#17300)


------------------------------------------
[...truncated 296.04 KB...]
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmi
Apr 21, 2022 4:00:53 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-21T16:00:50.620Z: Cancel request is committed for workflow job: 2022-04-21_05_32_39-4975189586864708056.
Apr 21, 2022 4:00:53 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-21T16:00:50.757Z: Cleaning up.
Apr 21, 2022 4:00:53 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-21T16:00:50.818Z: Stopping **** pool...
Apr 21, 2022 4:00:53 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-21T16:00:50.876Z: Stopping **** pool...
Apr 21, 2022 4:01:29 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-21T16:01:27.729Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Apr 21, 2022 4:01:29 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-21T16:01:27.773Z: Worker pool stopped.
Apr 21, 2022 4:01:35 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-04-21_05_32_39-4975189586864708056 finished with status CANCELLED.
Load test results for test (ID): 69cefdbc-5566-4ae6-8d36-deadae2f54a9 and timestamp: 2022-04-21T12:32:33.827000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                 12360.759
dataflow_v2_java17_total_bytes_count             3.86006401E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220421123040
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:d75db3a9f4e11ee1af2fac745b8daa23835fede6d9d30368df784c0733674d91
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220421123040]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:d75db3a9f4e11ee1af2fac745b8daa23835fede6d9d30368df784c0733674d91]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220421123040] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:d75db3a9f4e11ee1af2fac745b8daa23835fede6d9d30368df784c0733674d91])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:d75db3a9f4e11ee1af2fac745b8daa23835fede6d9d30368df784c0733674d91
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:d75db3a9f4e11ee1af2fac745b8daa23835fede6d9d30368df784c0733674d91
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:d75db3a9f4e11ee1af2fac745b8daa23835fede6d9d30368df784c0733674d91].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.4/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 31m 16s
110 actionable tasks: 73 executed, 33 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/xbd7uwrkmoupa

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #121

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/121/display/redirect?page=changes>

Changes:

[andyye333] Change func to PTransform

[noreply] Populate actual dataflow job id to bigquery write trace id (#17130)

[relax] mark static thread as a daemon thread

[noreply] [BEAM-13866] Add miscellaneous exec unit tests (#17363)


------------------------------------------
[...truncated 278.69 KB...]
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmi
Apr 20, 2022 12:39:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
SEVERE: 2022-04-20T12:39:53.912Z: generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
Apr 20, 2022 2:25:40 PM org.apache.beam.sdk.metrics.MetricsEnvironment getCurrentContainer
WARNING: Reporting metrics are not supported in the current execution environment.
Apr 20, 2022 4:00:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-20T16:00:41.522Z: Cancel request is committed for workflow job: 2022-04-20_05_33_06-426165222419621729.
Apr 20, 2022 4:00:47 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-20T16:00:46.465Z: Cleaning up.
Apr 20, 2022 4:00:47 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-20T16:00:46.544Z: Stopping **** pool...
Apr 20, 2022 4:00:47 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-20T16:00:46.600Z: Stopping **** pool...
Apr 20, 2022 4:01:39 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-20T16:01:38.405Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Apr 20, 2022 4:01:39 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-20T16:01:38.436Z: Worker pool stopped.
Apr 20, 2022 4:01:45 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-04-20_05_33_06-426165222419621729 finished with status CANCELLED.
Load test results for test (ID): b9981ed6-af2e-4594-8adb-e07903a06596 and timestamp: 2022-04-20T12:33:00.824000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                 12295.121
dataflow_v2_java17_total_bytes_count             4.18616716E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220420123035
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:ac7d89b321c2d902732fa25bb4c64744ce7de142c48e22272b4be7dc53db75fa
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220420123035]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:ac7d89b321c2d902732fa25bb4c64744ce7de142c48e22272b4be7dc53db75fa]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220420123035] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:ac7d89b321c2d902732fa25bb4c64744ce7de142c48e22272b4be7dc53db75fa])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:ac7d89b321c2d902732fa25bb4c64744ce7de142c48e22272b4be7dc53db75fa
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:ac7d89b321c2d902732fa25bb4c64744ce7de142c48e22272b4be7dc53db75fa
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:ac7d89b321c2d902732fa25bb4c64744ce7de142c48e22272b4be7dc53db75fa].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.4/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 31m 52s
110 actionable tasks: 74 executed, 32 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/ygqdlziw45zlg

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #120

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/120/display/redirect?page=changes>

Changes:

[Heejong Lee] [BEAM-14251] add output_coder_override to ExpansionRequest

[Heejong Lee] remove null

[rarokni] [BEAM-14307] Fix Slow Side input pattern bug in sample

[Heejong Lee] better error msg

[Heejong Lee] update from comments

[noreply] [BEAM-14316] Introducing KafkaIO.Read implementation compatibility

[noreply] [BEAM-14290] Address staticcheck warnings in the reflectx package

[noreply] [BEAM-14302] Simply bools in fn.go, genx_test.go (#17356)

[noreply] Merge pull request #17382: [BEAM-12356] Close DatasetService leak as


------------------------------------------
[...truncated 603.30 KB...]
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmi
Apr 19, 2022 4:00:37 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-19T16:00:35.094Z: Cancel request is committed for workflow job: 2022-04-19_05_32_43-4783407724810226384.
Apr 19, 2022 4:00:37 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-19T16:00:35.218Z: Cleaning up.
Apr 19, 2022 4:00:37 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-19T16:00:35.281Z: Stopping **** pool...
Apr 19, 2022 4:00:37 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-19T16:00:35.348Z: Stopping **** pool...
Apr 19, 2022 4:01:22 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-19T16:01:20.787Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Apr 19, 2022 4:01:22 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-19T16:01:20.825Z: Worker pool stopped.
Apr 19, 2022 4:01:27 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-04-19_05_32_43-4783407724810226384 finished with status CANCELLED.
Load test results for test (ID): 11d10eda-55e5-4e22-88c7-2210a04f5830 and timestamp: 2022-04-19T12:32:38.652000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                  12334.25
dataflow_v2_java17_total_bytes_count             3.96843308E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220419123039
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:398d8ea056f8445ad50eb460635608987789ba49af2871667a67805752ce3e14
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220419123039]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:398d8ea056f8445ad50eb460635608987789ba49af2871667a67805752ce3e14]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220419123039] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:398d8ea056f8445ad50eb460635608987789ba49af2871667a67805752ce3e14])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:398d8ea056f8445ad50eb460635608987789ba49af2871667a67805752ce3e14
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:398d8ea056f8445ad50eb460635608987789ba49af2871667a67805752ce3e14
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:398d8ea056f8445ad50eb460635608987789ba49af2871667a67805752ce3e14].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.4/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 31m 9s
110 actionable tasks: 74 executed, 32 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/nbrwfottzid4k

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #119

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/119/display/redirect>

Changes:


------------------------------------------
[...truncated 52.86 KB...]
Apr 18, 2022 12:32:36 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Staging portable pipeline proto to gs://temp-storage-for-perf-tests/loadtests/staging/
Apr 18, 2022 12:32:36 PM org.apache.beam.runners.dataflow.util.PackageUtil tryStagePackage
INFO: Uploading <115405 bytes, hash 32f81f2b8cc26e63e0172c1bf398abf417be3a6ad4c55b00f6fac5fec2b06166> to gs://temp-storage-for-perf-tests/loadtests/staging/pipeline-MvgfK4zCbmPgFywb85ir9Be-OmrUxVsA9vrF_sKwYWY.pb
Apr 18, 2022 12:32:37 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read input/DataflowRunner.StreamingUnboundedRead.ReadWithIds as step s1
Apr 18, 2022 12:32:37 PM org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource split
INFO: Split into 20 bundles of sizes: [org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@57272109, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@59696551, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@648d0e6d, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@79e66b2f, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@17273273, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5f69e2b, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@984169e, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@43f1bb92, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6d6bbd35, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5c5d6175, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7544ac86, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3b27b497, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@b1534d3, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3c74aa0d, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6c841199, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6a818392, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@489091bd, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@512d6e60, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1de9b505, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7b122839]
Apr 18, 2022 12:32:37 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read input/StripIds as step s2
Apr 18, 2022 12:32:37 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect start time metrics (input) as step s3
Apr 18, 2022 12:32:37 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Window.Into()/Window.Assign as step s4
Apr 18, 2022 12:32:37 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read co-input/DataflowRunner.StreamingUnboundedRead.ReadWithIds as step s5
Apr 18, 2022 12:32:37 PM org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource split
INFO: Split into 20 bundles of sizes: [org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6a6f6c7e, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6c5ddccd, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1dbd580, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6c101cc1, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7d0d91a1, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7fb48179, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@201c3cda, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4c86da0c, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5d97caa4, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6732726, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@474821de, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3d64c581, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5ec5ea63, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4190bc8a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@47d023b7, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5c83ae01, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2d64c100, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@69d45cca, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2fdf17dc, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6e6d4780]
Apr 18, 2022 12:32:37 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read co-input/StripIds as step s6
Apr 18, 2022 12:32:37 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect start time metrics (co-input) as step s7
Apr 18, 2022 12:32:37 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Window.Into()2/Window.Assign as step s8
Apr 18, 2022 12:32:37 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/MakeUnionTable0 as step s9
Apr 18, 2022 12:32:37 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/MakeUnionTable1 as step s10
Apr 18, 2022 12:32:37 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/Flatten as step s11
Apr 18, 2022 12:32:37 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/GBK as step s12
Apr 18, 2022 12:32:37 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/ConstructCoGbkResultFn as step s13
Apr 18, 2022 12:32:37 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Ungroup and reiterate as step s14
Apr 18, 2022 12:32:37 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect total bytes as step s15
Apr 18, 2022 12:32:37 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect end time metrics as step s16
Apr 18, 2022 12:32:38 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Dataflow SDK version: 2.39.0-SNAPSHOT
Apr 18, 2022 12:32:38 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To access the Dataflow monitoring console, please navigate to https://console.cloud.google.com/dataflow/jobs/us-central1/2022-04-18_05_32_38-2279234934055603378?project=apache-beam-testing
Apr 18, 2022 12:32:38 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Submitted job: 2022-04-18_05_32_38-2279234934055603378
Apr 18, 2022 12:32:38 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To cancel the job using the 'gcloud' tool, run:
> gcloud dataflow jobs --project=apache-beam-testing cancel --region=us-central1 2022-04-18_05_32_38-2279234934055603378
Apr 18, 2022 12:32:44 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
WARNING: 2022-04-18T12:32:43.788Z: The workflow name is not a valid Cloud Label. Labels applied to Cloud resources (such as GCE Instances) for monitoring will be labeled with this modified job name: load0tests0java170dataflow0v20streaming0cogbk01-jenkins-04-gxql. For the best monitoring experience, please name your job with a valid Cloud Label. For details, see: https://cloud.google.com/compute/docs/labeling-resources#restrictions
Apr 18, 2022 12:32:48 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-18T12:32:47.975Z: Worker configuration: e2-standard-2 in us-central1-b.
Apr 18, 2022 12:32:50 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-18T12:32:48.764Z: Expanding SplittableParDo operations into optimizable parts.
Apr 18, 2022 12:32:50 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-18T12:32:48.792Z: Expanding CollectionToSingleton operations into optimizable parts.
Apr 18, 2022 12:32:50 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-18T12:32:48.847Z: Expanding CoGroupByKey operations into optimizable parts.
Apr 18, 2022 12:32:50 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-18T12:32:48.928Z: Expanding SplittableProcessKeyed operations into optimizable parts.
Apr 18, 2022 12:32:50 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-18T12:32:48.957Z: Expanding GroupByKey operations into streaming Read/Write steps
Apr 18, 2022 12:32:50 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-18T12:32:49.014Z: Lifting ValueCombiningMappingFns into MergeBucketsMappingFns
Apr 18, 2022 12:32:50 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-18T12:32:49.139Z: Fusing adjacent ParDo, Read, Write, and Flatten operations
Apr 18, 2022 12:32:50 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-18T12:32:49.164Z: Unzipping flatten CoGroupByKey-Flatten for input CoGroupByKey-MakeUnionTable0-ParMultiDo-ConstructUnionTable-.output
Apr 18, 2022 12:32:50 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-18T12:32:49.186Z: Fusing unzipped copy of CoGroupByKey/GBK/WriteStream, through flatten CoGroupByKey/Flatten, into producer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable)
Apr 18, 2022 12:32:50 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-18T12:32:49.217Z: Fusing consumer CoGroupByKey/GBK/WriteStream into CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable)
Apr 18, 2022 12:32:50 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-18T12:32:49.248Z: Fusing consumer Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read input/Impulse
Apr 18, 2022 12:32:50 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-18T12:32:49.282Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
Apr 18, 2022 12:32:50 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-18T12:32:49.313Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
Apr 18, 2022 12:32:50 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-18T12:32:49.336Z: Fusing consumer Read input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
Apr 18, 2022 12:32:50 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-18T12:32:49.367Z: Fusing consumer Collect start time metrics (input)/ParMultiDo(TimeMonitor) into Read input/ParDo(StripIds)/ParMultiDo(StripIds)
Apr 18, 2022 12:32:50 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-18T12:32:49.389Z: Fusing consumer Window.Into()/Window.Assign into Collect start time metrics (input)/ParMultiDo(TimeMonitor)
Apr 18, 2022 12:32:50 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-18T12:32:49.421Z: Fusing consumer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable) into Window.Into()/Window.Assign
Apr 18, 2022 12:32:50 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-18T12:32:49.446Z: Fusing consumer Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read co-input/Impulse
Apr 18, 2022 12:32:50 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-18T12:32:49.493Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
Apr 18, 2022 12:32:50 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-18T12:32:49.528Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
Apr 18, 2022 12:32:50 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-18T12:32:49.559Z: Fusing consumer Read co-input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
Apr 18, 2022 12:32:50 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-18T12:32:49.591Z: Fusing consumer Collect start time metrics (co-input)/ParMultiDo(TimeMonitor) into Read co-input/ParDo(StripIds)/ParMultiDo(StripIds)
Apr 18, 2022 12:32:50 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-18T12:32:49.635Z: Fusing consumer Window.Into()2/Window.Assign into Collect start time metrics (co-input)/ParMultiDo(TimeMonitor)
Apr 18, 2022 12:32:50 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-18T12:32:49.666Z: Fusing consumer CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable) into Window.Into()2/Window.Assign
Apr 18, 2022 12:32:50 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-18T12:32:49.697Z: Fusing consumer CoGroupByKey/GBK/MergeBuckets into CoGroupByKey/GBK/ReadStream
Apr 18, 2022 12:32:50 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-18T12:32:49.720Z: Fusing consumer CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult) into CoGroupByKey/GBK/MergeBuckets
Apr 18, 2022 12:32:50 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-18T12:32:49.755Z: Fusing consumer Ungroup and reiterate/ParMultiDo(UngroupAndReiterate) into CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult)
Apr 18, 2022 12:32:50 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-18T12:32:49.789Z: Fusing consumer Collect total bytes/ParMultiDo(ByteMonitor) into Ungroup and reiterate/ParMultiDo(UngroupAndReiterate)
Apr 18, 2022 12:32:50 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-18T12:32:49.823Z: Fusing consumer Collect end time metrics/ParMultiDo(TimeMonitor) into Collect total bytes/ParMultiDo(ByteMonitor)
Apr 18, 2022 12:32:50 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-18T12:32:49.990Z: Running job using Streaming Engine
Apr 18, 2022 12:32:52 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-18T12:32:50.240Z: Starting 5 ****s in us-central1-b...
Apr 18, 2022 12:32:55 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-18T12:32:54.782Z: Your project already contains 100 Dataflow-created metric descriptors, so new user metrics of the form custom.googleapis.com/* will not be created. However, all user metrics are also available in the metric dataflow.googleapis.com/job/user_counter. If you rely on the custom metrics, you can delete old / unused metric descriptors. See https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.list and https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.delete
Apr 18, 2022 12:33:20 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-18T12:33:20.535Z: Autoscaling: Raised the number of ****s to 2 so that the pipeline can catch up with its backlog and keep up with its input rate.
Apr 18, 2022 12:33:20 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-18T12:33:20.553Z: Resized **** pool to 2, though goal was 5.  This could be a quota issue.
Apr 18, 2022 12:33:31 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-18T12:33:30.837Z: Autoscaling: Raised the number of ****s to 5 so that the pipeline can catch up with its backlog and keep up with its input rate.
Apr 18, 2022 12:34:21 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-18T12:34:20.974Z: Workers have started successfully.
Apr 18, 2022 12:37:49 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
SEVERE: 2022-04-18T12:37:47.439Z: generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
Apr 18, 2022 12:37:51 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
SEVERE: 2022-04-18T12:37:50.916Z: generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
Apr 18, 2022 4:00:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-18T16:00:42.714Z: Cancel request is committed for workflow job: 2022-04-18_05_32_38-2279234934055603378.
Apr 18, 2022 4:00:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-18T16:00:42.828Z: Cleaning up.
Apr 18, 2022 4:00:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-18T16:00:42.983Z: Stopping **** pool...
Apr 18, 2022 4:00:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-18T16:00:43.025Z: Stopping **** pool...
Apr 18, 2022 4:01:37 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-18T16:01:37.008Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Apr 18, 2022 4:01:37 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-18T16:01:37.038Z: Worker pool stopped.
Apr 18, 2022 4:01:42 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-04-18_05_32_38-2279234934055603378 finished with status CANCELLED.
Load test results for test (ID): 41fb1f1f-1d70-40a1-91aa-fc95be2b5cea and timestamp: 2022-04-18T12:32:32.627000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                 12297.749
dataflow_v2_java17_total_bytes_count              2.7185534E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220418123034
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:a4df32eaf063e0e131f7d9440f668f5e0c40c242f28640b1d74abac4601a5ffc
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220418123034]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:a4df32eaf063e0e131f7d9440f668f5e0c40c242f28640b1d74abac4601a5ffc]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220418123034] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:a4df32eaf063e0e131f7d9440f668f5e0c40c242f28640b1d74abac4601a5ffc])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:a4df32eaf063e0e131f7d9440f668f5e0c40c242f28640b1d74abac4601a5ffc
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:a4df32eaf063e0e131f7d9440f668f5e0c40c242f28640b1d74abac4601a5ffc
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:a4df32eaf063e0e131f7d9440f668f5e0c40c242f28640b1d74abac4601a5ffc].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.4/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 31m 28s
110 actionable tasks: 73 executed, 33 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/7scm7ayhwflta

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #118

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/118/display/redirect>

Changes:


------------------------------------------
[...truncated 637.69 KB...]
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
Apr 17, 2022 4:01:02 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-17T16:01:00.019Z: Cancel request is committed for workflow job: 2022-04-17_05_32_49-8133379480979873444.
Apr 17, 2022 4:01:02 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-17T16:01:00.101Z: Cleaning up.
Apr 17, 2022 4:01:02 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-17T16:01:00.175Z: Stopping **** pool...
Apr 17, 2022 4:01:02 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-17T16:01:00.235Z: Stopping **** pool...
Apr 17, 2022 4:01:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-17T16:01:55.255Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Apr 17, 2022 4:01:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-17T16:01:55.299Z: Worker pool stopped.
Apr 17, 2022 4:02:01 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-04-17_05_32_49-8133379480979873444 finished with status CANCELLED.
Load test results for test (ID): d70e5409-1870-4777-a37f-7c2c351e5b0d and timestamp: 2022-04-17T12:32:39.505000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                 12368.731
dataflow_v2_java17_total_bytes_count             3.97045431E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220417123040
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:9dbf27da19729e3ddbc7bc5f595b3268e4fb3920d7b8b0f79f8617edd668c153
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220417123040]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:9dbf27da19729e3ddbc7bc5f595b3268e4fb3920d7b8b0f79f8617edd668c153]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220417123040] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:9dbf27da19729e3ddbc7bc5f595b3268e4fb3920d7b8b0f79f8617edd668c153])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:9dbf27da19729e3ddbc7bc5f595b3268e4fb3920d7b8b0f79f8617edd668c153
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:9dbf27da19729e3ddbc7bc5f595b3268e4fb3920d7b8b0f79f8617edd668c153
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:9dbf27da19729e3ddbc7bc5f595b3268e4fb3920d7b8b0f79f8617edd668c153].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.4/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 31m 41s
110 actionable tasks: 73 executed, 33 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/7ysdlikbx7obw

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #117

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/117/display/redirect?page=changes>

Changes:

[pandiana] BigQueryServicesImpl: reduce number of threads spawned by

[noreply] [BEAM-13204] Fix website bug where code tabs do not appear if the


------------------------------------------
[...truncated 182.64 KB...]
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmi
Apr 16, 2022 12:39:15 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
SEVERE: 2022-04-16T12:39:13.646Z: generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
Apr 16, 2022 4:00:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-16T16:00:42.923Z: Cancel request is committed for workflow job: 2022-04-16_05_32_38-3458726814305771999.
Apr 16, 2022 4:00:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-16T16:00:43.055Z: Cleaning up.
Apr 16, 2022 4:00:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-16T16:00:43.463Z: Stopping **** pool...
Apr 16, 2022 4:00:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-16T16:00:43.571Z: Stopping **** pool...
Apr 16, 2022 4:01:35 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-16T16:01:35.236Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Apr 16, 2022 4:01:35 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-16T16:01:35.273Z: Worker pool stopped.
Apr 16, 2022 4:01:43 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-04-16_05_32_38-3458726814305771999 finished with status CANCELLED.
Load test results for test (ID): 18408d1d-a895-4b4d-93e6-dea02a74d3f0 and timestamp: 2022-04-16T12:32:32.679000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                 12274.909
dataflow_v2_java17_total_bytes_count             3.06803042E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220416123040
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:2f305dee5a665a5943e6b107852db2e393161be627e50c3c07a9468916d0b1cc
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220416123040]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:2f305dee5a665a5943e6b107852db2e393161be627e50c3c07a9468916d0b1cc]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220416123040] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:2f305dee5a665a5943e6b107852db2e393161be627e50c3c07a9468916d0b1cc])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:2f305dee5a665a5943e6b107852db2e393161be627e50c3c07a9468916d0b1cc
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:2f305dee5a665a5943e6b107852db2e393161be627e50c3c07a9468916d0b1cc
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:2f305dee5a665a5943e6b107852db2e393161be627e50c3c07a9468916d0b1cc].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.4/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 31m 23s
110 actionable tasks: 73 executed, 33 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/esyedybeb576a

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #116

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/116/display/redirect?page=changes>

Changes:

[Kenneth Knowles] Upgrade to Gradle 7.4

[Kenneth Knowles] Remove Python module dependency on Dataflow worker

[noreply] [BEAM-13925] Dont double assign committers if author or other reviewer

[noreply] [BEAM-13739] Remove deprecated shallow clone funcs (#17362)

[noreply] [BEAM-11104] Pipe Continuation to DataSource level (#17334)

[noreply] [BEAM-11105] Basic Watermark Estimation (Wall Clock Observing) (#17267)

[noreply] Respect output coder for TextIO. (#17367)

[noreply] Merge pull request #17200 from [BEAM-12164]: fix the autoscaling backlog

[noreply] [BEAM-17035] Call python3 directly when it is available. (#17366)

[noreply] Merge pull request #17375: [BEAM-8691] Declare newer


------------------------------------------
[...truncated 71.10 KB...]
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
Apr 15, 2022 12:41:19 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
SEVERE: 2022-04-15T12:41:18.300Z: generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
Apr 15, 2022 4:01:07 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-15T16:01:05.786Z: Cancel request is committed for workflow job: 2022-04-15_05_32_48-13724693753058666410.
Apr 15, 2022 4:01:07 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-15T16:01:05.883Z: Cleaning up.
Apr 15, 2022 4:01:07 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-15T16:01:05.980Z: Stopping **** pool...
Apr 15, 2022 4:01:07 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-15T16:01:06.071Z: Stopping **** pool...
Apr 15, 2022 4:02:00 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-15T16:02:00.436Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Apr 15, 2022 4:02:00 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-15T16:02:00.504Z: Worker pool stopped.
Apr 15, 2022 4:02:06 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-04-15_05_32_48-13724693753058666410 finished with status CANCELLED.
Load test results for test (ID): 0435a4fb-2a03-4fa6-af5e-87148d28d2fa and timestamp: 2022-04-15T12:32:43.470000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                  12355.25
dataflow_v2_java17_total_bytes_count             4.53040508E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220415123038
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:53461e94a63c37c564d048ebadf5ba0e688f41cb7d8cb63c337a89a1b8f3f47c
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220415123038]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:53461e94a63c37c564d048ebadf5ba0e688f41cb7d8cb63c337a89a1b8f3f47c]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220415123038] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:53461e94a63c37c564d048ebadf5ba0e688f41cb7d8cb63c337a89a1b8f3f47c])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:837acb140e871f602a1bfdea8e8c28ed0b1e9a458ab48d2d986ff464eca7dc12
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:837acb140e871f602a1bfdea8e8c28ed0b1e9a458ab48d2d986ff464eca7dc12
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:837acb140e871f602a1bfdea8e8c28ed0b1e9a458ab48d2d986ff464eca7dc12].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:53461e94a63c37c564d048ebadf5ba0e688f41cb7d8cb63c337a89a1b8f3f47c
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:53461e94a63c37c564d048ebadf5ba0e688f41cb7d8cb63c337a89a1b8f3f47c
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:53461e94a63c37c564d048ebadf5ba0e688f41cb7d8cb63c337a89a1b8f3f47c].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.4/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 31m 52s
110 actionable tasks: 74 executed, 32 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/k52jyknmsqc7g

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #115

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/115/display/redirect?page=changes>

Changes:

[relax] handle changing schemas in Storage API sink

[noreply] Fix a couple style issues (#17361)

[noreply] [BEAM-14287] Clean up staticcheck warnings in graph/coder (#17337)

[noreply] Improvements to dataflow job service for non-Python jobs. (#17338)

[noreply] Bump minimist (#17290)

[noreply] Bump ansi-regex (#17291)

[noreply] Bump nanoid (#17292)

[noreply] Bump lodash (#17293)

[noreply] Bump url-parse (#17294)

[noreply] Bump moment (#17328)

[noreply] Merge pull request #15549 from [BEAM-11997] Changed RedisIO


------------------------------------------
[...truncated 553.44 KB...]
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==
Apr 14, 2022 4:01:14 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-14T16:01:12.736Z: Cancel request is committed for workflow job: 2022-04-14_05_32_42-8770949662358378302.
Apr 14, 2022 4:01:14 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-14T16:01:13.004Z: Cleaning up.
Apr 14, 2022 4:01:14 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-14T16:01:13.099Z: Stopping **** pool...
Apr 14, 2022 4:01:14 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-14T16:01:13.153Z: Stopping **** pool...
Apr 14, 2022 4:02:08 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-14T16:02:08.314Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Apr 14, 2022 4:02:08 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-14T16:02:08.352Z: Worker pool stopped.
Apr 14, 2022 4:02:16 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-04-14_05_32_42-8770949662358378302 finished with status CANCELLED.
Load test results for test (ID): 177c6c38-46ba-410b-9699-1564a041fe5f and timestamp: 2022-04-14T12:32:37.729000000Z:
Exception in thread "main"                  Metric:                    Value:
dataflow_v2_java17_runtime_sec                   12359.0
dataflow_v2_java17_total_bytes_count             3.92879535E10
java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220414123041
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:c8c32183cc6707a7b84326de30eea9462c7a3984216ecd24f7da6731a7402c7b
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220414123041]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:c8c32183cc6707a7b84326de30eea9462c7a3984216ecd24f7da6731a7402c7b]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220414123041] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:c8c32183cc6707a7b84326de30eea9462c7a3984216ecd24f7da6731a7402c7b])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:c8c32183cc6707a7b84326de30eea9462c7a3984216ecd24f7da6731a7402c7b
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:c8c32183cc6707a7b84326de30eea9462c7a3984216ecd24f7da6731a7402c7b
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:c8c32183cc6707a7b84326de30eea9462c7a3984216ecd24f7da6731a7402c7b].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 31m 55s
110 actionable tasks: 74 executed, 32 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/klpmailx5y7ua

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #114

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/114/display/redirect?page=changes>

Changes:

[Valentyn Tymofieiev] Add remaining Dataflow test suites for Python 3.9.

[Heejong Lee] [BEAM-14232] Only resolve artifacts in expanded environments for Java

[noreply] Fix test ordering issue (#17350)

[buqian] Do not pass null to MoreObjects.firstNonNull as default value

[ningkang0957] [BEAM-14288] Fixed flaky test

[noreply] [BEAM-14277] Disables Spanner change streams tests (#17346)

[noreply] [BEAM-14219] Run cleanup script to remove stale prebuilt SDK container

[Heejong Lee] [BEAM-14300] Fix Java precommit failure

[noreply] [BEAM-14116] Rollback "Chunk commit requests dynamically (#17004)"

[noreply] [BEAM-13982] A base class for run inference (#16970)

[ningkang0957] Enumerates all possible expected strings when asserting

[noreply] [BEAM-13966] Add pivot(), a non-deferred column operation on categorical


------------------------------------------
[...truncated 102.69 KB...]
Traceback (most recent call last):
  File "<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/license_scripts/pull_licenses_java.py",> line 58, in pull_from_url
    url_read = urlopen(url)
  File "/usr/lib/python3.8/urllib/request.py", line 222, in urlopen
    return opener.open(url, data, timeout)
  File "/usr/lib/python3.8/urllib/request.py", line 531, in open
    response = meth(req, response)
  File "/usr/lib/python3.8/urllib/request.py", line 640, in http_response
    response = self.parent.error(
  File "/usr/lib/python3.8/urllib/request.py", line 563, in error
    result = self._call_chain(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 502, in _call_chain
    result = func(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 755, in http_error_302
    return self.parent.open(new, timeout=req.timeout)
  File "/usr/lib/python3.8/urllib/request.py", line 531, in open
    response = meth(req, response)
  File "/usr/lib/python3.8/urllib/request.py", line 640, in http_response
    response = self.parent.error(
  File "/usr/lib/python3.8/urllib/request.py", line 563, in error
    result = self._call_chain(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 502, in _call_chain
    result = func(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 755, in http_error_302
    return self.parent.open(new, timeout=req.timeout)
  File "/usr/lib/python3.8/urllib/request.py", line 531, in open
    response = meth(req, response)
  File "/usr/lib/python3.8/urllib/request.py", line 640, in http_response
    response = self.parent.error(
  File "/usr/lib/python3.8/urllib/request.py", line 569, in error
    return self._call_chain(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 502, in _call_chain
    result = func(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 649, in http_error_default
    raise HTTPError(req.full_url, code, msg, hdrs, fp)
urllib.error.HTTPError: HTTP Error 502: Bad Gateway
ERROR:root:Invalid url for slf4j-api-1.7.30: http://www.opensource.org/licenses/mit-license.php after 9 retries.
Traceback (most recent call last):
  File "<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/license_scripts/pull_licenses_java.py",> line 58, in pull_from_url
    url_read = urlopen(url)
  File "/usr/lib/python3.8/urllib/request.py", line 222, in urlopen
    return opener.open(url, data, timeout)
  File "/usr/lib/python3.8/urllib/request.py", line 531, in open
    response = meth(req, response)
  File "/usr/lib/python3.8/urllib/request.py", line 640, in http_response
    response = self.parent.error(
  File "/usr/lib/python3.8/urllib/request.py", line 563, in error
    result = self._call_chain(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 502, in _call_chain
    result = func(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 755, in http_error_302
    return self.parent.open(new, timeout=req.timeout)
  File "/usr/lib/python3.8/urllib/request.py", line 531, in open
    response = meth(req, response)
  File "/usr/lib/python3.8/urllib/request.py", line 640, in http_response
    response = self.parent.error(
  File "/usr/lib/python3.8/urllib/request.py", line 563, in error
    result = self._call_chain(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 502, in _call_chain
    result = func(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 755, in http_error_302
    return self.parent.open(new, timeout=req.timeout)
  File "/usr/lib/python3.8/urllib/request.py", line 531, in open
    response = meth(req, response)
  File "/usr/lib/python3.8/urllib/request.py", line 640, in http_response
    response = self.parent.error(
  File "/usr/lib/python3.8/urllib/request.py", line 569, in error
    return self._call_chain(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 502, in _call_chain
    result = func(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 649, in http_error_default
    raise HTTPError(req.full_url, code, msg, hdrs, fp)
urllib.error.HTTPError: HTTP Error 502: Bad Gateway
ERROR:root:Invalid url for slf4j-simple-1.7.30: http://www.opensource.org/licenses/mit-license.php after 9 retries.
Traceback (most recent call last):
  File "<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/license_scripts/pull_licenses_java.py",> line 58, in pull_from_url
    url_read = urlopen(url)
  File "/usr/lib/python3.8/urllib/request.py", line 222, in urlopen
    return opener.open(url, data, timeout)
  File "/usr/lib/python3.8/urllib/request.py", line 531, in open
    response = meth(req, response)
  File "/usr/lib/python3.8/urllib/request.py", line 640, in http_response
    response = self.parent.error(
  File "/usr/lib/python3.8/urllib/request.py", line 563, in error
    result = self._call_chain(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 502, in _call_chain
    result = func(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 755, in http_error_302
    return self.parent.open(new, timeout=req.timeout)
  File "/usr/lib/python3.8/urllib/request.py", line 531, in open
    response = meth(req, response)
  File "/usr/lib/python3.8/urllib/request.py", line 640, in http_response
    response = self.parent.error(
  File "/usr/lib/python3.8/urllib/request.py", line 569, in error
    return self._call_chain(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 502, in _call_chain
    result = func(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 649, in http_error_default
    raise HTTPError(req.full_url, code, msg, hdrs, fp)
urllib.error.HTTPError: HTTP Error 502: Bad Gateway
ERROR:root:Invalid url for classgraph-4.8.104: http://opensource.org/licenses/MIT. Retrying...
Traceback (most recent call last):
  File "<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/license_scripts/pull_licenses_java.py",> line 58, in pull_from_url
    url_read = urlopen(url)
  File "/usr/lib/python3.8/urllib/request.py", line 222, in urlopen
    return opener.open(url, data, timeout)
  File "/usr/lib/python3.8/urllib/request.py", line 531, in open
    response = meth(req, response)
  File "/usr/lib/python3.8/urllib/request.py", line 640, in http_response
    response = self.parent.error(
  File "/usr/lib/python3.8/urllib/request.py", line 563, in error
    result = self._call_chain(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 502, in _call_chain
    result = func(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 755, in http_error_302
    return self.parent.open(new, timeout=req.timeout)
  File "/usr/lib/python3.8/urllib/request.py", line 531, in open
    response = meth(req, response)
  File "/usr/lib/python3.8/urllib/request.py", line 640, in http_response
    response = self.parent.error(
  File "/usr/lib/python3.8/urllib/request.py", line 569, in error
    return self._call_chain(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 502, in _call_chain
    result = func(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 649, in http_error_default
    raise HTTPError(req.full_url, code, msg, hdrs, fp)
urllib.error.HTTPError: HTTP Error 502: Bad Gateway
ERROR:root:Invalid url for kafka-1.16.3: http://opensource.org/licenses/MIT after 9 retries.
Traceback (most recent call last):
  File "<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/license_scripts/pull_licenses_java.py",> line 58, in pull_from_url
    url_read = urlopen(url)
  File "/usr/lib/python3.8/urllib/request.py", line 222, in urlopen
    return opener.open(url, data, timeout)
  File "/usr/lib/python3.8/urllib/request.py", line 531, in open
    response = meth(req, response)
  File "/usr/lib/python3.8/urllib/request.py", line 640, in http_response
    response = self.parent.error(
  File "/usr/lib/python3.8/urllib/request.py", line 563, in error
    result = self._call_chain(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 502, in _call_chain
    result = func(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 755, in http_error_302
    return self.parent.open(new, timeout=req.timeout)
  File "/usr/lib/python3.8/urllib/request.py", line 531, in open
    response = meth(req, response)
  File "/usr/lib/python3.8/urllib/request.py", line 640, in http_response
    response = self.parent.error(
  File "/usr/lib/python3.8/urllib/request.py", line 569, in error
    return self._call_chain(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 502, in _call_chain
    result = func(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 649, in http_error_default
    raise HTTPError(req.full_url, code, msg, hdrs, fp)
urllib.error.HTTPError: HTTP Error 502: Bad Gateway
ERROR:root:Invalid url for classgraph-4.8.104: http://opensource.org/licenses/MIT. Retrying...
Traceback (most recent call last):
  File "<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/license_scripts/pull_licenses_java.py",> line 58, in pull_from_url
    url_read = urlopen(url)
  File "/usr/lib/python3.8/urllib/request.py", line 222, in urlopen
    return opener.open(url, data, timeout)
  File "/usr/lib/python3.8/urllib/request.py", line 531, in open
    response = meth(req, response)
  File "/usr/lib/python3.8/urllib/request.py", line 640, in http_response
    response = self.parent.error(
  File "/usr/lib/python3.8/urllib/request.py", line 563, in error
    result = self._call_chain(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 502, in _call_chain
    result = func(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 755, in http_error_302
    return self.parent.open(new, timeout=req.timeout)
  File "/usr/lib/python3.8/urllib/request.py", line 531, in open
    response = meth(req, response)
  File "/usr/lib/python3.8/urllib/request.py", line 640, in http_response
    response = self.parent.error(
  File "/usr/lib/python3.8/urllib/request.py", line 569, in error
    return self._call_chain(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 502, in _call_chain
    result = func(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 649, in http_error_default
    raise HTTPError(req.full_url, code, msg, hdrs, fp)
urllib.error.HTTPError: HTTP Error 502: Bad Gateway
ERROR:root:Invalid url for classgraph-4.8.104: http://opensource.org/licenses/MIT. Retrying...
Traceback (most recent call last):
  File "<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/license_scripts/pull_licenses_java.py",> line 58, in pull_from_url
    url_read = urlopen(url)
  File "/usr/lib/python3.8/urllib/request.py", line 222, in urlopen
    return opener.open(url, data, timeout)
  File "/usr/lib/python3.8/urllib/request.py", line 531, in open
    response = meth(req, response)
  File "/usr/lib/python3.8/urllib/request.py", line 640, in http_response
    response = self.parent.error(
  File "/usr/lib/python3.8/urllib/request.py", line 563, in error
    result = self._call_chain(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 502, in _call_chain
    result = func(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 755, in http_error_302
    return self.parent.open(new, timeout=req.timeout)
  File "/usr/lib/python3.8/urllib/request.py", line 531, in open
    response = meth(req, response)
  File "/usr/lib/python3.8/urllib/request.py", line 640, in http_response
    response = self.parent.error(
  File "/usr/lib/python3.8/urllib/request.py", line 569, in error
    return self._call_chain(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 502, in _call_chain
    result = func(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 649, in http_error_default
    raise HTTPError(req.full_url, code, msg, hdrs, fp)
urllib.error.HTTPError: HTTP Error 502: Bad Gateway
ERROR:root:Invalid url for classgraph-4.8.104: http://opensource.org/licenses/MIT after 9 retries.
ERROR:root:['slf4j-api-1.7.30', 'slf4j-simple-1.7.30', 'kafka-1.16.3', 'classgraph-4.8.104']
ERROR:root:**************************************** Licenses were not able to be pulled automatically for some dependencies. Please search source code of the dependencies on the internet and add "license" and "notice" (if available) field to <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/license_scripts/dep_urls_java.yaml> for each missing license. Dependency List: [classgraph-4.8.104,kafka-1.16.3,slf4j-api-1.7.30,slf4j-simple-1.7.30]
INFO:root:pull_licenses_java.py failed. It took 373.621474 seconds with 16 threads.
Traceback (most recent call last):
  File "<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/license_scripts/pull_licenses_java.py",> line 321, in <module>
    raise RuntimeError('{n} error(s) occurred.'.format(n=len(error_msg)),
RuntimeError: ('1 error(s) occurred.', ['**************************************** Licenses were not able to be pulled automatically for some dependencies. Please search source code of the dependencies on the internet and add "license" and "notice" (if available) field to <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/license_scripts/dep_urls_java.yaml> for each missing license. Dependency List: [classgraph-4.8.104,kafka-1.16.3,slf4j-api-1.7.30,slf4j-simple-1.7.30]'])

> Task :sdks:java:container:pullLicenses FAILED

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:container:pullLicenses'.
> Process 'command './license_scripts/license_script.sh'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 6m 51s
104 actionable tasks: 67 executed, 33 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/7apywnizdf2lq

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #113

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/113/display/redirect?page=changes>

Changes:

[kamil.bregula] [BEAM-14215] Improve argument validation in SnowflakeIO

[benjamin.gonzalez] [BEAM-14013] Add PreCommit Kotlin examples Jenkins Job

[Andrew Pilloud] [BEAM-13151] Support multiple layers of AutoValue nesting

[Heejong Lee] [BEAM-14233] Merge requirements from expanded response for Java External

[benjamin.gonzalez] [BEAM-14013] Add spark, direct, flink runners as triggers for Kotlin

[noreply] [BEAM-13898] Add tests to the pubsubx package. (#17324)

[noreply] [BEAM-14285] Clean up Staticcheck Warnings in io packages (#17336)

[noreply] [BEAM-14187] Fix concurrency issue in IsmReaderImpl (#17201)

[noreply] [BEAM-14288] Skip flaking test

[noreply] Simplify specifying additional dependencies in Go SDK in XLang IOs

[noreply] [BEAM-14240] Clean staticcheck warnings in runner packages (#17340)

[Daniel Oliveira] [BEAM-13538] Workaround to fix go-licenses crash.


------------------------------------------
[...truncated 556.31 KB...]
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_pro
Apr 12, 2022 4:00:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-12T16:00:40.629Z: Cancel request is committed for workflow job: 2022-04-12_05_32_38-16584136735451216641.
Apr 12, 2022 4:00:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-12T16:00:40.690Z: Cleaning up.
Apr 12, 2022 4:00:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-12T16:00:40.763Z: Stopping **** pool...
Apr 12, 2022 4:00:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-12T16:00:40.832Z: Stopping **** pool...
Apr 12, 2022 4:01:33 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-12T16:01:33.548Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Apr 12, 2022 4:01:33 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-12T16:01:33.651Z: Worker pool stopped.
Apr 12, 2022 4:01:40 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-04-12_05_32_38-16584136735451216641 finished with status CANCELLED.
Load test results for test (ID): f37135c9-cf31-48db-840a-2d9fcc7df2c6 and timestamp: 2022-04-12T12:32:32.859000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                 12307.586
dataflow_v2_java17_total_bytes_count             2.85776675E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220412123038
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:0297880347919b8404a0e4f2e0710b1d062f0edb1384b21986d0e776bbc6aa3f
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220412123038]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:0297880347919b8404a0e4f2e0710b1d062f0edb1384b21986d0e776bbc6aa3f]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220412123038] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:0297880347919b8404a0e4f2e0710b1d062f0edb1384b21986d0e776bbc6aa3f])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:0297880347919b8404a0e4f2e0710b1d062f0edb1384b21986d0e776bbc6aa3f
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:0297880347919b8404a0e4f2e0710b1d062f0edb1384b21986d0e776bbc6aa3f
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:0297880347919b8404a0e4f2e0710b1d062f0edb1384b21986d0e776bbc6aa3f].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 31m 22s
110 actionable tasks: 73 executed, 33 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/2huuoa7qg7lim

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #112

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/112/display/redirect>

Changes:


------------------------------------------
[...truncated 853.62 KB...]
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmi
Apr 11, 2022 4:01:39 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-11T16:01:39.528Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Apr 11, 2022 4:01:39 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-11T16:01:39.598Z: Worker pool stopped.
Apr 11, 2022 4:01:47 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-04-11_05_32_40-13881262804693704859 finished with status CANCELLED.
Load test results for test (ID): 3555057e-2d48-4bc3-ac06-f1e42cd3fd98 and timestamp: 2022-04-11T12:32:33.996000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                 12364.825
dataflow_v2_java17_total_bytes_count             2.65637155E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220411123036
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:0a4f249317af477039f94a21c8d09d2748f9a67a49e53e178a857fa36329751d
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220411123036]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:0a4f249317af477039f94a21c8d09d2748f9a67a49e53e178a857fa36329751d]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220411123036] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:0a4f249317af477039f94a21c8d09d2748f9a67a49e53e178a857fa36329751d])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:0a4f249317af477039f94a21c8d09d2748f9a67a49e53e178a857fa36329751d
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:0a4f249317af477039f94a21c8d09d2748f9a67a49e53e178a857fa36329751d
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:0a4f249317af477039f94a21c8d09d2748f9a67a49e53e178a857fa36329751d].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 31m 30s
110 actionable tasks: 73 executed, 33 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/jxqpntg3gnezk

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #111

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/111/display/redirect?page=changes>

Changes:

[chamikaramj] Re-raise exceptions swallowed in several Python I/O connectors

[noreply] Merge pull request #16928: [BEAM-11971] Re add reverted timer


------------------------------------------
[...truncated 85.92 KB...]
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
Apr 10, 2022 12:38:51 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
SEVERE: 2022-04-10T12:38:51.118Z: generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
Apr 10, 2022 4:00:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-10T16:00:41.808Z: Cancel request is committed for workflow job: 2022-04-10_05_32_58-12731610135292703172.
Apr 10, 2022 4:00:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-10T16:00:42.282Z: Cleaning up.
Apr 10, 2022 4:00:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-10T16:00:42.603Z: Stopping **** pool...
Apr 10, 2022 4:00:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-10T16:00:42.647Z: Stopping **** pool...
Apr 10, 2022 4:01:40 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-10T16:01:37.779Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Apr 10, 2022 4:01:40 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-10T16:01:37.815Z: Worker pool stopped.
Apr 10, 2022 4:01:43 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-04-10_05_32_58-12731610135292703172 finished with status CANCELLED.
Load test results for test (ID): e587a666-2285-4158-aa53-0f1d12ccb15d and timestamp: 2022-04-10T12:32:53.321000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                 12330.282
dataflow_v2_java17_total_bytes_count             4.13022115E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220410123035
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:72a303835871089201e801a436c8ee2eb92fc6c2dda32997e97e694aa7e1d69b
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220410123035]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:72a303835871089201e801a436c8ee2eb92fc6c2dda32997e97e694aa7e1d69b]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220410123035] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:72a303835871089201e801a436c8ee2eb92fc6c2dda32997e97e694aa7e1d69b])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:72a303835871089201e801a436c8ee2eb92fc6c2dda32997e97e694aa7e1d69b
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:72a303835871089201e801a436c8ee2eb92fc6c2dda32997e97e694aa7e1d69b
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:72a303835871089201e801a436c8ee2eb92fc6c2dda32997e97e694aa7e1d69b].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 31m 27s
110 actionable tasks: 73 executed, 33 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/xlrh2zgdmm7ii

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #110

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/110/display/redirect?page=changes>

Changes:

[benjamin.gonzalez] [BEAM-11714] Change spotBugs jenkins config

[Robert Bradshaw] Cleanup docs on Shared.

[Kyle Weaver] Nit: correct description for precommit cron jobs.

[benjamin.gonzalez] [BEAM-11714] Add dummy class for testing

[benjamin.gonzalez] [BEAM-11714] Remove dummy class used for testing

[benjamin.gonzalez] [BEAM-11714] Spotbugs print toJenkins UI precommit_Java17

[noreply] [BEAM-13767] Remove eclipse plugin as it generates a lot of unused tasks

[noreply] [BEAM-10708] Updated beam_sql error message (#17314)

[noreply] [BEAM-14281] add as_deterministic_coder to nullable coder (#17322)

[noreply] Improvements to Beam/Spark quickstart. (#17129)

[chamikaramj] Disable BigQueryIOStorageWriteIT for Runner v2 test suite


------------------------------------------
[...truncated 376.46 KB...]
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
Apr 09, 2022 4:00:35 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-09T16:00:34.827Z: Cancel request is committed for workflow job: 2022-04-09_05_32_53-12302577311782903590.
Apr 09, 2022 4:00:35 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-09T16:00:34.904Z: Cleaning up.
Apr 09, 2022 4:00:35 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-09T16:00:34.982Z: Stopping **** pool...
Apr 09, 2022 4:00:35 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-09T16:00:35.025Z: Stopping **** pool...
Apr 09, 2022 4:01:37 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-09T16:01:35.116Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Apr 09, 2022 4:01:37 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-09T16:01:35.184Z: Worker pool stopped.
Apr 09, 2022 4:01:41 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-04-09_05_32_53-12302577311782903590 finished with status CANCELLED.
Load test results for test (ID): 08c559da-265c-438e-8b4d-f608d8c83deb and timestamp: 2022-04-09T12:32:40.587000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                 12320.833
dataflow_v2_java17_total_bytes_count             4.00156759E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220409123043
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:df5bb150e9ffb293214f67ec0c73b3ecc1e326f7adf9a974bcfecfcfeec7d625
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220409123043]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:df5bb150e9ffb293214f67ec0c73b3ecc1e326f7adf9a974bcfecfcfeec7d625]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220409123043] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:df5bb150e9ffb293214f67ec0c73b3ecc1e326f7adf9a974bcfecfcfeec7d625])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:df5bb150e9ffb293214f67ec0c73b3ecc1e326f7adf9a974bcfecfcfeec7d625
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:df5bb150e9ffb293214f67ec0c73b3ecc1e326f7adf9a974bcfecfcfeec7d625
ERROR: (gcloud.container.images.delete) Not found: response: {'docker-distribution-api-version': 'registry/2.0', 'content-type': 'application/json', 'date': 'Sat, 09 Apr 2022 16:01:48 GMT', 'server': 'Docker Registry', 'cache-control': 'private', 'x-xss-protection': '0', 'x-frame-options': 'SAMEORIGIN', 'transfer-encoding': 'chunked', 'status': '404', 'content-length': '168', '-content-encoding': 'gzip'}
Failed to compute blob liveness for manifest: 'sha256:df5bb150e9ffb293214f67ec0c73b3ecc1e326f7adf9a974bcfecfcfeec7d625': None

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages FAILED

FAILURE: Build completed with 2 failures.

1: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.
==============================================================================

2: Task failed with an exception.
-----------
* Where:
Build file '<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/runners/google-cloud-dataflow-java/build.gradle'> line: 297

* What went wrong:
Execution failed for task ':runners:google-cloud-dataflow-java:cleanUpDockerJavaImages'.
> Process 'command './scripts/cleanup_untagged_gcr_images.sh'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.
==============================================================================

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 31m 22s
110 actionable tasks: 74 executed, 32 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/wmmfpbdzoqmja

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 - Build # 109 - Aborted!

Posted by Apache Jenkins Server <je...@builds.apache.org>.
beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 - Build # 109 - Aborted:

Check console output at https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/109/ to view the results.

Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #108

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/108/display/redirect?page=changes>

Changes:

[noreply] Avoid pr-bot state desync (#17299)

[noreply] [BEAM-14259] Clean up staticcheck warnings in the exec package (#17285)

[noreply] Minor: Prefer registered schema in SQL docs (#17298)

[Kyle Weaver] [BEAM-14262] Update plugins for Dockerized Jenkins.

[Kyle Weaver] Add ansicolor and ws-cleanup plugins.

[noreply] [Playground] add meta tags (#17207)

[noreply] fixes golint and deprecated issues in recent Go SDK import (#17304)

[noreply] [BEAM-14266] Replace deprecated ptypes package uses (#17302)

[noreply] [BEAM-11936] Fix rawtypes warnings in SnowflakeIO (#17257)

[noreply] Merge pull request #17262: [BEAM-14244] Use the supplied output


------------------------------------------
[...truncated 90.68 KB...]
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
Apr 07, 2022 12:42:02 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
SEVERE: 2022-04-07T12:42:00.955Z: generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
Apr 07, 2022 4:00:47 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-07T16:00:46.424Z: Cancel request is committed for workflow job: 2022-04-07_05_32_40-9479556150086241544.
Apr 07, 2022 4:00:47 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-07T16:00:46.508Z: Cleaning up.
Apr 07, 2022 4:00:47 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-07T16:00:46.593Z: Stopping **** pool...
Apr 07, 2022 4:00:47 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-07T16:00:46.653Z: Stopping **** pool...
Apr 07, 2022 4:01:44 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-07T16:01:43.819Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Apr 07, 2022 4:01:44 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-07T16:01:43.862Z: Worker pool stopped.
Apr 07, 2022 4:01:49 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-04-07_05_32_40-9479556150086241544 finished with status CANCELLED.
Load test results for test (ID): 36061655-0445-4a2e-8ace-0a89d2cb57c9 and timestamp: 2022-04-07T12:32:34.775000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                 12340.917
dataflow_v2_java17_total_bytes_count             2.70217214E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220407123036
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:ef93a3b6b99df0cbecc4e8d9dd63fda639570d96e4ee4cb9d7a37ad26e93235c
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220407123036]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:ef93a3b6b99df0cbecc4e8d9dd63fda639570d96e4ee4cb9d7a37ad26e93235c]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220407123036] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:ef93a3b6b99df0cbecc4e8d9dd63fda639570d96e4ee4cb9d7a37ad26e93235c])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:ef93a3b6b99df0cbecc4e8d9dd63fda639570d96e4ee4cb9d7a37ad26e93235c
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:ef93a3b6b99df0cbecc4e8d9dd63fda639570d96e4ee4cb9d7a37ad26e93235c
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:ef93a3b6b99df0cbecc4e8d9dd63fda639570d96e4ee4cb9d7a37ad26e93235c].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 31m 34s
110 actionable tasks: 73 executed, 33 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/cos6vkydokaz4

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #107

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/107/display/redirect?page=changes>

Changes:

[bingyeli] update query

[Robert Bradshaw] [BEAM-14250] Fix coder registration for types defined in __main__.

[johnjcasey] [BEAM-14256] update SpEL dependency to 5.3.18.RELEASE

[johnjcasey] [BEAM-14256] remove .RELEASE

[dannymccormick] Fix dependency issue causing failures

[Kyle Weaver] [BEAM-9649] Add region option to Mongo Dataflow test.

[noreply] Allow get_coder(None).

[noreply] [BEAM-13015] Disable retries for fnapi grpc channels which otherwise

[noreply] [BEAM-13952] Sickbay

[noreply] BEAM-14235 parquetio module does not parse PEP-440 compliant Pyarrow

[noreply] [Website] Contribution guide page indent bug fix (#17287)

[noreply] [BEAM-10976] Document go sdk bundle finalization (#17048)

[noreply] [BEAM-13829] Expose status API from Go SDK Harness (#16957)


------------------------------------------
[...truncated 636.76 KB...]
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmi
Apr 06, 2022 4:02:09 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-06T16:02:07.953Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Apr 06, 2022 4:02:09 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-06T16:02:08.003Z: Worker pool stopped.
Apr 06, 2022 4:02:18 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-04-06_05_34_14-17492783050096433347 finished with status CANCELLED.
Load test results for test (ID): 7c5381e5-9149-4612-9a16-feaa4e40361b and timestamp: 2022-04-06T12:34:08.158000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                 12270.523
dataflow_v2_java17_total_bytes_count             2.59884868E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220406123036
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:5233ac8e4bfedcb4cff524cb83131c0b1e4fba7448433c636ca8b5ddbd966942
ERROR: (gcloud.container.images.untag) Image could not be found: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220406123036]

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages FAILED

FAILURE: Build completed with 2 failures.

1: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.
==============================================================================

2: Task failed with an exception.
-----------
* Where:
Build file '<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/runners/google-cloud-dataflow-java/build.gradle'> line: 294

* What went wrong:
Execution failed for task ':runners:google-cloud-dataflow-java:cleanUpDockerJavaImages'.
> Process 'command 'gcloud'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.
==============================================================================

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 31m 57s
110 actionable tasks: 74 executed, 32 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/3pof7unckot2g

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #106

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/106/display/redirect?page=changes>

Changes:

[benjamin.gonzalez] [BEAM-8970] Add docs to run wordcount example on portable Spark Runner

[Kiley Sok] Update python container version

[benjamin.gonzalez] [BEAM-8970] Add period to end of sentence

[Kyle Weaver] Add self-descriptive message for expected errors.

[noreply] Add --dataflowServiceOptions=enable_prime to useUnifiedWorker conditions

[noreply] [BEAM-10529] nullable xlang coder (#16923)

[noreply] Fix go fmt break in core/typex/special.go (#17266)

[noreply] [BEAM-5436] Add doc page on Go cross compilation. (#17256)

[noreply] Pr-bot Don't count all reviews as approvals (#17269)

[noreply] Fix postcommits (#17263)

[noreply] [BEAM-14241] Address staticcheck warnings in boot.go (#17264)

[noreply] [BEAM-14157] GrpcWindmillServer: Use stream specific boolean to do

[noreply] [BEAM-10582] Allow (and test) pyarrow 7 (#17229)

[noreply] [BEAM-13519] Solve race issues when the server responds with an error


------------------------------------------
[...truncated 754.22 KB...]
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/d
Apr 05, 2022 4:00:58 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-05T16:00:57.319Z: Cancel request is committed for workflow job: 2022-04-05_05_32_51-3666642527379357457.
Apr 05, 2022 4:00:58 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-05T16:00:57.444Z: Cleaning up.
Apr 05, 2022 4:00:58 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-05T16:00:57.512Z: Stopping **** pool...
Apr 05, 2022 4:00:58 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-05T16:00:57.572Z: Stopping **** pool...
Apr 05, 2022 4:01:57 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-05T16:01:56.238Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Apr 05, 2022 4:01:57 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-05T16:01:56.273Z: Worker pool stopped.
Apr 05, 2022 4:02:02 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-04-05_05_32_51-3666642527379357457 finished with status CANCELLED.
Load test results for test (ID): 61edf798-1b2b-4693-ac07-01cede3f0333 and timestamp: 2022-04-05T12:32:46.389000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                  12341.89
dataflow_v2_java17_total_bytes_count             3.73281742E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220405123037
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:ef2e32b5e5b62e6653188e891ea43a7f16a95758d2a2a60ca5a5bf9a17408f92
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220405123037]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:ef2e32b5e5b62e6653188e891ea43a7f16a95758d2a2a60ca5a5bf9a17408f92]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220405123037] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:ef2e32b5e5b62e6653188e891ea43a7f16a95758d2a2a60ca5a5bf9a17408f92])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:ef2e32b5e5b62e6653188e891ea43a7f16a95758d2a2a60ca5a5bf9a17408f92
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:ef2e32b5e5b62e6653188e891ea43a7f16a95758d2a2a60ca5a5bf9a17408f92
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:ef2e32b5e5b62e6653188e891ea43a7f16a95758d2a2a60ca5a5bf9a17408f92].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 32m 4s
110 actionable tasks: 73 executed, 33 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/6e3fmf37nc6ek

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #105

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/105/display/redirect>

Changes:


------------------------------------------
[...truncated 458.36 KB...]
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/strea
Apr 04, 2022 4:03:21 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-04T16:03:21.050Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Apr 04, 2022 4:03:21 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-04-04T16:03:21.112Z: Worker pool stopped.
Apr 04, 2022 4:03:28 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-04-04_05_32_41-386720858990461328 finished with status CANCELLED.
Load test results for test (ID): 2383320e-7504-40a1-ad91-cb6e77e8b3e2 and timestamp: 2022-04-04T12:32:35.986000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                 12339.642
dataflow_v2_java17_total_bytes_count             4.13967225E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220404123037
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:45c5397c5fa9a9de944159d473d9f56fc2f54a4148a3c72764fddfe40e2d0a3e
Deleted: sha256:f38cc01db745182ff7a12a33f7d29a94450e7955475e328dbfd5d6679df2632e
Deleted: sha256:073336a4ab9c6ac10620325143a6a21cd777bd14844d213ea25473db6ab36516
Deleted: sha256:4fe965c2eae1d1cc36d550e0ca852eeb3d6bc4eaf65fb308a84da3841e726169
Deleted: sha256:283e8a8e965c8042cbbeb09d5447041b9a412c1c466192098f579946d3ae0275
Deleted: sha256:4c7d9132b334603c78570b9394031051f9a2da5bb8ddd534c8af8976eb736425
Deleted: sha256:393cb90935a2dfd2ccfd37144d81ca63e2b5336f326e77cc39c300a0dc219b99
Deleted: sha256:ec5d3fea0ae0c7b9fd52a3ccb56eed6cb0b005933c4b1b5bf5cfb343e808d5c9
Deleted: sha256:4b43eebb0f4fa7225f7e64ce89bf2af024b9d52e0d305232224111caa9d3be1b
Deleted: sha256:ce5d7e50f946084f9c3275530cbb6925429e2680d70801ab7d4c8c4f7eb75167
Deleted: sha256:e2d9238788a94ae0b3b31fd7050972e7296b293881ee1e1a83462373ce3cc522
Deleted: sha256:d6c833b54a19c4bc44f94a20259bfa11e72935f683b20f362135af2b055dd32d
Deleted: sha256:ab584f8cc7c000a513240b6ccb35f88a83a2e66acf5fd862185bf354bd83d941
Deleted: sha256:3a7378f2f2a5bfd4a21b203280a311ec06441366a9872beb37e22300f62dde04
Deleted: sha256:0673b74f8cc304eacb3a5ccd2061ac5ce9e7e47b1cf287163a4f88e360f65b47
Deleted: sha256:de2eb27e40434e484c7d44348ba38c33d8bab9b69e8579d1b427d162850794ad
Deleted: sha256:d684a28793d0c737a5dd9630ffd9b736c611776f7343bdde4a0d33b6033256aa
Deleted: sha256:7a2e75438ed419d51bd23de46fccfd3510b05f98ee8445ef09d8060990bbf5e7
Deleted: sha256:f4e8a583db4780006d08aaf5430a075314cb384a1a670cabfae26badce7332b9
Deleted: sha256:9d847d2693a2ff84a4ab706e3aebaabad342902028512191fb08eeb3848178fc
Deleted: sha256:77fb90446d06ecf3c9fb2668059da8ac083ac72061a1940e6fe7f44b202deb06
Deleted: sha256:3fb64211c2f47dc0cf9aa4fe9f87a4b967c7c01f5fa60313cb84bbd104debc9e
Deleted: sha256:f7f6b7b8fe430f0221ce5cbdffbad04f6f7ea5707cc57a58b2e633909e341f48
Deleted: sha256:c45fc619af0b548f520a2fbcc8feb4db24c79b21f4880bfd490cdc6891cc1ead
Deleted: sha256:4e401894ffb4918a306baeca124fcf46f8e3cbf56a2bcbc326dfac281f154f9a
Deleted: sha256:c37c8a4d65d03507aaa3c88508270d1f290ed91b7a46e33f05bb4b6715b7b610
Deleted: sha256:5508bea2f0be5b66f54591f22053e73883d51fdf4834de963e3e8dab94534052
Deleted: sha256:6778903f746a88bf72c1443c0d8634a82a3634ac73c94fac6102f1e388bfa57e
Deleted: sha256:be70fcf12cfee61004019b2438f91712281f81c250e89035b68b8169726a0fe9
Deleted: sha256:8cbb3ead3217b35e9217e8041c7ad34e59350c1b49da667855d2fc04b9969efe
Deleted: sha256:e0ae11ea9dbfa3398c7555362a8726f00e02b7e79a32cdab5d3d49fa9d902449
Deleted: sha256:4e6578a82049418c5e65286a1896da15b1689fa95f7bda334942339922bf5f53
Deleted: sha256:f1b27c2e719be61ea47e42183321baab4a64a048db130414c3f337f34ea9ca7e
Deleted: sha256:702d6242cc7552eb938a97b9064a875d4db995eb546216e74945a8844bf3388a
Deleted: sha256:cc61c28c05d3668e15b2c408ab0a939a955f7aabdd21f01c60fcec93456b9c08
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220404123037]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:45c5397c5fa9a9de944159d473d9f56fc2f54a4148a3c72764fddfe40e2d0a3e]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220404123037] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:45c5397c5fa9a9de944159d473d9f56fc2f54a4148a3c72764fddfe40e2d0a3e])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:45c5397c5fa9a9de944159d473d9f56fc2f54a4148a3c72764fddfe40e2d0a3e
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:45c5397c5fa9a9de944159d473d9f56fc2f54a4148a3c72764fddfe40e2d0a3e
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:45c5397c5fa9a9de944159d473d9f56fc2f54a4148a3c72764fddfe40e2d0a3e].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 33m 12s
110 actionable tasks: 73 executed, 33 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/6aumrwiig2kai

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 - Build # 104 - Aborted!

Posted by Apache Jenkins Server <je...@builds.apache.org>.
beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 - Build # 104 - Aborted:

Check console output at https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/104/ to view the results.

beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 - Build # 103 - Aborted!

Posted by Apache Jenkins Server <je...@builds.apache.org>.
beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 - Build # 103 - Aborted:

Check console output at https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/103/ to view the results.

beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 - Build # 102 - Aborted!

Posted by Apache Jenkins Server <je...@builds.apache.org>.
beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 - Build # 102 - Aborted:

Check console output at https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/102/ to view the results.

beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 - Build # 101 - Aborted!

Posted by Apache Jenkins Server <je...@builds.apache.org>.
beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 - Build # 101 - Aborted:

Check console output at https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/101/ to view the results.

Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #100

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/100/display/redirect?page=changes>

Changes:

[vachan] Update display data to include BQ information.

[noreply] Revert "[BEAM-14084] iterable_input_value_types changed from list to

[chamikaramj] Convert URLs to local jars when constructing filesToStage

[Valentyn Tymofieiev] Ensure the removed option prebuild_sdk_container_base_image not used on

[noreply] [BEAM-13314]Revise recommendations to manage Python pipeline

[noreply] Merge pull request #17202 from [BEAM-14194]: Disallow autoscaling for

[noreply] Merge pull request #17080 from [BEAM-13880] [Playground] Increase test

[noreply] Merge pull request #17050 from [BEAM-13877] [Playground] Increase test

[noreply] [BEAM-14200] Improve SamzaJobInvoker extensibility (#17212)

[noreply] Merge pull request #17148 from [BEAM-14042] [playground] Scroll imports

[noreply] [BEAM-13918] Increase datastoreio go sdk unit test coverage (#17173)

[noreply] Merge pull request #16819: [BEAM-13806] Adding test suite for Go x-lang


------------------------------------------
[...truncated 652.11 KB...]
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:706
==>
    dist_proc/dax/workf
Mar 30, 2022 4:00:44 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-30T16:00:43.739Z: Cancel request is committed for workflow job: 2022-03-30_05_35_27-15656292221588933800.
Mar 30, 2022 4:00:44 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-30T16:00:43.808Z: Cleaning up.
Mar 30, 2022 4:00:44 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-30T16:00:43.953Z: Stopping **** pool...
Mar 30, 2022 4:00:44 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-30T16:00:44.008Z: Stopping **** pool...
Mar 30, 2022 4:03:17 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-30T16:03:14.772Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Mar 30, 2022 4:03:17 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-30T16:03:14.850Z: Worker pool stopped.
Mar 30, 2022 4:03:20 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-03-30_05_35_27-15656292221588933800 finished with status CANCELLED.
Load test results for test (ID): 4b3f459a-dd4b-4d6b-b75a-025a680163b1 and timestamp: 2022-03-30T12:35:21.642000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                 12183.626
dataflow_v2_java17_total_bytes_count             3.28077625E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220330123257
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:13907352a71480629641f00abc705583ef645ae17d526a6741011f9942ec9435
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220330123257]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:13907352a71480629641f00abc705583ef645ae17d526a6741011f9942ec9435]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220330123257] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:13907352a71480629641f00abc705583ef645ae17d526a6741011f9942ec9435])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:13907352a71480629641f00abc705583ef645ae17d526a6741011f9942ec9435
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:13907352a71480629641f00abc705583ef645ae17d526a6741011f9942ec9435
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:13907352a71480629641f00abc705583ef645ae17d526a6741011f9942ec9435].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 30m 45s
110 actionable tasks: 74 executed, 32 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/3irgbggo3nnqu

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #99

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/99/display/redirect?page=changes>

Changes:

[hengfeng] feat: remove the metadata table after the pipeline finishes

[thiagotnunes] test: add test for metadata table dropping

[noreply] Minor: Add warning about pubsub client to Beam 2.36.0 blog (#17188)

[noreply] [BEAM-14177] Fix GBK re-iteration caching for portable runners. (#17184)

[noreply] Merge pull request #17187: [BEAM-14181] Make sure to evict connections

[noreply] Only reset transform.label if it is correctly assigned (#17192)

[noreply] [BEAM-12641] Use google-auth instead of oauth2client for GCP auth

[Robert Bradshaw] [BEAM-14163] Fix typo in single core per container logic.

[thiagotnunes] test: disable SpannerIO.readChangeStream test

[noreply] Merge pull request #17164 from [BEAM-14140][Playground] Fix Deploy

[noreply] Merge pull request #16855 from [BEAM-13938][Playground] Increase test


------------------------------------------
[...truncated 83.41 KB...]
    return self.do_open(http.client.HTTPConnection, req)
  File "/usr/lib/python3.8/urllib/request.py", line 1357, in do_open
    raise URLError(err)
urllib.error.URLError: <urlopen error [Errno 101] Network is unreachable>
ERROR:root:Invalid url for jFormatString-3.0.0: http://www.gnu.org/licenses/lgpl.html. Retrying...
Traceback (most recent call last):
  File "/usr/lib/python3.8/urllib/request.py", line 1354, in do_open
    h.request(req.get_method(), req.selector, req.data, headers,
  File "/usr/lib/python3.8/http/client.py", line 1256, in request
    self._send_request(method, url, body, headers, encode_chunked)
  File "/usr/lib/python3.8/http/client.py", line 1302, in _send_request
    self.endheaders(body, encode_chunked=encode_chunked)
  File "/usr/lib/python3.8/http/client.py", line 1251, in endheaders
    self._send_output(message_body, encode_chunked=encode_chunked)
  File "/usr/lib/python3.8/http/client.py", line 1011, in _send_output
    self.send(msg)
  File "/usr/lib/python3.8/http/client.py", line 951, in send
    self.connect()
  File "/usr/lib/python3.8/http/client.py", line 1418, in connect
    super().connect()
  File "/usr/lib/python3.8/http/client.py", line 922, in connect
    self.sock = self._create_connection(
  File "/usr/lib/python3.8/socket.py", line 808, in create_connection
    raise err
  File "/usr/lib/python3.8/socket.py", line 796, in create_connection
    sock.connect(sa)
OSError: [Errno 101] Network is unreachable

During handling of the above exception, another exception occurred:

Traceback (most recent call last):
  File "<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/license_scripts/pull_licenses_java.py",> line 58, in pull_from_url
    url_read = urlopen(url)
  File "/usr/lib/python3.8/urllib/request.py", line 222, in urlopen
    return opener.open(url, data, timeout)
  File "/usr/lib/python3.8/urllib/request.py", line 525, in open
    response = self._open(req, data)
  File "/usr/lib/python3.8/urllib/request.py", line 542, in _open
    result = self._call_chain(self.handle_open, protocol, protocol +
  File "/usr/lib/python3.8/urllib/request.py", line 502, in _call_chain
    result = func(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 1397, in https_open
    return self.do_open(http.client.HTTPSConnection, req,
  File "/usr/lib/python3.8/urllib/request.py", line 1357, in do_open
    raise URLError(err)
urllib.error.URLError: <urlopen error [Errno 101] Network is unreachable>
ERROR:root:Invalid url for spotbugs-annotations-4.0.6: https://www.gnu.org/licenses/old-licenses/lgpl-2.1.en.html. Retrying...
Traceback (most recent call last):
  File "/usr/lib/python3.8/urllib/request.py", line 1354, in do_open
    h.request(req.get_method(), req.selector, req.data, headers,
  File "/usr/lib/python3.8/http/client.py", line 1256, in request
    self._send_request(method, url, body, headers, encode_chunked)
  File "/usr/lib/python3.8/http/client.py", line 1302, in _send_request
    self.endheaders(body, encode_chunked=encode_chunked)
  File "/usr/lib/python3.8/http/client.py", line 1251, in endheaders
    self._send_output(message_body, encode_chunked=encode_chunked)
  File "/usr/lib/python3.8/http/client.py", line 1011, in _send_output
    self.send(msg)
  File "/usr/lib/python3.8/http/client.py", line 951, in send
    self.connect()
  File "/usr/lib/python3.8/http/client.py", line 922, in connect
    self.sock = self._create_connection(
  File "/usr/lib/python3.8/socket.py", line 808, in create_connection
    raise err
  File "/usr/lib/python3.8/socket.py", line 796, in create_connection
    sock.connect(sa)
OSError: [Errno 101] Network is unreachable

During handling of the above exception, another exception occurred:

Traceback (most recent call last):
  File "<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/license_scripts/pull_licenses_java.py",> line 58, in pull_from_url
    url_read = urlopen(url)
  File "/usr/lib/python3.8/urllib/request.py", line 222, in urlopen
    return opener.open(url, data, timeout)
  File "/usr/lib/python3.8/urllib/request.py", line 525, in open
    response = self._open(req, data)
  File "/usr/lib/python3.8/urllib/request.py", line 542, in _open
    result = self._call_chain(self.handle_open, protocol, protocol +
  File "/usr/lib/python3.8/urllib/request.py", line 502, in _call_chain
    result = func(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 1383, in http_open
    return self.do_open(http.client.HTTPConnection, req)
  File "/usr/lib/python3.8/urllib/request.py", line 1357, in do_open
    raise URLError(err)
urllib.error.URLError: <urlopen error [Errno 101] Network is unreachable>
ERROR:root:Invalid url for checkstyle-8.23: http://www.gnu.org/licenses/old-licenses/lgpl-2.1.txt after 9 retries.
Traceback (most recent call last):
  File "/usr/lib/python3.8/urllib/request.py", line 1354, in do_open
    h.request(req.get_method(), req.selector, req.data, headers,
  File "/usr/lib/python3.8/http/client.py", line 1256, in request
    self._send_request(method, url, body, headers, encode_chunked)
  File "/usr/lib/python3.8/http/client.py", line 1302, in _send_request
    self.endheaders(body, encode_chunked=encode_chunked)
  File "/usr/lib/python3.8/http/client.py", line 1251, in endheaders
    self._send_output(message_body, encode_chunked=encode_chunked)
  File "/usr/lib/python3.8/http/client.py", line 1011, in _send_output
    self.send(msg)
  File "/usr/lib/python3.8/http/client.py", line 951, in send
    self.connect()
  File "/usr/lib/python3.8/http/client.py", line 922, in connect
    self.sock = self._create_connection(
  File "/usr/lib/python3.8/socket.py", line 808, in create_connection
    raise err
  File "/usr/lib/python3.8/socket.py", line 796, in create_connection
    sock.connect(sa)
OSError: [Errno 101] Network is unreachable

During handling of the above exception, another exception occurred:

Traceback (most recent call last):
  File "<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/license_scripts/pull_licenses_java.py",> line 58, in pull_from_url
    url_read = urlopen(url)
  File "/usr/lib/python3.8/urllib/request.py", line 222, in urlopen
    return opener.open(url, data, timeout)
  File "/usr/lib/python3.8/urllib/request.py", line 525, in open
    response = self._open(req, data)
  File "/usr/lib/python3.8/urllib/request.py", line 542, in _open
    result = self._call_chain(self.handle_open, protocol, protocol +
  File "/usr/lib/python3.8/urllib/request.py", line 502, in _call_chain
    result = func(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 1383, in http_open
    return self.do_open(http.client.HTTPConnection, req)
  File "/usr/lib/python3.8/urllib/request.py", line 1357, in do_open
    raise URLError(err)
urllib.error.URLError: <urlopen error [Errno 101] Network is unreachable>
ERROR:root:Invalid url for jFormatString-3.0.0: http://www.gnu.org/licenses/lgpl.html. Retrying...
Traceback (most recent call last):
  File "/usr/lib/python3.8/urllib/request.py", line 1354, in do_open
    h.request(req.get_method(), req.selector, req.data, headers,
  File "/usr/lib/python3.8/http/client.py", line 1256, in request
    self._send_request(method, url, body, headers, encode_chunked)
  File "/usr/lib/python3.8/http/client.py", line 1302, in _send_request
    self.endheaders(body, encode_chunked=encode_chunked)
  File "/usr/lib/python3.8/http/client.py", line 1251, in endheaders
    self._send_output(message_body, encode_chunked=encode_chunked)
  File "/usr/lib/python3.8/http/client.py", line 1011, in _send_output
    self.send(msg)
  File "/usr/lib/python3.8/http/client.py", line 951, in send
    self.connect()
  File "/usr/lib/python3.8/http/client.py", line 922, in connect
    self.sock = self._create_connection(
  File "/usr/lib/python3.8/socket.py", line 808, in create_connection
    raise err
  File "/usr/lib/python3.8/socket.py", line 796, in create_connection
    sock.connect(sa)
OSError: [Errno 101] Network is unreachable

During handling of the above exception, another exception occurred:

Traceback (most recent call last):
  File "<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/license_scripts/pull_licenses_java.py",> line 58, in pull_from_url
    url_read = urlopen(url)
  File "/usr/lib/python3.8/urllib/request.py", line 222, in urlopen
    return opener.open(url, data, timeout)
  File "/usr/lib/python3.8/urllib/request.py", line 525, in open
    response = self._open(req, data)
  File "/usr/lib/python3.8/urllib/request.py", line 542, in _open
    result = self._call_chain(self.handle_open, protocol, protocol +
  File "/usr/lib/python3.8/urllib/request.py", line 502, in _call_chain
    result = func(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 1383, in http_open
    return self.do_open(http.client.HTTPConnection, req)
  File "/usr/lib/python3.8/urllib/request.py", line 1357, in do_open
    raise URLError(err)
urllib.error.URLError: <urlopen error [Errno 101] Network is unreachable>
ERROR:root:Invalid url for jFormatString-3.0.0: http://www.gnu.org/licenses/lgpl.html. Retrying...
Traceback (most recent call last):
  File "/usr/lib/python3.8/urllib/request.py", line 1354, in do_open
    h.request(req.get_method(), req.selector, req.data, headers,
  File "/usr/lib/python3.8/http/client.py", line 1256, in request
    self._send_request(method, url, body, headers, encode_chunked)
  File "/usr/lib/python3.8/http/client.py", line 1302, in _send_request
    self.endheaders(body, encode_chunked=encode_chunked)
  File "/usr/lib/python3.8/http/client.py", line 1251, in endheaders
    self._send_output(message_body, encode_chunked=encode_chunked)
  File "/usr/lib/python3.8/http/client.py", line 1011, in _send_output
    self.send(msg)
  File "/usr/lib/python3.8/http/client.py", line 951, in send
    self.connect()
  File "/usr/lib/python3.8/http/client.py", line 1418, in connect
    super().connect()
  File "/usr/lib/python3.8/http/client.py", line 922, in connect
    self.sock = self._create_connection(
  File "/usr/lib/python3.8/socket.py", line 808, in create_connection
    raise err
  File "/usr/lib/python3.8/socket.py", line 796, in create_connection
    sock.connect(sa)
OSError: [Errno 101] Network is unreachable

During handling of the above exception, another exception occurred:

Traceback (most recent call last):
  File "<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/license_scripts/pull_licenses_java.py",> line 58, in pull_from_url
    url_read = urlopen(url)
  File "/usr/lib/python3.8/urllib/request.py", line 222, in urlopen
    return opener.open(url, data, timeout)
  File "/usr/lib/python3.8/urllib/request.py", line 525, in open
    response = self._open(req, data)
  File "/usr/lib/python3.8/urllib/request.py", line 542, in _open
    result = self._call_chain(self.handle_open, protocol, protocol +
  File "/usr/lib/python3.8/urllib/request.py", line 502, in _call_chain
    result = func(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 1397, in https_open
    return self.do_open(http.client.HTTPSConnection, req,
  File "/usr/lib/python3.8/urllib/request.py", line 1357, in do_open
    raise URLError(err)
urllib.error.URLError: <urlopen error [Errno 101] Network is unreachable>
ERROR:root:Invalid url for spotbugs-annotations-4.0.6: https://www.gnu.org/licenses/old-licenses/lgpl-2.1.en.html after 9 retries.
ERROR:root:['checkstyle-8.23', 'spotbugs-annotations-4.0.6']
ERROR:root:**************************************** Licenses were not able to be pulled automatically for some dependencies. Please search source code of the dependencies on the internet and add "license" and "notice" (if available) field to <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/license_scripts/dep_urls_java.yaml> for each missing license. Dependency List: [checkstyle-8.23,spotbugs-annotations-4.0.6]
INFO:root:pull_licenses_java.py failed. It took 614.65573 seconds with 16 threads.
Traceback (most recent call last):
  File "<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/license_scripts/pull_licenses_java.py",> line 321, in <module>
    raise RuntimeError('{n} error(s) occurred.'.format(n=len(error_msg)),
RuntimeError: ('1 error(s) occurred.', ['**************************************** Licenses were not able to be pulled automatically for some dependencies. Please search source code of the dependencies on the internet and add "license" and "notice" (if available) field to <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/license_scripts/dep_urls_java.yaml> for each missing license. Dependency List: [checkstyle-8.23,spotbugs-annotations-4.0.6]'])

> Task :sdks:java:container:pullLicenses FAILED

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:container:pullLicenses'.
> Process 'command './license_scripts/license_script.sh'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 10m 59s
104 actionable tasks: 68 executed, 32 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/grjzz3vlca26s

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #98

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/98/display/redirect>

Changes:


------------------------------------------
[...truncated 1.06 MB...]
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_pro
Mar 28, 2022 4:00:47 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-28T16:00:44.524Z: Cancel request is committed for workflow job: 2022-03-28_05_32_47-3934483608835347313.
Mar 28, 2022 4:00:47 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-28T16:00:44.577Z: Cleaning up.
Mar 28, 2022 4:00:47 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-28T16:00:44.660Z: Stopping **** pool...
Mar 28, 2022 4:00:47 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-28T16:00:44.726Z: Stopping **** pool...
Mar 28, 2022 4:03:03 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-28T16:03:03.814Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Mar 28, 2022 4:03:03 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-28T16:03:03.860Z: Worker pool stopped.
Mar 28, 2022 4:03:10 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-03-28_05_32_47-3934483608835347313 finished with status CANCELLED.
Load test results for test (ID): b8163ec5-dc15-4dd8-a8bc-7d424ebc0f17 and timestamp: 2022-03-28T12:32:41.897000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                 12335.488
dataflow_v2_java17_total_bytes_count             3.63243584E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220328123051
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:02e6e8f84f765194982e755bfd876ac00069dc342f5020007f3e5e549024ec0f
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220328123051]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:02e6e8f84f765194982e755bfd876ac00069dc342f5020007f3e5e549024ec0f]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220328123051] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:02e6e8f84f765194982e755bfd876ac00069dc342f5020007f3e5e549024ec0f])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:02e6e8f84f765194982e755bfd876ac00069dc342f5020007f3e5e549024ec0f
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:02e6e8f84f765194982e755bfd876ac00069dc342f5020007f3e5e549024ec0f
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:02e6e8f84f765194982e755bfd876ac00069dc342f5020007f3e5e549024ec0f].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 32m 39s
110 actionable tasks: 73 executed, 33 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/n2bigukdx2ooi

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #97

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/97/display/redirect>

Changes:


------------------------------------------
[...truncated 943.59 KB...]
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn
Mar 27, 2022 4:00:52 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-27T16:00:51.854Z: Cancel request is committed for workflow job: 2022-03-27_05_33_43-6680142917713648222.
Mar 27, 2022 4:00:52 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-27T16:00:51.885Z: Cleaning up.
Mar 27, 2022 4:00:52 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-27T16:00:51.948Z: Stopping **** pool...
Mar 27, 2022 4:00:52 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-27T16:00:52.001Z: Stopping **** pool...
Mar 27, 2022 4:03:20 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-27T16:03:18.868Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Mar 27, 2022 4:03:20 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-27T16:03:18.903Z: Worker pool stopped.
Mar 27, 2022 4:03:24 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-03-27_05_33_43-6680142917713648222 finished with status CANCELLED.
Load test results for test (ID): 3bd7bf6c-a793-4d51-b8d2-14dcf2454afb and timestamp: 2022-03-27T12:33:38.191000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                 12286.003
dataflow_v2_java17_total_bytes_count             3.09719541E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220327123138
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:398e8dc5b7c4fcef2f8613a6688c05d18e5ef40303bac8aa61d1d61823225305
Deleted: sha256:0b6a1f8f6d10b0fdcfd8168beb4ae49459515164e4e22bb515458b2014dbc320
Deleted: sha256:2f2da1af414622f50f66cccc77fd2d5b5225c1155e95e6bbb29a1e8b71754cea
Deleted: sha256:2893ec7ef62f9a34e5e3b16a174db913d81619deb75c997cc9962c3364ab5235
Deleted: sha256:47a4287f2258ef7811881928ad92a0543ca5fc9869b2a002c392b79bd570ea83
Deleted: sha256:0d552e068042ab4c2c4011d368843d17450d0aef2ae12ff19ddbca4b931416b8
Deleted: sha256:f447e6510d455ff8f1b19dca2f55116bccf6d56e4479b847f38ced228d4aac7a
Deleted: sha256:bf443eb577738c0de5325ad7cb006008630103a4161ee3f4289b24d59169d08a
Deleted: sha256:55a73df2c5e7fd30361112b7db2970c527401b0163edcf8761ad06e8648eb57b
Deleted: sha256:c6522d4317bd13cc6aadd63f952f9faa08865ac6d07e0c42b0b6457ffb48ae99
Deleted: sha256:ab2cd32cb713dc29ccd6130b904a794928397f8aa884b6e0f2e6d842da77d023
Deleted: sha256:831ff308237c401c35a7de0558a583e20de0914153653125d4a96c36d80bc8e3
Deleted: sha256:f12976b467fba819bc5564613211c031749a0325968a41077a4a0c9d4fc0aa94
Deleted: sha256:358c8e07b081c23b024ee94119c48c93e64ddec0edb7fdbeb72cee3cc252624c
Deleted: sha256:36058e360b2eecfffa725576d9213d1e8dd8ac244c47b7db63b417ed5db264a6
Deleted: sha256:b8b19762c81a9cf25e01daabc471f998be8bc134a36f9078581c33cd6867a72f
Deleted: sha256:3481db85a20eea410579cd98875f5cd1905c995ecbb105647b4649b33c2de3b7
Deleted: sha256:62055386f601c9cf2fc0c90dbafd9b22ec312e01c8ba730ff56568d5dfa53516
Deleted: sha256:301303e664510663613767781b6d6171e8fb209556b38b3e1f78b81781eafbff
Deleted: sha256:d65f5f536d6e79e0e41da49d5352792be8ee6897ce23107f6f6280ab088ce454
Deleted: sha256:0aa66a7dc212d23710b8c5fdc03bb48b70f0380b69d1cc94906d39046595d40d
Deleted: sha256:a2b3be430ee2c6b9c25df2f4fba1386a0fdf98f93f40c452ac79dea031715bfd
Deleted: sha256:ae3ff6f50bff18fe57ee5b46f477be3c74c2afc7079f936066b6aca41bfe68dd
Deleted: sha256:041a938f599049bda12a29f5f4b15a29a9005c03b33bce688c77ca13fcf3e5c7
Deleted: sha256:7f9f36aac687364ca6513cf9ed4c3e99e74592b5edf16c5ab8e6a22485d1462d
Deleted: sha256:1004f1e9381697848cab583256376c70a4a071984015245472bd6d1c7d812203
Deleted: sha256:e69fb9dc85ed1c9389795b98a1b2499797ca3ebb2083e2d2c50bd988b47d7500
Deleted: sha256:5695242723b1897464d8eed5f72af2e7c319646865b639cf294d68d35b340288
Deleted: sha256:1007b5179605ab470b9e0898091aa9b2822527240065ba18e8bd94948eb85c03
Deleted: sha256:fff5d503c0f76ce570a28ba5b40eae1a480fdc6cb3f164f8597c72151e381033
Deleted: sha256:372104d8edc54ca3a674ea63ead17c1d40357a6c02657c3ef9604914d675e545
Deleted: sha256:646667eedc6c7276a2441d0fa31936365f38000a73103717dc19605e33254f08
Deleted: sha256:a9aab06405db7673cde97bbf169b83a2535bb9caf899ed862fce7035b82f2c25
Deleted: sha256:39e6fc670d4318d84a24979856816ce60e3345c6bfd3b985b022d15b4fd1d3c4
Deleted: sha256:21962f4c31badbf6efc4fff5b6f735105dd18d2bb484871ae3a6c110da27437e
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220327123138]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:398e8dc5b7c4fcef2f8613a6688c05d18e5ef40303bac8aa61d1d61823225305]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220327123138] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:398e8dc5b7c4fcef2f8613a6688c05d18e5ef40303bac8aa61d1d61823225305])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:398e8dc5b7c4fcef2f8613a6688c05d18e5ef40303bac8aa61d1d61823225305
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:398e8dc5b7c4fcef2f8613a6688c05d18e5ef40303bac8aa61d1d61823225305
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:398e8dc5b7c4fcef2f8613a6688c05d18e5ef40303bac8aa61d1d61823225305].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 32m 8s
110 actionable tasks: 73 executed, 33 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/hhknqxs5qlmka

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #96

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/96/display/redirect?page=changes>

Changes:

[ryanthompson591] iterable_input_value_types will now be an iterable, I don't anticipate

[marco.robles] [BEAM-8218] PulsarIO Connector

[benjamin.gonzalez] [BEAM-12572] Change examples jobs to run as cron jobs

[benjamin.gonzalez] [BEAM-12572] SpotlessApply

[Robert Bradshaw] [BEAM-14171] More explicit asserts in CoGBKResult.

[Robert Bradshaw] Add some comments.

[noreply] [BEAM-14160] Parse filesToStage in Java expansion service (#17167)

[chamikaramj] Mapped JOB_STATE_RESOURCE_CLEANING_UP to RESOURCE_CLEANING_UP in Python

[noreply] Explicitly import estimator from tensorflow (#17168)


------------------------------------------
[...truncated 577.11 KB...]
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
Mar 26, 2022 4:01:05 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-26T16:01:03.919Z: Cancel request is committed for workflow job: 2022-03-26_05_39_20-3837141771591477710.
Mar 26, 2022 4:01:05 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-26T16:01:03.982Z: Cleaning up.
Mar 26, 2022 4:01:05 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-26T16:01:04.078Z: Stopping **** pool...
Mar 26, 2022 4:01:05 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-26T16:01:04.145Z: Stopping **** pool...
Mar 26, 2022 4:03:29 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-26T16:03:27.129Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Mar 26, 2022 4:03:29 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-26T16:03:27.192Z: Worker pool stopped.
Mar 26, 2022 4:03:36 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-03-26_05_39_20-3837141771591477710 finished with status CANCELLED.
Load test results for test (ID): 02c69fc2-b043-4e07-9be0-eed70a2dbff2 and timestamp: 2022-03-26T12:39:15.441000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                 11954.871
dataflow_v2_java17_total_bytes_count             3.20120702E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220326123710
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:f7a5d35fd580d3c14615df776edc0b219e3ac855178c0072bc8b78901af7f25e
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220326123710]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:f7a5d35fd580d3c14615df776edc0b219e3ac855178c0072bc8b78901af7f25e]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220326123710] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:f7a5d35fd580d3c14615df776edc0b219e3ac855178c0072bc8b78901af7f25e])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:f7a5d35fd580d3c14615df776edc0b219e3ac855178c0072bc8b78901af7f25e
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:f7a5d35fd580d3c14615df776edc0b219e3ac855178c0072bc8b78901af7f25e
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:f7a5d35fd580d3c14615df776edc0b219e3ac855178c0072bc8b78901af7f25e].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 26m 47s
110 actionable tasks: 73 executed, 33 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/pbqse3l46mgea

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 - Build # 95 - Aborted!

Posted by Apache Jenkins Server <je...@builds.apache.org>.
beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 - Build # 95 - Aborted:

Check console output at https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/95/ to view the results.

Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #94

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/94/display/redirect?page=changes>

Changes:

[bulat.safiullin] [BEAM-13976] [Website] update homepage

[bulat.safiullin] [BEAM-13976] [Website] update homepage, add logo

[bulat.safiullin] [BEAM-13976] [Website] update text

[bulat.safiullin] [BEAM-13976] [Website] Update Community landing page

[bulat.safiullin] [BEAM-13979] [Website] Update Community/Contact us page

[bulat.safiullin] [BEAM-13979] [Website] update title

[bulat.safiullin] [BEAM-13979] [Website] delete space

[bulat.safiullin] [BEAM-13979] [Website] add Beam Playground

[bulat.safiullin] [BEAM-13976] [Website] delete Beam Playground

[bulat.safiullin] [BEAM-13976] [Website] change navbar css links rules, delete links from

[bulat.safiullin] [BEAM-13977] [Website] delete available-contact-channels on mobile

[bulat.safiullin] [BEAM-13976] [Website] change padding size between the sections

[bulat.safiullin] [BEAM-13976] [Website] change title to capital letters

[bulat.safiullin] [BEAM-13976] [Website] change title

[bulat.safiullin] [BEAM-14040] [Website] create new page, add link

[bulat.safiullin] [BEAM-13977] [Website] change title

[bulat.safiullin] [BEAM-13979] [Website] change text

[bulat.safiullin] [BEAM-13976] [Website] change text

[bulat.safiullin] [BEAM-13977] [Website] change text, add capital letters

[bulat.safiullin] [BEAM-13976] [Website] add playground sass, change text-align

[bulat.safiullin] [BEAM-14040] [Website] add io connectors table

[bulat.safiullin] [BEAM-13976] [Website] add playground section, add empty line

[bulat.safiullin] [BEAM-14040] [Website] add overflow to css, add table content

[bulat.safiullin] [BEAM-14040] [Website] change ✘ for ✔, add license, add br

[bulat.safiullin] [BEAM-14040] [Website] add empty line

[bulat.safiullin] [BEAM-14040] [Website] change td

[bulat.safiullin] [BEAM-14041] [Website] update built io transforms

[bulat.safiullin] [BEAM-14041] [Website] move connectors from Miscellaneous to Database

[bulat.safiullin] [BEAM-14040] [Website] change links color

[danielamartinmtz] Updated metrics' CronJob API to use the latest batch version.

[bulat.safiullin] [BEAM-14041] [Website] change IO from go to java

[bulat.safiullin] [BEAM-14040] [Website] change links, change specific version to current

[danielamartinmtz] Updated cluster to test in metrics-upgrade-clone in BeamMetrics_Publish

[aydar.zaynutdinov] [BEAM-13976][Website]

[aydar.zaynutdinov] [BEAM-14040][Website]

[aydar.zaynutdinov] [BEAM-14041][Website]

[danielamartinmtz] Updated StateFulSet k8s obejct in cassandra-svc-statefulset.yaml file in

[danielamartinmtz] Updated documentation including cluster specs.

[noreply] Beam 13058 k8s apis upgrade - elasticsearch (#18)

[danielamartinmtz] Removed code used for testing.

[danielamartinmtz] Removed code used for testing in job_PostCommit_BeamMetrics_Publish

[noreply] Beam 13058 k8s apis upgrade - Adding Basic Auth details in documentation

[Pablo Estrada] [BEAM-14151] Excluding Spanner CDC tests from Dataflow V1 suite

[danielamartinmtz] Added comments in initContainers and remove unused code in elasticsearch

[noreply] [BEAM-14134] Optimize memory allocations for various core coders

[noreply] [BEAM-14129] Restructure PubsubLiteIO Read side to produce smaller

[noreply] [BEAM-12697] Add primitive field generation from IR to SBE extension

[noreply] [BEAM-13889] Add test cases to jsonx package (#17124)

[noreply] Remove unreachable code in container.go (#17166)

[noreply] Add ability to handle streaming input to AvroSchemaIOProvider (#17126)

[noreply] [BEAM-12898] Flink Load Tests failure- UncheckedExecutionException -

[Daniel Oliveira] Moving to 2.39.0-SNAPSHOT on master branch.


------------------------------------------
[...truncated 736.14 KB...]
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
==>
    dist_proc/dax/workflow/****/streaming/fnapi_streaming_operators.cc:439
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
==>
    dist_proc/dax/workflow/****/streaming/fnapi_streaming_operators.cc:439
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
==>
    dist_proc/dax/workflow/****/streaming/fnapi_streaming_operators.cc:439
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
==>
    dist_proc/dax/workflow/****/streaming/fnapi_streaming_operators.cc:439
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
==>
    dist_proc/dax/workflow/****/streaming/fnapi_streaming_operators.cc:439
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
==>
    dist_proc/dax/workflow/****/streaming/fnapi_streaming_operators.cc:439
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
==>
    dist_proc/dax/workflow/****/streaming/fnapi_streaming_operators.cc:439
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
==>
    dist_proc/dax/workflow/****/streaming/fnapi_streaming_operators.cc:439
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
==>
    dist_proc/dax/workflow/****/streaming/fnapi_streaming_operators.cc:439
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
==>
    dist_proc/dax/workflow/****/streaming/fnapi_streaming_operators.cc:439
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
==>
    dist_proc/dax/workflow/****/streaming/fnapi_streaming_operators.cc:439
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
==>
    dist_proc/dax/workflow/****/streaming/fnapi_streaming_operators.cc:439
Mar 24, 2022 4:01:07 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-24T16:01:05.123Z: Cancel request is committed for workflow job: 2022-03-24_05_47_36-985224193372712323.
Mar 24, 2022 4:01:07 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-24T16:01:05.204Z: Cleaning up.
Mar 24, 2022 4:01:07 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-24T16:01:05.264Z: Stopping **** pool...
Mar 24, 2022 4:01:07 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-24T16:01:05.322Z: Stopping **** pool...
Mar 24, 2022 4:03:36 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-24T16:03:35.912Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Mar 24, 2022 4:03:36 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-24T16:03:35.948Z: Worker pool stopped.
Mar 24, 2022 4:03:41 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-03-24_05_47_36-985224193372712323 finished with status CANCELLED.
Load test results for test (ID): 171bff61-9a56-4b07-afe1-c023c511d34e and timestamp: 2022-03-24T12:47:30.070000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                 11459.946
dataflow_v2_java17_total_bytes_count             2.61693813E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220324124524
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:fc4a710fc685b2e7cc2e80f42dcf1be850e0987e361eebcd7bd68c1677ad334b
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220324124524]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:fc4a710fc685b2e7cc2e80f42dcf1be850e0987e361eebcd7bd68c1677ad334b]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220324124524] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:fc4a710fc685b2e7cc2e80f42dcf1be850e0987e361eebcd7bd68c1677ad334b])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:23dd33b71af4dc30d27e6504abb9ad7a95ca776d900ce9bac2e911c239c6010d
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:23dd33b71af4dc30d27e6504abb9ad7a95ca776d900ce9bac2e911c239c6010d
ERROR: (gcloud.container.images.delete) Not found: response: {'docker-distribution-api-version': 'registry/2.0', 'content-type': 'application/json', 'date': 'Thu, 24 Mar 2022 16:03:48 GMT', 'server': 'Docker Registry', 'cache-control': 'private', 'x-xss-protection': '0', 'x-frame-options': 'SAMEORIGIN', 'transfer-encoding': 'chunked', 'status': '404', 'content-length': '168', '-content-encoding': 'gzip'}
Failed to compute blob liveness for manifest: 'sha256:23dd33b71af4dc30d27e6504abb9ad7a95ca776d900ce9bac2e911c239c6010d': None

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages FAILED

FAILURE: Build completed with 2 failures.

1: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.
==============================================================================

2: Task failed with an exception.
-----------
* Where:
Build file '<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/runners/google-cloud-dataflow-java/build.gradle'> line: 297

* What went wrong:
Execution failed for task ':runners:google-cloud-dataflow-java:cleanUpDockerJavaImages'.
> Process 'command './scripts/cleanup_untagged_gcr_images.sh'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.
==============================================================================

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 18m 37s
110 actionable tasks: 73 executed, 33 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/wasgz2j63xfhs

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #93

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/93/display/redirect?page=changes>

Changes:

[mmack] [BEAM-13232] Close clients properly in KinesisSource. Also use lazy init

[noreply] [BEAM-14141] Set Interactive Beam to use the default Dataproc image

[noreply] BEAM-14115 - Update find criteria limited to _id (#17102)

[chamikaramj] Disable BigQueryIOStorageWriteIT for general Java post-commit

[noreply] Revert "[BEAM-14038] Auto-startup for Python expansion service.

[noreply] Minor: Bump timeout for Java PreCommit (#17157)

[noreply] [BEAM-14152] Disable flaky

[noreply] Fixing a small bug in TypedSchemaTransformTest that caused it to flake.

[noreply] [BEAM-14116] Catch MonitoringInfoMetricName null keys or values in the

[noreply] [BEAM-14129] Restructure SubscriptionPartitionLoader to use a manual SDF

[noreply] [BEAM-13015] Avoid repeated weighing of StateKey in

[noreply] Add option to add modules to JDK add-open (#17110)

[noreply] [BEAM-13015] Clarify ownership of the list for state caching across


------------------------------------------
[...truncated 197.06 KB...]
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
Mar 23, 2022 4:01:02 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-23T16:01:01.986Z: Cancel request is committed for workflow job: 2022-03-23_05_56_54-2353223608011268682.
Mar 23, 2022 4:01:02 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-23T16:01:02.098Z: Cleaning up.
Mar 23, 2022 4:01:02 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-23T16:01:02.206Z: Stopping **** pool...
Mar 23, 2022 4:01:02 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-23T16:01:02.293Z: Stopping **** pool...
Mar 23, 2022 4:03:19 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-23T16:03:16.590Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Mar 23, 2022 4:03:19 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-23T16:03:16.658Z: Worker pool stopped.
Mar 23, 2022 4:03:23 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-03-23_05_56_54-2353223608011268682 finished with status CANCELLED.
Load test results for test (ID): 6f5b5239-1fa8-4da4-a565-c1c0308e840a and timestamp: 2022-03-23T12:56:46.573000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                 10862.829
dataflow_v2_java17_total_bytes_count             2.47387186E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220323125323
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:81e163f2e158709d2904623ea4faf7b22312e1efb360dee4dfac2ff583b50612
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220323125323]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:81e163f2e158709d2904623ea4faf7b22312e1efb360dee4dfac2ff583b50612]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220323125323] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:81e163f2e158709d2904623ea4faf7b22312e1efb360dee4dfac2ff583b50612])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:81e163f2e158709d2904623ea4faf7b22312e1efb360dee4dfac2ff583b50612
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:81e163f2e158709d2904623ea4faf7b22312e1efb360dee4dfac2ff583b50612
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:81e163f2e158709d2904623ea4faf7b22312e1efb360dee4dfac2ff583b50612].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 10m 57s
110 actionable tasks: 76 executed, 30 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/5ppxvujrfkaqw

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #92

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/92/display/redirect?page=changes>

Changes:

[Kyle Weaver] [BEAM-14124] Add display data to BQ storage reads.

[mmack] [adhoc] Move aws IT tests to testing package according to best practices

[noreply] fixes static checks and go lint issues (#17138)

[Kyle Weaver] Don't print in task configuration.

[noreply] [BEAM-14136] Clean up staticcheck and linter warnings in the Go SDK

[noreply] Merge pull request #17063 from [BEAM-12164] Fix flaky tests

[noreply] Revert "[BEAM-14112] Avoid storing a generator in _CustomBigQuerySource

[Kyle Weaver] [BEAM-4106] Remove filesToStage from Flink pipeline option list.

[noreply] [BEAM-14071] Enabling Flink on Dataproc for Interactive Beam (#17044)

[noreply] Minor: Bypass schema registry in schemas_test.py (#17108)


------------------------------------------
[...truncated 96.30 KB...]
    result = self._call_chain(self.handle_open, protocol, protocol +
  File "/usr/lib/python3.8/urllib/request.py", line 502, in _call_chain
    result = func(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 1383, in http_open
    return self.do_open(http.client.HTTPConnection, req)
  File "/usr/lib/python3.8/urllib/request.py", line 1357, in do_open
    raise URLError(err)
urllib.error.URLError: <urlopen error [Errno 101] Network is unreachable>
ERROR:root:Invalid url for checkstyle-8.23: http://www.gnu.org/licenses/old-licenses/lgpl-2.1.txt. Retrying...
Traceback (most recent call last):
  File "/usr/lib/python3.8/urllib/request.py", line 1354, in do_open
    h.request(req.get_method(), req.selector, req.data, headers,
  File "/usr/lib/python3.8/http/client.py", line 1256, in request
    self._send_request(method, url, body, headers, encode_chunked)
  File "/usr/lib/python3.8/http/client.py", line 1302, in _send_request
    self.endheaders(body, encode_chunked=encode_chunked)
  File "/usr/lib/python3.8/http/client.py", line 1251, in endheaders
    self._send_output(message_body, encode_chunked=encode_chunked)
  File "/usr/lib/python3.8/http/client.py", line 1011, in _send_output
    self.send(msg)
  File "/usr/lib/python3.8/http/client.py", line 951, in send
    self.connect()
  File "/usr/lib/python3.8/http/client.py", line 922, in connect
    self.sock = self._create_connection(
  File "/usr/lib/python3.8/socket.py", line 808, in create_connection
    raise err
  File "/usr/lib/python3.8/socket.py", line 796, in create_connection
    sock.connect(sa)
OSError: [Errno 101] Network is unreachable

During handling of the above exception, another exception occurred:

Traceback (most recent call last):
  File "<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/license_scripts/pull_licenses_java.py",> line 58, in pull_from_url
    url_read = urlopen(url)
  File "/usr/lib/python3.8/urllib/request.py", line 222, in urlopen
    return opener.open(url, data, timeout)
  File "/usr/lib/python3.8/urllib/request.py", line 525, in open
    response = self._open(req, data)
  File "/usr/lib/python3.8/urllib/request.py", line 542, in _open
    result = self._call_chain(self.handle_open, protocol, protocol +
  File "/usr/lib/python3.8/urllib/request.py", line 502, in _call_chain
    result = func(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 1383, in http_open
    return self.do_open(http.client.HTTPConnection, req)
  File "/usr/lib/python3.8/urllib/request.py", line 1357, in do_open
    raise URLError(err)
urllib.error.URLError: <urlopen error [Errno 101] Network is unreachable>
ERROR:root:Invalid url for jFormatString-3.0.0: http://www.gnu.org/licenses/lgpl.html. Retrying...
Traceback (most recent call last):
  File "/usr/lib/python3.8/urllib/request.py", line 1354, in do_open
    h.request(req.get_method(), req.selector, req.data, headers,
  File "/usr/lib/python3.8/http/client.py", line 1256, in request
    self._send_request(method, url, body, headers, encode_chunked)
  File "/usr/lib/python3.8/http/client.py", line 1302, in _send_request
    self.endheaders(body, encode_chunked=encode_chunked)
  File "/usr/lib/python3.8/http/client.py", line 1251, in endheaders
    self._send_output(message_body, encode_chunked=encode_chunked)
  File "/usr/lib/python3.8/http/client.py", line 1011, in _send_output
    self.send(msg)
  File "/usr/lib/python3.8/http/client.py", line 951, in send
    self.connect()
  File "/usr/lib/python3.8/http/client.py", line 922, in connect
    self.sock = self._create_connection(
  File "/usr/lib/python3.8/socket.py", line 808, in create_connection
    raise err
  File "/usr/lib/python3.8/socket.py", line 796, in create_connection
    sock.connect(sa)
OSError: [Errno 101] Network is unreachable

During handling of the above exception, another exception occurred:

Traceback (most recent call last):
  File "<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/license_scripts/pull_licenses_java.py",> line 58, in pull_from_url
    url_read = urlopen(url)
  File "/usr/lib/python3.8/urllib/request.py", line 222, in urlopen
    return opener.open(url, data, timeout)
  File "/usr/lib/python3.8/urllib/request.py", line 525, in open
    response = self._open(req, data)
  File "/usr/lib/python3.8/urllib/request.py", line 542, in _open
    result = self._call_chain(self.handle_open, protocol, protocol +
  File "/usr/lib/python3.8/urllib/request.py", line 502, in _call_chain
    result = func(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 1383, in http_open
    return self.do_open(http.client.HTTPConnection, req)
  File "/usr/lib/python3.8/urllib/request.py", line 1357, in do_open
    raise URLError(err)
urllib.error.URLError: <urlopen error [Errno 101] Network is unreachable>
ERROR:root:Invalid url for checkstyle-8.23: http://www.gnu.org/licenses/old-licenses/lgpl-2.1.txt. Retrying...
Traceback (most recent call last):
  File "/usr/lib/python3.8/urllib/request.py", line 1354, in do_open
    h.request(req.get_method(), req.selector, req.data, headers,
  File "/usr/lib/python3.8/http/client.py", line 1256, in request
    self._send_request(method, url, body, headers, encode_chunked)
  File "/usr/lib/python3.8/http/client.py", line 1302, in _send_request
    self.endheaders(body, encode_chunked=encode_chunked)
  File "/usr/lib/python3.8/http/client.py", line 1251, in endheaders
    self._send_output(message_body, encode_chunked=encode_chunked)
  File "/usr/lib/python3.8/http/client.py", line 1011, in _send_output
    self.send(msg)
  File "/usr/lib/python3.8/http/client.py", line 951, in send
    self.connect()
  File "/usr/lib/python3.8/http/client.py", line 922, in connect
    self.sock = self._create_connection(
  File "/usr/lib/python3.8/socket.py", line 808, in create_connection
    raise err
  File "/usr/lib/python3.8/socket.py", line 796, in create_connection
    sock.connect(sa)
OSError: [Errno 101] Network is unreachable

During handling of the above exception, another exception occurred:

Traceback (most recent call last):
  File "<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/license_scripts/pull_licenses_java.py",> line 58, in pull_from_url
    url_read = urlopen(url)
  File "/usr/lib/python3.8/urllib/request.py", line 222, in urlopen
    return opener.open(url, data, timeout)
  File "/usr/lib/python3.8/urllib/request.py", line 525, in open
    response = self._open(req, data)
  File "/usr/lib/python3.8/urllib/request.py", line 542, in _open
    result = self._call_chain(self.handle_open, protocol, protocol +
  File "/usr/lib/python3.8/urllib/request.py", line 502, in _call_chain
    result = func(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 1383, in http_open
    return self.do_open(http.client.HTTPConnection, req)
  File "/usr/lib/python3.8/urllib/request.py", line 1357, in do_open
    raise URLError(err)
urllib.error.URLError: <urlopen error [Errno 101] Network is unreachable>
ERROR:root:Invalid url for jFormatString-3.0.0: http://www.gnu.org/licenses/lgpl.html after 9 retries.
Traceback (most recent call last):
  File "/usr/lib/python3.8/urllib/request.py", line 1354, in do_open
    h.request(req.get_method(), req.selector, req.data, headers,
  File "/usr/lib/python3.8/http/client.py", line 1256, in request
    self._send_request(method, url, body, headers, encode_chunked)
  File "/usr/lib/python3.8/http/client.py", line 1302, in _send_request
    self.endheaders(body, encode_chunked=encode_chunked)
  File "/usr/lib/python3.8/http/client.py", line 1251, in endheaders
    self._send_output(message_body, encode_chunked=encode_chunked)
  File "/usr/lib/python3.8/http/client.py", line 1011, in _send_output
    self.send(msg)
  File "/usr/lib/python3.8/http/client.py", line 951, in send
    self.connect()
  File "/usr/lib/python3.8/http/client.py", line 922, in connect
    self.sock = self._create_connection(
  File "/usr/lib/python3.8/socket.py", line 808, in create_connection
    raise err
  File "/usr/lib/python3.8/socket.py", line 796, in create_connection
    sock.connect(sa)
OSError: [Errno 101] Network is unreachable

During handling of the above exception, another exception occurred:

Traceback (most recent call last):
  File "<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/license_scripts/pull_licenses_java.py",> line 58, in pull_from_url
    url_read = urlopen(url)
  File "/usr/lib/python3.8/urllib/request.py", line 222, in urlopen
    return opener.open(url, data, timeout)
  File "/usr/lib/python3.8/urllib/request.py", line 525, in open
    response = self._open(req, data)
  File "/usr/lib/python3.8/urllib/request.py", line 542, in _open
    result = self._call_chain(self.handle_open, protocol, protocol +
  File "/usr/lib/python3.8/urllib/request.py", line 502, in _call_chain
    result = func(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 1383, in http_open
    return self.do_open(http.client.HTTPConnection, req)
  File "/usr/lib/python3.8/urllib/request.py", line 1357, in do_open
    raise URLError(err)
urllib.error.URLError: <urlopen error [Errno 101] Network is unreachable>
ERROR:root:Invalid url for checkstyle-8.23: http://www.gnu.org/licenses/old-licenses/lgpl-2.1.txt. Retrying...
Traceback (most recent call last):
  File "/usr/lib/python3.8/urllib/request.py", line 1354, in do_open
    h.request(req.get_method(), req.selector, req.data, headers,
  File "/usr/lib/python3.8/http/client.py", line 1256, in request
    self._send_request(method, url, body, headers, encode_chunked)
  File "/usr/lib/python3.8/http/client.py", line 1302, in _send_request
    self.endheaders(body, encode_chunked=encode_chunked)
  File "/usr/lib/python3.8/http/client.py", line 1251, in endheaders
    self._send_output(message_body, encode_chunked=encode_chunked)
  File "/usr/lib/python3.8/http/client.py", line 1011, in _send_output
    self.send(msg)
  File "/usr/lib/python3.8/http/client.py", line 951, in send
    self.connect()
  File "/usr/lib/python3.8/http/client.py", line 922, in connect
    self.sock = self._create_connection(
  File "/usr/lib/python3.8/socket.py", line 808, in create_connection
    raise err
  File "/usr/lib/python3.8/socket.py", line 796, in create_connection
    sock.connect(sa)
OSError: [Errno 101] Network is unreachable

During handling of the above exception, another exception occurred:

Traceback (most recent call last):
  File "<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/license_scripts/pull_licenses_java.py",> line 58, in pull_from_url
    url_read = urlopen(url)
  File "/usr/lib/python3.8/urllib/request.py", line 222, in urlopen
    return opener.open(url, data, timeout)
  File "/usr/lib/python3.8/urllib/request.py", line 525, in open
    response = self._open(req, data)
  File "/usr/lib/python3.8/urllib/request.py", line 542, in _open
    result = self._call_chain(self.handle_open, protocol, protocol +
  File "/usr/lib/python3.8/urllib/request.py", line 502, in _call_chain
    result = func(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 1383, in http_open
    return self.do_open(http.client.HTTPConnection, req)
  File "/usr/lib/python3.8/urllib/request.py", line 1357, in do_open
    raise URLError(err)
urllib.error.URLError: <urlopen error [Errno 101] Network is unreachable>
ERROR:root:Invalid url for checkstyle-8.23: http://www.gnu.org/licenses/old-licenses/lgpl-2.1.txt after 9 retries.
ERROR:root:['spotbugs-annotations-4.0.6', 'jFormatString-3.0.0', 'checkstyle-8.23']
ERROR:root:**************************************** Licenses were not able to be pulled automatically for some dependencies. Please search source code of the dependencies on the internet and add "license" and "notice" (if available) field to <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/license_scripts/dep_urls_java.yaml> for each missing license. Dependency List: [checkstyle-8.23,jFormatString-3.0.0,spotbugs-annotations-4.0.6]
INFO:root:pull_licenses_java.py failed. It took 490.211551 seconds with 16 threads.
Traceback (most recent call last):
  File "<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/license_scripts/pull_licenses_java.py",> line 321, in <module>
    raise RuntimeError('{n} error(s) occurred.'.format(n=len(error_msg)),
RuntimeError: ('1 error(s) occurred.', ['**************************************** Licenses were not able to be pulled automatically for some dependencies. Please search source code of the dependencies on the internet and add "license" and "notice" (if available) field to <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/license_scripts/dep_urls_java.yaml> for each missing license. Dependency List: [checkstyle-8.23,jFormatString-3.0.0,spotbugs-annotations-4.0.6]'])

> Task :sdks:java:container:pullLicenses FAILED

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:container:pullLicenses'.
> Process 'command './license_scripts/license_script.sh'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 8m 51s
103 actionable tasks: 67 executed, 32 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/tyynoqqjcz5vs

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #91

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/91/display/redirect>

Changes:


------------------------------------------
[...truncated 173.78 KB...]
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmi
Mar 21, 2022 4:00:53 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-21T16:00:53.114Z: Cancel request is committed for workflow job: 2022-03-21_05_33_13-12093101199060966286.
Mar 21, 2022 4:00:53 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-21T16:00:53.193Z: Cleaning up.
Mar 21, 2022 4:00:53 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-21T16:00:53.266Z: Stopping **** pool...
Mar 21, 2022 4:00:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-21T16:00:53.311Z: Stopping **** pool...
Mar 21, 2022 4:03:26 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-21T16:03:25.962Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Mar 21, 2022 4:03:26 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-21T16:03:26.036Z: Worker pool stopped.
Mar 21, 2022 4:03:43 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-03-21_05_33_13-12093101199060966286 finished with status CANCELLED.
Load test results for test (ID): 6080656d-f371-45a0-949b-ad62381f9747 and timestamp: 2022-03-21T12:33:07.798000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                 12318.889
dataflow_v2_java17_total_bytes_count             2.29082851E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220321123038
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:eac6502b09713bce9bb50bfc7780b9aa55ce9784880beb5540fb72bd9f18d9cf
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220321123038]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:eac6502b09713bce9bb50bfc7780b9aa55ce9784880beb5540fb72bd9f18d9cf]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220321123038] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:eac6502b09713bce9bb50bfc7780b9aa55ce9784880beb5540fb72bd9f18d9cf])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:eac6502b09713bce9bb50bfc7780b9aa55ce9784880beb5540fb72bd9f18d9cf
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:eac6502b09713bce9bb50bfc7780b9aa55ce9784880beb5540fb72bd9f18d9cf
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:eac6502b09713bce9bb50bfc7780b9aa55ce9784880beb5540fb72bd9f18d9cf].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 33m 29s
109 actionable tasks: 72 executed, 33 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/vxvbseb7qwn4w

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #90

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/90/display/redirect?page=changes>

Changes:

[noreply] [BEAM-14122] Upgrade pip-licenses dependency (#17132)


------------------------------------------
[...truncated 75.81 KB...]
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
Mar 20, 2022 4:03:40 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-20T16:03:38.051Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Mar 20, 2022 4:03:40 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-20T16:03:38.104Z: Worker pool stopped.
Mar 20, 2022 4:03:46 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-03-20_05_33_54-15103221717270704761 finished with status CANCELLED.
Load test results for test (ID): c8c7faad-9563-4f16-b9bd-fef8eba4466a and timestamp: 2022-03-20T12:33:46.406000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                 12282.641
dataflow_v2_java17_total_bytes_count             3.37656372E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220320123107
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:6fee85bcf825f0edb186dda1f9b55ac700431e2dd6b361ee854f4c4fb5f21ec4
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220320123107]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:6fee85bcf825f0edb186dda1f9b55ac700431e2dd6b361ee854f4c4fb5f21ec4]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220320123107] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:6fee85bcf825f0edb186dda1f9b55ac700431e2dd6b361ee854f4c4fb5f21ec4])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:6fee85bcf825f0edb186dda1f9b55ac700431e2dd6b361ee854f4c4fb5f21ec4
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:6fee85bcf825f0edb186dda1f9b55ac700431e2dd6b361ee854f4c4fb5f21ec4
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:6fee85bcf825f0edb186dda1f9b55ac700431e2dd6b361ee854f4c4fb5f21ec4].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 33m 3s
109 actionable tasks: 72 executed, 33 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/lzqsnwicozqvg

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #89

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/89/display/redirect?page=changes>

Changes:

[Kiley Sok] Add Java 17 Nexmark metrics to Grafana

[yiru] .

[yiru] .

[yiru] .

[yiru] format fix

[yiru] .

[yiru] make DoFn into a separate class

[yiru] .

[yiru] fix setting

[mmack] [adhoc] Minor cleanup for aws2 tests

[mmack] [BEAM-14125] Update website IO matrix to recommend aws2 IOs

[noreply] [BEAM-14128] Eliminating quadratic behavior of

[noreply] [BEAM-13972] Add RunInference interface (#16917)

[noreply] Merge pull request #17116 from [BEAM-12164] Remove change_stream in

[yiru] fix checkstyle

[yiru] spotlessapply

[noreply] Deprecate tags.go (#17025)

[noreply] [BEAM-12753] and [BEAM-12815] Fix Flink Integration Tests (#17067)

[noreply] Merge pull request #16895 from [BEAM-13882][Playground] More tests for

[noreply] [BEAM-13925] Add weekly automation to update our reviewer config

[noreply] Merge pull request #17076 from Beam 14082 update payground for mobile

[noreply] [BEAM-13925] Assign committers in the scheduled action (#17062)

[noreply] Pin setup-gcloud to v0 instead of master (#17123)

[noreply] [BEAM-3304] documentation for PaneInfo in BPG (#17047)

[noreply] Merge pull request #17016 from [BEAM-14049][Playground] Add new API

[noreply] Merge pull request #17077 from [BEAM-14078] [Website] change link

[noreply] Merge pull request #17085 from [BEAM-14077] [Website] add beam

[noreply] Update Changes.md w/Go pipeline pre-process fix.

[noreply] [BEAM-14098] wrapper for postgres on JDBC IO GO SDK (#17088)

[noreply] Merge pull request #17023 from [BEAM-12164]: Remove child partition


------------------------------------------
[...truncated 451.20 KB...]
==>
    dist_proc/dax/workflow/****/streaming/fnapi_streaming_operators.cc:439
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
==>
    dist_proc/dax/workflow/****/streaming/fnapi_streaming_operators.cc:439
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
==>
    dist_proc/dax/workflow/****/streaming/fnapi_streaming_operators.cc:439
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
==>
    dist_proc/dax/workflow/****/streaming/fnapi_streaming_operators.cc:439
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
==>
    dist_proc/dax/workflow/****/streaming/fnapi_streaming_operators.cc:439
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
==>
    dist_proc/dax/workflow/****/streaming/fnapi_streaming_operators.cc:439
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
==>
    dist_proc/dax/workflow/****/streaming/fnapi_streaming_operators.cc:439
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
==>
    dist_proc/dax/workflow/****/streaming/fnapi_streaming_operators.cc:439
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
==>
    dist_proc/dax/workflow/****/streaming/fnapi_streaming_operators.cc:439
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
==>
    dist_proc/dax/workflow/****/streaming/fnapi_streaming_operators.cc:439
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
==>
    dist_proc/dax/workflow/****/streaming/fnapi_streaming_operators.cc:439
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
==>
    dist_proc/dax/workflow/****/streaming/fnapi_streaming_operators.cc:439
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
==>
    dist_proc/dax/workflow/****/streaming/fnapi_streaming_operators.cc:439
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
==>
    dist_proc/dax/workflow/****/streaming/fnapi_streaming_operators.cc:439
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
=
Mar 19, 2022 4:03:17 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-19T16:03:17.250Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Mar 19, 2022 4:03:19 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-19T16:03:17.305Z: Worker pool stopped.
Mar 19, 2022 4:03:22 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-03-19_05_36_24-5316936995767417513 finished with status CANCELLED.
Load test results for test (ID): ebac2d50-d651-4f88-a492-490151db556c and timestamp: 2022-03-19T12:36:19.431000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                  12121.36
dataflow_v2_java17_total_bytes_count             2.86365661E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220319123354
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:22e481d6bacf51c889aa54c383557e0ee611f25620b80e7c6dd1da7459e8cfbd
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220319123354]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:22e481d6bacf51c889aa54c383557e0ee611f25620b80e7c6dd1da7459e8cfbd]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220319123354] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:22e481d6bacf51c889aa54c383557e0ee611f25620b80e7c6dd1da7459e8cfbd])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:22e481d6bacf51c889aa54c383557e0ee611f25620b80e7c6dd1da7459e8cfbd
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:22e481d6bacf51c889aa54c383557e0ee611f25620b80e7c6dd1da7459e8cfbd
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:22e481d6bacf51c889aa54c383557e0ee611f25620b80e7c6dd1da7459e8cfbd].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 29m 49s
109 actionable tasks: 72 executed, 33 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/liad3x36e75gm

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #88

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/88/display/redirect?page=changes>

Changes:

[Luke Cwik] [BEAM-10212] Clean-up comments, remove rawtypes usage.

[noreply] [BEAM-11934] Add enable_file_dynamic_sharding to allow DataflowRunner

[noreply] [BEAM-12777] Create symlink for `current` directory (#17105)

[noreply] [BEAM-14020] Adding SchemaTransform, SchemaTransformProvider,

[noreply] [BEAM-13015] Modify metrics to begin and reset to a non-dirty state.

[noreply] [BEAM-14112] Avoid storing a generator in _CustomBigQuerySource (#17100)

[noreply] Populate environment capabilities in v1beta3 protos. (#17042)

[Kyle Weaver] [BEAM-12976] Test a whole pipeline using projection pushdown in BQ IO.

[Kyle Weaver] [BEAM-12976] Enable projection pushdown for Java pipelines on Dataflow,

[noreply] [BEAM-14038] Auto-startup for Python expansion service. (#17035)

[Kyle Weaver] [BEAM-14123] Fix typo in hdfsIntegrationTest task name.

[noreply] [BEAM-13893] improved coverage of jobopts package (#17003)

[noreply] Merge pull request #16977 from [BEAM-12164]  Added integration test for


------------------------------------------
[...truncated 158.99 KB...]
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
Mar 18, 2022 12:46:22 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
SEVERE: 2022-03-18T12:46:19.877Z: generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
Mar 18, 2022 4:00:58 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-18T16:00:58.030Z: Cancel request is committed for workflow job: 2022-03-18_05_38_38-10321091201345048590.
Mar 18, 2022 4:00:58 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-18T16:00:58.160Z: Cleaning up.
Mar 18, 2022 4:00:59 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-18T16:00:58.433Z: Stopping **** pool...
Mar 18, 2022 4:00:59 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-18T16:00:58.519Z: Stopping **** pool...
Mar 18, 2022 4:03:15 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-18T16:03:15.317Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Mar 18, 2022 4:03:15 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-18T16:03:15.409Z: Worker pool stopped.
Mar 18, 2022 4:03:21 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-03-18_05_38_38-10321091201345048590 finished with status CANCELLED.
Load test results for test (ID): bf0dd004-d93c-43d0-8e9c-88e0ffb6dd4b and timestamp: 2022-03-18T12:38:26.092000000Z:
Exception in thread "main"                  Metric:                    Value:
dataflow_v2_java17_runtime_sec                 11989.959
java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
dataflow_v2_java17_total_bytes_count             3.07381204E10
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220318123610
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:285da2d04e1dba971666500a8615417be3837c780c160035b2a5e3f9be5365a7
Deleted: sha256:4f791f237666227d50370490c25951202c7d9eb70e1a6821ff4599b97440c2be
Deleted: sha256:cedb018695816e4152849a83aafa34588bf745823ffb9cfbcea90f84ce6731c7
Deleted: sha256:d5a75997369103c972dfdaf21d29f2f2fb853471d8afb9d6404d2df913a93958
Deleted: sha256:1568afad16ed6f8dd7bda800c1581430d9fec5e8fc1a966ccb41ccd97a113ea8
Deleted: sha256:f20c16a2d9195bfba3adc483d98eb12bf70dfba113b77fd86609d763c8586c86
Deleted: sha256:31e4b8d8145bc014f424545d07226a45e6e4b475a6ef75e3ef4b49ecb83c3178
Deleted: sha256:307d29ecd9be3c96874baae5d9155377dc1d90e0cd5314c00dad3c7f69633d09
Deleted: sha256:7d57958fc3c3bbd2677e9183fba5741faaf447c0d38695ad2cc40b61436b73a4
Deleted: sha256:ff2f276961425725e8961c2c048233186c0ab0d57a63eb5df5fe6a1c9fcbee23
Deleted: sha256:d827fe1ba91df62aff1e013f379749f51cf1c3decbc38e99977d8a16019fead7
Deleted: sha256:6d5c1aaa2cc08abad7f0412460ed56611581e283ce953818ac01238170737376
Deleted: sha256:5e275235333f94be51ab3f2cd95dd4604b9a86876e4d6c75795369af11f22f9f
Deleted: sha256:7d401f06147577f36fe4ba0c88e8c15128244524804979d58ba0972d5e828d18
Deleted: sha256:fa9c88d08090f2a77af22be795df76771875344a262118a8c1067bcbba6228ad
Deleted: sha256:5444047e11eb0135ba711f91327fa5f10a44aa4464dd980fb57eb9e2409ec414
Deleted: sha256:ac73387e66d3b062b425e84c6c256d57fdaff556fe8fd7dbe546b49ae7cff901
Deleted: sha256:ee201c319449b40e7efe283100a5fe7fab679c630db29a00bc71dda9c66354df
Deleted: sha256:14db9c3b1c771e5666bfac159351e3986ee7a9fc0c888a784f1d2c7ee0f3c8e6
Deleted: sha256:1ec9d674004b48fa020770297fe4d3675f29dc337aec8b9b7f6f587d075515d8
Deleted: sha256:2155081bd5f4c30eedaec35ddf57244708551f0df0544691e7f564908ff6108a
Deleted: sha256:10bfc6d88eeff9e82a93bf539195f116b7fe459bac4d55e1f7722076c8aec204
Deleted: sha256:233f2f34f4db1df6466eddb2dce1ab0e9f4c18013d1d3ab75bef30cb25b8ca55
Deleted: sha256:296ec1bebad6c36c9bd11544dfd3f98e8e8d1275a0f830916238f9842e29d15e
Deleted: sha256:7be9c6cec3062bfdb9b39d8fc0bc55da3eeeaaf5ab5a0ba4b6dd6a93ca828d78
Deleted: sha256:6a03fd9328e09ff073b2890ec38d99baff5d6890a52681a331b92ab468912351
Deleted: sha256:75f858462c4f17617d41446a7848705eded17a2f2a314bb764d9b14c89a74f73
Deleted: sha256:3cd85c6ffc59068d18f73016c8167d1a7cf06e41adc57b7a19391114034cfbac
Deleted: sha256:c81626450806646ecea50750e84940811a17ff18c5e7766fe48a48a569fbe62c
Deleted: sha256:36bea53e17bfa5d5d054a093a6e85840eb86ac6f61d977b5d0968464a57558b4
Deleted: sha256:e49c32c396234d40f7e32171bf12383bf69d6ee1185e9e20a25d1ae029f54636
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220318123610]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:285da2d04e1dba971666500a8615417be3837c780c160035b2a5e3f9be5365a7]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220318123610] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:285da2d04e1dba971666500a8615417be3837c780c160035b2a5e3f9be5365a7])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:381df7627f67811583b68d338f48b25b8f4b4b8cf3e17c3473aff098bddd27c5
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:381df7627f67811583b68d338f48b25b8f4b4b8cf3e17c3473aff098bddd27c5
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:381df7627f67811583b68d338f48b25b8f4b4b8cf3e17c3473aff098bddd27c5].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:285da2d04e1dba971666500a8615417be3837c780c160035b2a5e3f9be5365a7
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:285da2d04e1dba971666500a8615417be3837c780c160035b2a5e3f9be5365a7
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:285da2d04e1dba971666500a8615417be3837c780c160035b2a5e3f9be5365a7].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 27m 26s
109 actionable tasks: 72 executed, 33 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/i3yeutqsaa3bm

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #87

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/87/display/redirect?page=changes>

Changes:

[noreply] Mapped JOB_STATE_RESOURCE_CLEANING_UP to State.RUNNING.

[ryanthompson591] fixed typo in typehints

[zyichi] Remove unused prebuild_sdk_container_base_iamge option from validate

[hengfeng] feat: add more custom metrics

[noreply] [BEAM-14103][Playgrounf][Bugfix] Fix google analytics id (#17092)

[noreply] Minor: Make ScopedReadStateSupplier final (#16992)

[noreply] [BEAM-14113] Improve SamzaJobServerDriver extensibility (#17099)

[noreply] [BEAM-14116] Chunk commit requests dynamically (#17004)

[noreply] Merge pull request #17079 from [BEAM-13660] Add types and queries in

[noreply] [BEAM-13888] Add unit testing to ioutilx (#17058)

[noreply] Merge pull request #16822 from [BEAM-13841][Playground] Add Application

[noreply] Minor: Make serializableCoder warning gramatically correct english

[noreply] [BEAM-14091] Fixing Interactive Beam show/collect for remote runners


------------------------------------------
[...truncated 775.47 KB...]
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:222
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmi
Mar 17, 2022 4:03:45 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-17T16:03:44.621Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Mar 17, 2022 4:03:45 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-17T16:03:44.697Z: Worker pool stopped.
Mar 17, 2022 4:03:53 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-03-17_05_33_17-15156894379266467156 finished with status CANCELLED.
Load test results for test (ID): 0d417360-61a6-455b-a253-ba7f15da465e and timestamp: 2022-03-17T12:33:12.270000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                 12326.478
dataflow_v2_java17_total_bytes_count             3.78690776E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220317123039
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:9e2ac9cf49716cd991f377754cbb64532d5dd6ce720188b64d6dae3e6265e63b
Deleted: sha256:2de597a5274795d155795a62dfa26489640b9b2db76f38e7e1ad5568cec22d79
Deleted: sha256:01084484f98299e349e53fe5ac3d0468c389b4357544c4bdd4c1cee6f476f351
Deleted: sha256:89d9b569987b1f98b16e7b6f900a4e937b2860125ffc07b913d0449d51e1c580
Deleted: sha256:0a581de2c9cb020efcad0c45bc14876be8862720d1a491375a71a9091c7c449c
Deleted: sha256:004d63a5319e8b05590cedaf14ba8ff6ec8f32e06a58e5f38ca5a1c2df1cc6d0
Deleted: sha256:b4f2906541ac4508a3c3aaa06eaaa6add07a766fb0f8d052ee2d5590136ac6a8
Deleted: sha256:0790d3b9f0382b339fb1016e3aca7a961660f3dd581e478984bf40528888ead3
Deleted: sha256:0831af310198d337d5da00acd010dcc5bb6aab58a707655c0b0d79f4bea1ef15
Deleted: sha256:d8244a87ef8fa53bb10314914f8b76ade27c21f40de468fd437a58f8aa0c107e
Deleted: sha256:ec34ca1abcace3651aac3e6a7a3cb7d04c50791435da3ce95e47cd2735e5ce01
Deleted: sha256:9cbed822988f60f6dd0b4bacd4e7b08ea9efff54dc335a5966121e7ae6780ec4
Deleted: sha256:0a59d9baa1e84796f175a6a08b11b46164e4322b8bdf132acca2a56d3feecc09
Deleted: sha256:9e14f34a681ac4f16d27f374cae4db9ebc456ab7189fdb163785eafe7356f577
Deleted: sha256:8657509d0a6348875c8a22f05bd6a3ec7b1dc57801336c94690d23e0873e4bb3
Deleted: sha256:1d081b2880ee0d6db24adb7917bbac3981067a2255693ad4f657d2b62e1f61a7
Deleted: sha256:bd6f7808f780ff473c3aad7ac1f6f11d061bf631eb0f6a7bf900d3d4c09dc51d
Deleted: sha256:f429fa328a2ee2dfad587b60f1bc68c152b2cb34222a0daa2a0bf6ab91b7986c
Deleted: sha256:8371e94ca7b4b5b038956f1bf0b7bcd2426c64d6aed554d17d29222c1c5688bd
Deleted: sha256:413ed0400770024664d125c05385108f1466e9c3a960703d13c7c571654f3104
Deleted: sha256:7e81ea86e717b0479c5b72720337a0e7edf5158ee694cdb00290c4e269a9bc1e
Deleted: sha256:6c2fd24af21f98df03092e4159b81a8b0bfcaef87c1cc30bc0f7a689003e1713
Deleted: sha256:078d178abe02a8ab89831fa0767505a0636a49c663918fe5220fc13a39ba89a6
Deleted: sha256:9c38377a6a7fecb9f98c9240116f623627e84f4d5d7b5b69193a44df81636c79
Deleted: sha256:db53446f9f811022cab20c1d75dd372a991186051e6ad348610814cb815077da
Deleted: sha256:68fb003f5bbdb1f66495121518243349c3a263883089e1a20ac9b0ed22ec9a6c
Deleted: sha256:f3c7a214be31b57435a2764a9e90b6730bfe2da7185c49c0f464cab3e490938a
Deleted: sha256:ab606285db362a76287610b6e3bb6c881f45b6dbcf8c37586ff2c9939c614952
Deleted: sha256:8ef2719f33c1c3a77f67455a41ff50aa90582dca6c754bf4362ec1755ef70697
Deleted: sha256:571b0b18dbc5937f7c4edb7fc79cd3091e79251e42951cd67460497d44272176
Deleted: sha256:fe04087d24cb16b097048bf92fe682ecc64ffd4cfbb2becf1c6c7757f9c47ba0
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220317123039]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:9e2ac9cf49716cd991f377754cbb64532d5dd6ce720188b64d6dae3e6265e63b]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220317123039] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:9e2ac9cf49716cd991f377754cbb64532d5dd6ce720188b64d6dae3e6265e63b])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:9e2ac9cf49716cd991f377754cbb64532d5dd6ce720188b64d6dae3e6265e63b
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:9e2ac9cf49716cd991f377754cbb64532d5dd6ce720188b64d6dae3e6265e63b
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:9e2ac9cf49716cd991f377754cbb64532d5dd6ce720188b64d6dae3e6265e63b].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 33m 34s
109 actionable tasks: 72 executed, 33 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/qsanxfo3bf5ty

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #86

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/86/display/redirect?page=changes>

Changes:

[Chamikara Madhusanka Jayalath] Updates x-lang release validation to use staged jars

[dhuntsperger] documented maven-to-gradle conversion for Dataflow; refactored java

[dhuntsperger] adding a list of example pipelines

[dhuntsperger] removing unnecessary `ls` command from maven project generation

[dhuntsperger] fixing filename formatting in response to feedback

[dhuntsperger] adding extra step emphasizing runner setupt

[dhuntsperger] reorganized instructions to emphasize setup steps for runners

[noreply] [BEAM-13767] Move a bunch of python tasks to use gradle configuration…

[noreply] Merge pull request #17052 from [BEAM-13818] [SnowflakeIO] Add support

[noreply] Adding pydoc for StateHandler (#17091)

[noreply] BEAM-3165 Bypass split if numSplit is zero (#17084)


------------------------------------------
[...truncated 658.21 KB...]
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed throu
Mar 16, 2022 1:50:37 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
SEVERE: 2022-03-16T13:50:37.335Z: Staged package gax-grpc-2.8.1-Oaqma5KQUeKekzhWvuQ9BCsF3Uf1NF_DCF0Um3fxw3s.jar at location 'gs://temp-storage-for-perf-tests/loadtests/staging/gax-grpc-2.8.1-Oaqma5KQUeKekzhWvuQ9BCsF3Uf1NF_DCF0Um3fxw3s.jar' is inaccessible.
Mar 16, 2022 1:50:40 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
SEVERE: 2022-03-16T13:50:37.863Z: Staged package google-cloud-spanner-6.17.4-NvG_91cO2RfpEp9bgh1dl-N3SCcaKXMAoFmtVb0nBzY.jar at location 'gs://temp-storage-for-perf-tests/loadtests/staging/google-cloud-spanner-6.17.4-NvG_91cO2RfpEp9bgh1dl-N3SCcaKXMAoFmtVb0nBzY.jar' is inaccessible.
Mar 16, 2022 1:50:42 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
WARNING: 2022-03-16T13:50:41.910Z: One or more access checks for temp location or staged files failed. Please refer to other error messages for details. For more information on security and permissions, please see https://cloud.google.com/dataflow/security-and-permissions.
Mar 16, 2022 1:53:40 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
WARNING: 2022-03-16T13:53:40.223Z: One or more access checks for temp location or staged files failed. Please refer to other error messages for details. For more information on security and permissions, please see https://cloud.google.com/dataflow/security-and-permissions.
Mar 16, 2022 1:56:38 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
SEVERE: 2022-03-16T13:56:37.366Z: Staged package gax-grpc-2.8.1-Oaqma5KQUeKekzhWvuQ9BCsF3Uf1NF_DCF0Um3fxw3s.jar at location 'gs://temp-storage-for-perf-tests/loadtests/staging/gax-grpc-2.8.1-Oaqma5KQUeKekzhWvuQ9BCsF3Uf1NF_DCF0Um3fxw3s.jar' is inaccessible.
Mar 16, 2022 1:56:38 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
SEVERE: 2022-03-16T13:56:37.961Z: Staged package google-cloud-spanner-6.17.4-NvG_91cO2RfpEp9bgh1dl-N3SCcaKXMAoFmtVb0nBzY.jar at location 'gs://temp-storage-for-perf-tests/loadtests/staging/google-cloud-spanner-6.17.4-NvG_91cO2RfpEp9bgh1dl-N3SCcaKXMAoFmtVb0nBzY.jar' is inaccessible.
Mar 16, 2022 1:56:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
WARNING: 2022-03-16T13:56:40.389Z: One or more access checks for temp location or staged files failed. Please refer to other error messages for details. For more information on security and permissions, please see https://cloud.google.com/dataflow/security-and-permissions.
Mar 16, 2022 1:59:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
WARNING: 2022-03-16T13:59:40.349Z: One or more access checks for temp location or staged files failed. Please refer to other error messages for details. For more information on security and permissions, please see https://cloud.google.com/dataflow/security-and-permissions.
Mar 16, 2022 2:02:39 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
SEVERE: 2022-03-16T14:02:37.254Z: Staged package gax-grpc-2.8.1-Oaqma5KQUeKekzhWvuQ9BCsF3Uf1NF_DCF0Um3fxw3s.jar at location 'gs://temp-storage-for-perf-tests/loadtests/staging/gax-grpc-2.8.1-Oaqma5KQUeKekzhWvuQ9BCsF3Uf1NF_DCF0Um3fxw3s.jar' is inaccessible.
Mar 16, 2022 2:02:39 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
SEVERE: 2022-03-16T14:02:37.915Z: Staged package google-cloud-spanner-6.17.4-NvG_91cO2RfpEp9bgh1dl-N3SCcaKXMAoFmtVb0nBzY.jar at location 'gs://temp-storage-for-perf-tests/loadtests/staging/google-cloud-spanner-6.17.4-NvG_91cO2RfpEp9bgh1dl-N3SCcaKXMAoFmtVb0nBzY.jar' is inaccessible.
Mar 16, 2022 2:02:42 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
WARNING: 2022-03-16T14:02:41.257Z: One or more access checks for temp location or staged files failed. Please refer to other error messages for details. For more information on security and permissions, please see https://cloud.google.com/dataflow/security-and-permissions.
Mar 16, 2022 4:00:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-16T16:00:41.762Z: Cancel request is committed for workflow job: 2022-03-16_05_38_23-15995521566841701660.
Mar 16, 2022 4:00:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-16T16:00:41.842Z: Cleaning up.
Mar 16, 2022 4:00:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-16T16:00:41.907Z: Stopping **** pool...
Mar 16, 2022 4:00:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-16T16:00:41.969Z: Stopping **** pool...
Mar 16, 2022 4:03:04 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-16T16:03:03.977Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Mar 16, 2022 4:03:04 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-16T16:03:04.038Z: Worker pool stopped.
Mar 16, 2022 4:03:12 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-03-16_05_38_23-15995521566841701660 finished with status CANCELLED.
Load test results for test (ID): 6373d5ad-c367-4eef-84ef-5c58d43e9e5c and timestamp: 2022-03-16T12:38:15.003000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                 11998.798
dataflow_v2_java17_total_bytes_count             3.85616657E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220316123318
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:755b451e21bfd4401c18285e083a62275b5ce703da24181359eb4c944a253d7a
Deleted: sha256:99d7891d48cc443d6a358f439bcb816e8f63a5a4166eb66018f8aafb545cbe61
Deleted: sha256:742212c51b4fa0330dd442f77987923982fce8ec8a8f4b038269799f6e85fa5f
Deleted: sha256:4116c836156cce9c256a44381d775d3fb81874e482d27d41d4f03e6ab14b7413
Deleted: sha256:1b451daf9296fdd500b861c4bd69c7f0f056ea198a39c688648b29c03f2ae376
Deleted: sha256:653aa0a6eb689e8027d533e6b1ffec694dd347ef2c076fb85da91e4bf717b99f
Deleted: sha256:88803eae37b618e212a51bb786f5cba2e2146bd044bc547df57c1775a3f2be0c
Deleted: sha256:43c2bd393d3acc9a1b7c5265b971810b3b4806904ee29a14fb9ccb623b59574c
Deleted: sha256:161519c15be22dec82353e02d8cd659ea59c9e0968b732efd53b36a50f422a99
Deleted: sha256:df03f1ee02980630a6b5e0866e087b39f8feb8cd12825079e7a3bd06c6dab936
Deleted: sha256:8df9b413828447252c33c404bae3f208731b607bd0123fa6ff1a20e0f0bade79
Deleted: sha256:2eae11648cf0e833a0aa5fa537bdb137a6e803bb503c47a25987c3d6316bf00e
Deleted: sha256:a9127daf293d01d77c8b25679942ab522b83c39a0a09b2cd938a41e45c453c5d
Deleted: sha256:54412efe6e4c0989acd62b77f4c1b9d29e5b64c3a2255175c136d8c2d17ec0c0
Deleted: sha256:6342b180381f682ade34d6a59e41cd52dabdbe95c195a34a840fbb733705eeea
Deleted: sha256:5c6d687965d08f82ef52d826547684a4d9e5a80d9c57a3854dbdab64ba27c0e3
Deleted: sha256:5ad6714552931d9ee90e4453eef7d7465b8bf12d42a428ffb25188a2e6f50090
Deleted: sha256:b0e3593ac21cf088eeb970fc7a77a9c254f5de6108f36877fb2dc311c1e1d81b
Deleted: sha256:9c4f3c24c8a13abdf3c7b6a0ea62a4cf12ed36bb4f3e63bc6fb95b9b737dba53
Deleted: sha256:ad32abdb1ac0340b5efc9eaef5c7fa05853d478c58d7db41259b15b74d473b18
Deleted: sha256:037a659cddcf04b255fefdd01d5d2766058e7e9860fa4b2e1127813361981af2
Deleted: sha256:ed1326f60212ece16cda2a087309a7489ea34b1801b4bbac1830be1349036bd6
Deleted: sha256:7766963ebeec99526aca40459c6141564f8c861b501be2da419e69ff433fa5dd
Deleted: sha256:b3972f01df7b241581aca4ca02d6df52ed4360e2748c8c39dd4ab9da00c8fde4
Deleted: sha256:b2707cc50924e988bea57821522349ce7842076e4c796871986b3fc39bebfad7
Deleted: sha256:a79a39c1c763f77f22fd89300af620927a4cfda3bc8451d95b74df0d5a29a1a9
Deleted: sha256:304cec98013b85aef225aac5ac3a16eee9ed7f09e0b6fda43e26d17126339de4
Deleted: sha256:cb21fddc667612b85b76af32fdb4426c6cc2ba8a58656a9e6e38b442a4f5628e
Deleted: sha256:90e38ca5281d9c7081f5b891ce22171db61d6ef2995d74fcecb098cb534b70ac
Deleted: sha256:af2c7163de2757cbc53dc96b92806e22df6a4cdee2de2b0b263d0f822957e057
Deleted: sha256:79ffe215c22a97517a0eb0c863ba18fd7adf1b029639d9bfaa2cc4f6ea813a53
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220316123318]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:755b451e21bfd4401c18285e083a62275b5ce703da24181359eb4c944a253d7a]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220316123318] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:755b451e21bfd4401c18285e083a62275b5ce703da24181359eb4c944a253d7a])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:755b451e21bfd4401c18285e083a62275b5ce703da24181359eb4c944a253d7a
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:755b451e21bfd4401c18285e083a62275b5ce703da24181359eb4c944a253d7a
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:755b451e21bfd4401c18285e083a62275b5ce703da24181359eb4c944a253d7a].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 31m 29s
109 actionable tasks: 72 executed, 33 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/iccchbawqaoje

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 - Build # 85 - Aborted!

Posted by Apache Jenkins Server <je...@builds.apache.org>.
beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 - Build # 85 - Aborted:

Check console output at https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/85/ to view the results.

Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #84

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/84/display/redirect>

Changes:


------------------------------------------
[...truncated 1.03 MB...]
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmi
Mar 14, 2022 4:03:14 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-14T16:03:13.308Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Mar 14, 2022 4:03:14 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-14T16:03:13.345Z: Worker pool stopped.
Mar 14, 2022 4:03:22 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-03-14_05_33_22-5994472670153606868 finished with status CANCELLED.
Load test results for test (ID): 755c3d5b-44c9-4bec-8c07-b955e3a3e16c and timestamp: 2022-03-14T12:33:16.763000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                 12291.923
dataflow_v2_java17_total_bytes_count             3.20948431E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220314123043
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:4494b5e74b9b3a018ef73ee9d3e1a5b7681e9f9d818a7a07978a20750a5f3529
Deleted: sha256:3f0803366264330082a3949bdb8a5e8af4d18e8ca1894fae3980709c98b4fa36
Deleted: sha256:8f3fa83055dbdc3cabb7c9ca418f3d99c55f3e86f220d0f6b47daf3a40a55f20
Deleted: sha256:5e6c658db4f0818a3bbf5354367cffaffa2dccf13fc065bde2c30c09a776d5d2
Deleted: sha256:55d805fdfe0a82f734c7f1663e6f708014ce9fdd113f90aad966e86ca132507e
Deleted: sha256:5f58cd23442369c10685510c5ac6f404c7d2a2c537733709926c4d50660be673
Deleted: sha256:88aa3c8028fdcbff0248a4816ef48fac379c75614cad6f5b2814b39fc6b11ab8
Deleted: sha256:ce2d78eb7d7084d58c781967895a553171c4401a343f4d5044c7a6680e78d1e1
Deleted: sha256:2748acea1ab4a9b2cf96603b162ca4aea004da28ac2cd4c1496f75b1faf0a4fe
Deleted: sha256:7a0d38ce59c68f4999c7e2d2884e4260df546d8c1f74cc42a60543cf58cad0c6
Deleted: sha256:e0665d422cbbf899498a389630970000e48e35eeba43edc0c4a30052f8334699
Deleted: sha256:e52ab3b1770b9f3f2f558295e9a571224fff8c610f901ef45317f0e706c9f243
Deleted: sha256:09baf9236230dc15506a934a8cad4f533efe004bcf10c1ebc254dbffd9f7b840
Deleted: sha256:3225ce81b32ba81b7b7b752a0ee872ef83b1ea499368603b1b253ddec4165675
Deleted: sha256:add856f869d7c6334e616774cff79e1b0ba7c53d89d59a8b61f443f4196f619f
Deleted: sha256:eaf24fa0599a9504e31d71ffd4bd101d82390c6f7e423985624a3f1cde41366a
Deleted: sha256:a882f9c3389f8f278a5b40d2ddd3a27877e22dfd859bb9f80686042232d145ee
Deleted: sha256:f57f476743b1ab54865afbca59f1cd1c8fd4fa98d0b092d80061971840be6b69
Deleted: sha256:32cfbeaf134e7ead0a7c833ac2bf5ba35f165881b459e36c5fcdf04f5d7037e9
Deleted: sha256:c11b3eedbfeb478edbfcfc43c81bfbfc5a37ded7a882247b96ad816d9501b80c
Deleted: sha256:636c57ff482cdf8dc74463f38c8fdb57932f8e918444fa6596251d3028025d99
Deleted: sha256:ea8be04eb61504d0dddb3519c188d0a7c6f6ce8d75a26396e00eeddde2e26832
Deleted: sha256:03664ecae282030087c87baf9a89ba600704b7d01300415d2356aacea14d068a
Deleted: sha256:c3fcf65d8df684ca9326ca9e902fd9d74718b19a81800337e140426f99395ff0
Deleted: sha256:2fa50efdde0460911f25b48a71fe6e3d1b1452291fd92d5067fc3dd630555aa7
Deleted: sha256:f353603293c9c360b70caf894a5dac4fd704d1335e9da528f66b13efc89aae2c
Deleted: sha256:dd5b0305221c3518f9c07d8067bc7a3b7f1013f27d5d95bca63eafa60b186956
Deleted: sha256:3ed8e24bdca55f926a8fc5b089a1687759bfa4ca6b56e8076b2f26a300b4f492
Deleted: sha256:681d9370c2f60dbc5ef254d8488958adf860ffa69ffb17af20079fc28d4cf064
Deleted: sha256:69dfb61db43717e4c5e1b5d62165ab52e648d3ad2db107baa6448c77c77f9073
Deleted: sha256:5f802981645c8c09fbc26b8b173c2b83db6aaedfa46b75a7af8e4a894a38d83f
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220314123043]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:4494b5e74b9b3a018ef73ee9d3e1a5b7681e9f9d818a7a07978a20750a5f3529]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220314123043] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:4494b5e74b9b3a018ef73ee9d3e1a5b7681e9f9d818a7a07978a20750a5f3529])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:4494b5e74b9b3a018ef73ee9d3e1a5b7681e9f9d818a7a07978a20750a5f3529
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:4494b5e74b9b3a018ef73ee9d3e1a5b7681e9f9d818a7a07978a20750a5f3529
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:4494b5e74b9b3a018ef73ee9d3e1a5b7681e9f9d818a7a07978a20750a5f3529].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 33m 3s
109 actionable tasks: 72 executed, 33 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/uaqmdwpha2lv4

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #83

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/83/display/redirect?page=changes>

Changes:

[noreply] [BEAM-14072] [BEAM-13993] [BEAM-10039] Import beam plugins before


------------------------------------------
[...truncated 57.75 KB...]
Mar 13, 2022 12:32:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-13T12:32:39.698Z: Worker configuration: e2-standard-2 in us-central1-b.
Mar 13, 2022 12:32:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-13T12:32:40.434Z: Expanding SplittableParDo operations into optimizable parts.
Mar 13, 2022 12:32:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-13T12:32:40.472Z: Expanding CollectionToSingleton operations into optimizable parts.
Mar 13, 2022 12:32:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-13T12:32:40.529Z: Expanding CoGroupByKey operations into optimizable parts.
Mar 13, 2022 12:32:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-13T12:32:40.596Z: Expanding SplittableProcessKeyed operations into optimizable parts.
Mar 13, 2022 12:32:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-13T12:32:40.627Z: Expanding GroupByKey operations into streaming Read/Write steps
Mar 13, 2022 12:32:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-13T12:32:40.693Z: Lifting ValueCombiningMappingFns into MergeBucketsMappingFns
Mar 13, 2022 12:32:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-13T12:32:40.785Z: Fusing adjacent ParDo, Read, Write, and Flatten operations
Mar 13, 2022 12:32:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-13T12:32:40.821Z: Unzipping flatten CoGroupByKey-Flatten for input CoGroupByKey-MakeUnionTable0-ParMultiDo-ConstructUnionTable-.output
Mar 13, 2022 12:32:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-13T12:32:40.842Z: Fusing unzipped copy of CoGroupByKey/GBK/WriteStream, through flatten CoGroupByKey/Flatten, into producer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable)
Mar 13, 2022 12:32:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-13T12:32:40.865Z: Fusing consumer CoGroupByKey/GBK/WriteStream into CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable)
Mar 13, 2022 12:32:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-13T12:32:40.898Z: Fusing consumer Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read input/Impulse
Mar 13, 2022 12:32:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-13T12:32:40.918Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
Mar 13, 2022 12:32:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-13T12:32:40.951Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
Mar 13, 2022 12:32:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-13T12:32:40.986Z: Fusing consumer Read input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
Mar 13, 2022 12:32:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-13T12:32:41.015Z: Fusing consumer Collect start time metrics (input)/ParMultiDo(TimeMonitor) into Read input/ParDo(StripIds)/ParMultiDo(StripIds)
Mar 13, 2022 12:32:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-13T12:32:41.050Z: Fusing consumer Window.Into()/Window.Assign into Collect start time metrics (input)/ParMultiDo(TimeMonitor)
Mar 13, 2022 12:32:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-13T12:32:41.124Z: Fusing consumer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable) into Window.Into()/Window.Assign
Mar 13, 2022 12:32:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-13T12:32:41.158Z: Fusing consumer Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read co-input/Impulse
Mar 13, 2022 12:32:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-13T12:32:41.202Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
Mar 13, 2022 12:32:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-13T12:32:41.233Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
Mar 13, 2022 12:32:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-13T12:32:41.256Z: Fusing consumer Read co-input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
Mar 13, 2022 12:32:44 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-13T12:32:41.289Z: Fusing consumer Collect start time metrics (co-input)/ParMultiDo(TimeMonitor) into Read co-input/ParDo(StripIds)/ParMultiDo(StripIds)
Mar 13, 2022 12:32:44 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-13T12:32:41.318Z: Fusing consumer Window.Into()2/Window.Assign into Collect start time metrics (co-input)/ParMultiDo(TimeMonitor)
Mar 13, 2022 12:32:44 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-13T12:32:41.343Z: Fusing consumer CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable) into Window.Into()2/Window.Assign
Mar 13, 2022 12:32:44 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-13T12:32:41.380Z: Fusing consumer CoGroupByKey/GBK/MergeBuckets into CoGroupByKey/GBK/ReadStream
Mar 13, 2022 12:32:44 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-13T12:32:41.410Z: Fusing consumer CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult) into CoGroupByKey/GBK/MergeBuckets
Mar 13, 2022 12:32:44 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-13T12:32:41.437Z: Fusing consumer Ungroup and reiterate/ParMultiDo(UngroupAndReiterate) into CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult)
Mar 13, 2022 12:32:44 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-13T12:32:41.467Z: Fusing consumer Collect total bytes/ParMultiDo(ByteMonitor) into Ungroup and reiterate/ParMultiDo(UngroupAndReiterate)
Mar 13, 2022 12:32:44 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-13T12:32:41.491Z: Fusing consumer Collect end time metrics/ParMultiDo(TimeMonitor) into Collect total bytes/ParMultiDo(ByteMonitor)
Mar 13, 2022 12:32:44 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-13T12:32:41.636Z: Running job using Streaming Engine
Mar 13, 2022 12:32:44 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-13T12:32:41.853Z: Starting 5 ****s in us-central1-b...
Mar 13, 2022 12:32:50 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-13T12:32:48.686Z: Your project already contains 100 Dataflow-created metric descriptors, so new user metrics of the form custom.googleapis.com/* will not be created. However, all user metrics are also available in the metric dataflow.googleapis.com/job/user_counter. If you rely on the custom metrics, you can delete old / unused metric descriptors. See https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.list and https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.delete
Mar 13, 2022 12:33:22 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-13T12:33:21.484Z: Autoscaling: Raised the number of ****s to 5 so that the pipeline can catch up with its backlog and keep up with its input rate.
Mar 13, 2022 12:34:20 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-13T12:34:19.392Z: Workers have started successfully.
Mar 13, 2022 12:34:20 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-13T12:34:19.434Z: Workers have started successfully.
Mar 13, 2022 12:37:22 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
SEVERE: 2022-03-13T12:37:20.675Z: generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
Mar 13, 2022 4:00:35 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-13T16:00:34.486Z: Cancel request is committed for workflow job: 2022-03-13_05_32_28-11238988837831007706.
Mar 13, 2022 4:00:35 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-13T16:00:34.560Z: Cleaning up.
Mar 13, 2022 4:00:35 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-13T16:00:34.728Z: Stopping **** pool...
Mar 13, 2022 4:00:35 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-13T16:00:34.783Z: Stopping **** pool...
Mar 13, 2022 4:03:04 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-13T16:03:01.939Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Mar 13, 2022 4:03:04 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-13T16:03:01.987Z: Worker pool stopped.
Mar 13, 2022 4:03:08 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-03-13_05_32_28-11238988837831007706 finished with status CANCELLED.
Load test results for test (ID): 944cca84-7605-4d25-bc23-397d46a9fe07 and timestamp: 2022-03-13T12:32:23.515000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                 12345.032
dataflow_v2_java17_total_bytes_count             1.99644701E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220313123035
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:8382c89a8536286988a3d1425f2629a9ac3977cc9ea64e5a577076832009de78
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220313123035]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:8382c89a8536286988a3d1425f2629a9ac3977cc9ea64e5a577076832009de78]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220313123035] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:8382c89a8536286988a3d1425f2629a9ac3977cc9ea64e5a577076832009de78])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:8382c89a8536286988a3d1425f2629a9ac3977cc9ea64e5a577076832009de78
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:8382c89a8536286988a3d1425f2629a9ac3977cc9ea64e5a577076832009de78
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:8382c89a8536286988a3d1425f2629a9ac3977cc9ea64e5a577076832009de78].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 32m 54s
109 actionable tasks: 72 executed, 33 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/gkvnhrapq5la2

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #82

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/82/display/redirect?page=changes>

Changes:

[Ismaël Mejía] [BEAM-13981] Remove Spark Runner specific code for event logging

[vitaly.terentyev] [BEAM-2766] Support null key/values in HadoopFormatIO

[vitaly.terentyev] [BEAM-2766] Fix checkstyle


------------------------------------------
[...truncated 555.24 KB...]
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is
Mar 12, 2022 4:00:42 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-12T16:00:41.726Z: Cancel request is committed for workflow job: 2022-03-12_04_35_47-11478298439972272621.
Mar 12, 2022 4:00:42 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-12T16:00:41.762Z: Cleaning up.
Mar 12, 2022 4:00:42 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-12T16:00:41.819Z: Stopping **** pool...
Mar 12, 2022 4:00:42 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-12T16:00:41.863Z: Stopping **** pool...
Mar 12, 2022 4:03:10 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-12T16:03:09.336Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Mar 12, 2022 4:03:10 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-12T16:03:09.378Z: Worker pool stopped.
Mar 12, 2022 4:03:15 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-03-12_04_35_47-11478298439972272621 finished with status CANCELLED.
Load test results for test (ID): 0e3fa2ff-3f83-438a-abd7-331f342b6fcd and timestamp: 2022-03-12T12:35:42.591000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                 12158.628
dataflow_v2_java17_total_bytes_count             3.43665774E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220312123324
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:e71c98942309e00c066bdcd90ffe47e1bf70a4d7affce29b92d2fda3e97f8040
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220312123324]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:e71c98942309e00c066bdcd90ffe47e1bf70a4d7affce29b92d2fda3e97f8040]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220312123324] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:e71c98942309e00c066bdcd90ffe47e1bf70a4d7affce29b92d2fda3e97f8040])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:e71c98942309e00c066bdcd90ffe47e1bf70a4d7affce29b92d2fda3e97f8040
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:e71c98942309e00c066bdcd90ffe47e1bf70a4d7affce29b92d2fda3e97f8040
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:e71c98942309e00c066bdcd90ffe47e1bf70a4d7affce29b92d2fda3e97f8040].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 30m 12s
109 actionable tasks: 72 executed, 33 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/qnuq5zwokozpk

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #81

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/81/display/redirect?page=changes>

Changes:

[ihr] [BEAM-13923] Fix the answers placeholders locations in the Java katas

[jakub.kukul] [BEAM-14039] Propagate ignore_unknown_columns parameter.

[stranniknm] [BEAM-14079] playground - improve accessibility

[noreply] [BEAM-13925] Find and address prs that havent been reviewed in a week

[noreply] Fix import path

[noreply] [BEAM-13925] Fix one more import path

[noreply] Add a StatefulDoFn test that sets event time timer within allowed

[noreply] Merge pull request #17056 from [BEAM-14076] [SnowflakeIO] Add support


------------------------------------------
[...truncated 443.50 KB...]
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmi
Mar 11, 2022 4:00:44 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-11T16:00:43.962Z: Cancel request is committed for workflow job: 2022-03-11_04_44_02-13238566233246168968.
Mar 11, 2022 4:00:44 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-11T16:00:44.050Z: Cleaning up.
Mar 11, 2022 4:00:44 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-11T16:00:44.137Z: Stopping **** pool...
Mar 11, 2022 4:00:44 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-11T16:00:44.203Z: Stopping **** pool...
Mar 11, 2022 4:03:18 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-11T16:03:17.258Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Mar 11, 2022 4:03:18 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-11T16:03:17.450Z: Worker pool stopped.
Mar 11, 2022 4:03:23 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-03-11_04_44_02-13238566233246168968 finished with status CANCELLED.
Load test results for test (ID): 56bff4b6-c267-4fda-be4e-74e9b5466fff and timestamp: 2022-03-11T12:43:39.169000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                 11660.004
dataflow_v2_java17_total_bytes_count             3.25241189E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220311123850
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:a9fb7aecee1e4a55bbd67e4e0e63ec645008068ef744805414f1af6e7ec43afc
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220311123850]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:a9fb7aecee1e4a55bbd67e4e0e63ec645008068ef744805414f1af6e7ec43afc]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220311123850] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:a9fb7aecee1e4a55bbd67e4e0e63ec645008068ef744805414f1af6e7ec43afc])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:a9fb7aecee1e4a55bbd67e4e0e63ec645008068ef744805414f1af6e7ec43afc
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:a9fb7aecee1e4a55bbd67e4e0e63ec645008068ef744805414f1af6e7ec43afc
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:a9fb7aecee1e4a55bbd67e4e0e63ec645008068ef744805414f1af6e7ec43afc].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 25m 55s
109 actionable tasks: 72 executed, 33 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/tnfrvmym766ai

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #80

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/80/display/redirect?page=changes>

Changes:

[hengfeng] [BEAM-12164]: display the metadata table's name on UI

[noreply] Revert "[BEAM-13993] [BEAM-10039] Import beam plugins before starting

[noreply] Merge pull request #17036 from [BEAM-12164] Convert all static instances

[noreply] fix variable reference (#16991)

[noreply] Merge pull request #16844 from [BEAM-12164]: allow for nanosecond

[noreply] [BEAM-13904] Increase unit testing in the reflectx package (#17024)


------------------------------------------
[...truncated 62.89 KB...]
  File "/usr/lib/python3.8/urllib/request.py", line 1383, in http_open
    return self.do_open(http.client.HTTPConnection, req)
  File "/usr/lib/python3.8/urllib/request.py", line 1358, in do_open
    r = h.getresponse()
  File "/usr/lib/python3.8/http/client.py", line 1348, in getresponse
    response.begin()
  File "/usr/lib/python3.8/http/client.py", line 316, in begin
    version, status, reason = self._read_status()
  File "/usr/lib/python3.8/http/client.py", line 277, in _read_status
    line = str(self.fp.readline(_MAXLINE + 1), "iso-8859-1")
  File "/usr/lib/python3.8/socket.py", line 669, in readinto
    return self._sock.recv_into(b)
ConnectionResetError: [Errno 104] Connection reset by peer
ERROR:root:Error occurred when pull <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/build/target/java_third_party_licenses/checkstyle-8.23.jar/LICENSE> from http://www.gnu.org/licenses/old-licenses/lgpl-2.1.txt for checkstyle-8.23. Retrying...
Traceback (most recent call last):
  File "/usr/lib/python3.8/urllib/request.py", line 1354, in do_open
    h.request(req.get_method(), req.selector, req.data, headers,
  File "/usr/lib/python3.8/http/client.py", line 1256, in request
    self._send_request(method, url, body, headers, encode_chunked)
  File "/usr/lib/python3.8/http/client.py", line 1302, in _send_request
    self.endheaders(body, encode_chunked=encode_chunked)
  File "/usr/lib/python3.8/http/client.py", line 1251, in endheaders
    self._send_output(message_body, encode_chunked=encode_chunked)
  File "/usr/lib/python3.8/http/client.py", line 1011, in _send_output
    self.send(msg)
  File "/usr/lib/python3.8/http/client.py", line 951, in send
    self.connect()
  File "/usr/lib/python3.8/http/client.py", line 922, in connect
    self.sock = self._create_connection(
  File "/usr/lib/python3.8/socket.py", line 808, in create_connection
    raise err
  File "/usr/lib/python3.8/socket.py", line 796, in create_connection
    sock.connect(sa)
OSError: [Errno 101] Network is unreachable

During handling of the above exception, another exception occurred:

Traceback (most recent call last):
  File "<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/license_scripts/pull_licenses_java.py",> line 58, in pull_from_url
    url_read = urlopen(url)
  File "/usr/lib/python3.8/urllib/request.py", line 222, in urlopen
    return opener.open(url, data, timeout)
  File "/usr/lib/python3.8/urllib/request.py", line 525, in open
    response = self._open(req, data)
  File "/usr/lib/python3.8/urllib/request.py", line 542, in _open
    result = self._call_chain(self.handle_open, protocol, protocol +
  File "/usr/lib/python3.8/urllib/request.py", line 502, in _call_chain
    result = func(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 1383, in http_open
    return self.do_open(http.client.HTTPConnection, req)
  File "/usr/lib/python3.8/urllib/request.py", line 1357, in do_open
    raise URLError(err)
urllib.error.URLError: <urlopen error [Errno 101] Network is unreachable>
ERROR:root:Invalid url for jFormatString-3.0.0: http://www.gnu.org/licenses/lgpl.html. Retrying...
Traceback (most recent call last):
  File "/usr/lib/python3.8/urllib/request.py", line 1354, in do_open
    h.request(req.get_method(), req.selector, req.data, headers,
  File "/usr/lib/python3.8/http/client.py", line 1256, in request
    self._send_request(method, url, body, headers, encode_chunked)
  File "/usr/lib/python3.8/http/client.py", line 1302, in _send_request
    self.endheaders(body, encode_chunked=encode_chunked)
  File "/usr/lib/python3.8/http/client.py", line 1251, in endheaders
    self._send_output(message_body, encode_chunked=encode_chunked)
  File "/usr/lib/python3.8/http/client.py", line 1011, in _send_output
    self.send(msg)
  File "/usr/lib/python3.8/http/client.py", line 951, in send
    self.connect()
  File "/usr/lib/python3.8/http/client.py", line 922, in connect
    self.sock = self._create_connection(
  File "/usr/lib/python3.8/socket.py", line 808, in create_connection
    raise err
  File "/usr/lib/python3.8/socket.py", line 796, in create_connection
    sock.connect(sa)
OSError: [Errno 101] Network is unreachable

During handling of the above exception, another exception occurred:

Traceback (most recent call last):
  File "<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/license_scripts/pull_licenses_java.py",> line 58, in pull_from_url
    url_read = urlopen(url)
  File "/usr/lib/python3.8/urllib/request.py", line 222, in urlopen
    return opener.open(url, data, timeout)
  File "/usr/lib/python3.8/urllib/request.py", line 525, in open
    response = self._open(req, data)
  File "/usr/lib/python3.8/urllib/request.py", line 542, in _open
    result = self._call_chain(self.handle_open, protocol, protocol +
  File "/usr/lib/python3.8/urllib/request.py", line 502, in _call_chain
    result = func(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 1383, in http_open
    return self.do_open(http.client.HTTPConnection, req)
  File "/usr/lib/python3.8/urllib/request.py", line 1357, in do_open
    raise URLError(err)
urllib.error.URLError: <urlopen error [Errno 101] Network is unreachable>
ERROR:root:Invalid url for jFormatString-3.0.0: http://www.gnu.org/licenses/lgpl.html. Retrying...
Traceback (most recent call last):
  File "<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/license_scripts/pull_licenses_java.py",> line 58, in pull_from_url
    url_read = urlopen(url)
  File "/usr/lib/python3.8/urllib/request.py", line 222, in urlopen
    return opener.open(url, data, timeout)
  File "/usr/lib/python3.8/urllib/request.py", line 525, in open
    response = self._open(req, data)
  File "/usr/lib/python3.8/urllib/request.py", line 542, in _open
    result = self._call_chain(self.handle_open, protocol, protocol +
  File "/usr/lib/python3.8/urllib/request.py", line 502, in _call_chain
    result = func(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 1383, in http_open
    return self.do_open(http.client.HTTPConnection, req)
  File "/usr/lib/python3.8/urllib/request.py", line 1358, in do_open
    r = h.getresponse()
  File "/usr/lib/python3.8/http/client.py", line 1348, in getresponse
    response.begin()
  File "/usr/lib/python3.8/http/client.py", line 316, in begin
    version, status, reason = self._read_status()
  File "/usr/lib/python3.8/http/client.py", line 277, in _read_status
    line = str(self.fp.readline(_MAXLINE + 1), "iso-8859-1")
  File "/usr/lib/python3.8/socket.py", line 669, in readinto
    return self._sock.recv_into(b)
ConnectionResetError: [Errno 104] Connection reset by peer
ERROR:root:Error occurred when pull <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/build/target/java_third_party_licenses/jFormatString-3.0.0.jar/LICENSE> from http://www.gnu.org/licenses/lgpl.html for jFormatString-3.0.0. Retrying...
Traceback (most recent call last):
  File "/usr/lib/python3.8/urllib/request.py", line 1354, in do_open
    h.request(req.get_method(), req.selector, req.data, headers,
  File "/usr/lib/python3.8/http/client.py", line 1256, in request
    self._send_request(method, url, body, headers, encode_chunked)
  File "/usr/lib/python3.8/http/client.py", line 1302, in _send_request
    self.endheaders(body, encode_chunked=encode_chunked)
  File "/usr/lib/python3.8/http/client.py", line 1251, in endheaders
    self._send_output(message_body, encode_chunked=encode_chunked)
  File "/usr/lib/python3.8/http/client.py", line 1011, in _send_output
    self.send(msg)
  File "/usr/lib/python3.8/http/client.py", line 951, in send
    self.connect()
  File "/usr/lib/python3.8/http/client.py", line 922, in connect
    self.sock = self._create_connection(
  File "/usr/lib/python3.8/socket.py", line 808, in create_connection
    raise err
  File "/usr/lib/python3.8/socket.py", line 796, in create_connection
    sock.connect(sa)
OSError: [Errno 101] Network is unreachable

During handling of the above exception, another exception occurred:

Traceback (most recent call last):
  File "<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/license_scripts/pull_licenses_java.py",> line 58, in pull_from_url
    url_read = urlopen(url)
  File "/usr/lib/python3.8/urllib/request.py", line 222, in urlopen
    return opener.open(url, data, timeout)
  File "/usr/lib/python3.8/urllib/request.py", line 525, in open
    response = self._open(req, data)
  File "/usr/lib/python3.8/urllib/request.py", line 542, in _open
    result = self._call_chain(self.handle_open, protocol, protocol +
  File "/usr/lib/python3.8/urllib/request.py", line 502, in _call_chain
    result = func(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 1383, in http_open
    return self.do_open(http.client.HTTPConnection, req)
  File "/usr/lib/python3.8/urllib/request.py", line 1357, in do_open
    raise URLError(err)
urllib.error.URLError: <urlopen error [Errno 101] Network is unreachable>
ERROR:root:Invalid url for jFormatString-3.0.0: http://www.gnu.org/licenses/lgpl.html. Retrying...
Traceback (most recent call last):
  File "/usr/lib/python3.8/urllib/request.py", line 1354, in do_open
    h.request(req.get_method(), req.selector, req.data, headers,
  File "/usr/lib/python3.8/http/client.py", line 1256, in request
    self._send_request(method, url, body, headers, encode_chunked)
  File "/usr/lib/python3.8/http/client.py", line 1302, in _send_request
    self.endheaders(body, encode_chunked=encode_chunked)
  File "/usr/lib/python3.8/http/client.py", line 1251, in endheaders
    self._send_output(message_body, encode_chunked=encode_chunked)
  File "/usr/lib/python3.8/http/client.py", line 1011, in _send_output
    self.send(msg)
  File "/usr/lib/python3.8/http/client.py", line 951, in send
    self.connect()
  File "/usr/lib/python3.8/http/client.py", line 922, in connect
    self.sock = self._create_connection(
  File "/usr/lib/python3.8/socket.py", line 808, in create_connection
    raise err
  File "/usr/lib/python3.8/socket.py", line 796, in create_connection
    sock.connect(sa)
OSError: [Errno 101] Network is unreachable

During handling of the above exception, another exception occurred:

Traceback (most recent call last):
  File "<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/license_scripts/pull_licenses_java.py",> line 58, in pull_from_url
    url_read = urlopen(url)
  File "/usr/lib/python3.8/urllib/request.py", line 222, in urlopen
    return opener.open(url, data, timeout)
  File "/usr/lib/python3.8/urllib/request.py", line 531, in open
    response = meth(req, response)
  File "/usr/lib/python3.8/urllib/request.py", line 640, in http_response
    response = self.parent.error(
  File "/usr/lib/python3.8/urllib/request.py", line 563, in error
    result = self._call_chain(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 502, in _call_chain
    result = func(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 755, in http_error_302
    return self.parent.open(new, timeout=req.timeout)
  File "/usr/lib/python3.8/urllib/request.py", line 525, in open
    response = self._open(req, data)
  File "/usr/lib/python3.8/urllib/request.py", line 542, in _open
    result = self._call_chain(self.handle_open, protocol, protocol +
  File "/usr/lib/python3.8/urllib/request.py", line 502, in _call_chain
    result = func(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 1383, in http_open
    return self.do_open(http.client.HTTPConnection, req)
  File "/usr/lib/python3.8/urllib/request.py", line 1357, in do_open
    raise URLError(err)
urllib.error.URLError: <urlopen error [Errno 101] Network is unreachable>
ERROR:root:Invalid url for jFormatString-3.0.0: http://www.gnu.org/licenses/lgpl.html after 9 retries.
ERROR:root:['jFormatString-3.0.0']
ERROR:root:**************************************** Licenses were not able to be pulled automatically for some dependencies. Please search source code of the dependencies on the internet and add "license" and "notice" (if available) field to <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/license_scripts/dep_urls_java.yaml> for each missing license. Dependency List: [jFormatString-3.0.0]
INFO:root:pull_licenses_java.py failed. It took 1224.162859 seconds with 16 threads.
Traceback (most recent call last):
  File "<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/license_scripts/pull_licenses_java.py",> line 321, in <module>
    raise RuntimeError('{n} error(s) occurred.'.format(n=len(error_msg)),
RuntimeError: ('1 error(s) occurred.', ['**************************************** Licenses were not able to be pulled automatically for some dependencies. Please search source code of the dependencies on the internet and add "license" and "notice" (if available) field to <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/license_scripts/dep_urls_java.yaml> for each missing license. Dependency List: [jFormatString-3.0.0]'])

> Task :sdks:java:container:pullLicenses FAILED

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:container:pullLicenses'.
> Process 'command './license_scripts/license_script.sh'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 20m 59s
103 actionable tasks: 66 executed, 33 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/ircfssnvqdfhk

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #79

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/79/display/redirect?page=changes>

Changes:

[Robert Bradshaw] Update dataflow API client.

[Robert Bradshaw] Instructions for updating apitools generated files.

[noreply] Merge pull request #17027: [BEAM-11205] Upgrade GCP Libraries BOM

[noreply] [BEAM-13709] Inconsistent behavior when parsing boolean flags across

[noreply] [BEAM-10976] Bundle finalization: Harness and some exec changes (#16980)

[noreply] Merge pull request #16976 from [BEAM-14010] [Website] Add Playground

[noreply] [BEAM-12447] Upgrade cloud build client and add/cleanup options (#17032)


------------------------------------------
[...truncated 163.31 KB...]
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmi
Mar 09, 2022 4:03:32 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-09T16:03:31.800Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Mar 09, 2022 4:03:32 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-09T16:03:31.842Z: Worker pool stopped.
Mar 09, 2022 4:03:38 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-03-09_04_34_52-17954488066678003077 finished with status CANCELLED.
Load test results for test (ID): 855fcf78-cbe0-4e64-86fb-2872ada99183 and timestamp: 2022-03-09T12:34:47.150000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                 12224.009
dataflow_v2_java17_total_bytes_count             2.94236799E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220309123252
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:af958bd776e36b12bff13d15d729e1d1d1b5155d1e29eb75425a2327db32fef0
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220309123252]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:af958bd776e36b12bff13d15d729e1d1d1b5155d1e29eb75425a2327db32fef0]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220309123252] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:af958bd776e36b12bff13d15d729e1d1d1b5155d1e29eb75425a2327db32fef0])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:af958bd776e36b12bff13d15d729e1d1d1b5155d1e29eb75425a2327db32fef0
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:af958bd776e36b12bff13d15d729e1d1d1b5155d1e29eb75425a2327db32fef0
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:af958bd776e36b12bff13d15d729e1d1d1b5155d1e29eb75425a2327db32fef0].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 31m 8s
109 actionable tasks: 72 executed, 33 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/3dnyr5wsgqvsk

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #78

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/78/display/redirect?page=changes>

Changes:

[dannymccormick] [BEAM-11085] Test that windows are correctly observed in DoFns

[jrmccluskey] [BEAM-14050] Update taxi.go example instructions

[noreply] Give pr bot write permissions on pr update

[noreply] Adding a logical type for Schemas using proto serialization. (#16940)

[noreply] BEAM-13765 missing PAssert methods (#16668)

[noreply] [BEAM-13909] improve coverage of Provision package (#17014)


------------------------------------------
[...truncated 1.22 MB...]
==>
    dist_proc/dax/workflow/****/streaming/fnapi_streaming_operators.cc:439
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
==>
    dist_proc/dax/workflow/****/streaming/fnapi_streaming_operators.cc:439
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
==>
    dist_proc/dax/workflow/****/streaming/fnapi_streaming_operators.cc:439
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
==>
    dist_proc/dax/workflow/****/streaming/fnapi_streaming_operators.cc:439
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
==>
    dist_proc/dax/workflow/****/streaming/fnapi_streaming_operators.cc:439
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
==>
    dist_proc/dax/workflow/****/streaming/fnapi_streaming_operators.cc:439
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
==>
    dist_proc/dax/workflow/****/streaming/fnapi_streaming_operators.cc:439
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
==>
    dist_proc/dax/workflow/****/streaming/fnapi_streaming_operators.cc:439
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
==>
    dist_proc/dax/workflow/****/streaming/fnapi_streaming_operators.cc:439
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
==>
    dist_proc/dax/workflow/****/streaming/fnapi_streaming_operators.cc:439
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
==>
    dist_proc/dax/workflow/****/streaming/fnapi_streaming_operators.cc:439
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
==>
    dist_proc/dax/workflow/****/streaming/fnapi_streaming_operators.cc:439
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
==>
    dist_proc/dax/workflow/****/streaming/fnapi_streaming_operators.cc:439
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
==>
    dist_proc/dax/workflow/****/streaming/fnapi_streaming_operators.cc:439
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
=
Mar 08, 2022 4:10:18 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-08T16:10:18.438Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Mar 08, 2022 4:10:18 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-08T16:10:18.479Z: Worker pool stopped.
Mar 08, 2022 4:10:25 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-03-08_04_36_53-765526506217188100 finished with status CANCELLED.
Load test results for test (ID): c0ebca97-cc88-40bc-865e-feb1ec091c69 and timestamp: 2022-03-08T12:36:48.185000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                 12496.256
dataflow_v2_java17_total_bytes_count             4.31670887E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220308123447
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:c562245b098ee175f52ef8d20fa37e2cd98bc5b7ffa6fe2a514eb10f1e7586da
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220308123447]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:c562245b098ee175f52ef8d20fa37e2cd98bc5b7ffa6fe2a514eb10f1e7586da]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220308123447] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:c562245b098ee175f52ef8d20fa37e2cd98bc5b7ffa6fe2a514eb10f1e7586da])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:c562245b098ee175f52ef8d20fa37e2cd98bc5b7ffa6fe2a514eb10f1e7586da
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:c562245b098ee175f52ef8d20fa37e2cd98bc5b7ffa6fe2a514eb10f1e7586da
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:c562245b098ee175f52ef8d20fa37e2cd98bc5b7ffa6fe2a514eb10f1e7586da].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 35m 58s
109 actionable tasks: 72 executed, 33 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/sonkgkm5xckxq

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #77

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/77/display/redirect?page=changes>

Changes:

[noreply] [BEAM-13925] Add ability to get metrics on pr-bot performance (#16985)


------------------------------------------
[...truncated 80.40 KB...]
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
Mar 07, 2022 12:37:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
SEVERE: 2022-03-07T12:37:41.364Z: generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
Mar 07, 2022 12:37:45 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
SEVERE: 2022-03-07T12:37:44.730Z: generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
Mar 07, 2022 4:00:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-07T16:00:42.300Z: Cancel request is committed for workflow job: 2022-03-07_04_32_35-1741415835065682294.
Mar 07, 2022 4:00:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-07T16:00:42.392Z: Cleaning up.
Mar 07, 2022 4:00:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-07T16:00:42.474Z: Stopping **** pool...
Mar 07, 2022 4:00:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-07T16:00:42.528Z: Stopping **** pool...
Mar 07, 2022 4:03:14 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-07T16:03:13.767Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Mar 07, 2022 4:03:14 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-07T16:03:13.815Z: Worker pool stopped.
Mar 07, 2022 4:03:20 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-03-07_04_32_35-1741415835065682294 finished with status CANCELLED.
Load test results for test (ID): 7edca3b4-634f-46d6-9e45-bf2e0241ce8e and timestamp: 2022-03-07T12:32:30.154000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                 12346.571
dataflow_v2_java17_total_bytes_count             3.00751037E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220307123036
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:2ea2121aa9be3ea55b9a124d683324c5cd577729dc50c7e67278536f571c46df
Deleted: sha256:3191ac007ae400def8a9bff7b0125fac17f369248a00d641626adfa51fcb570f
Deleted: sha256:4e2515d07d5a8e2b161f591d41d2fb873fb70b68814374638e147dd67d4401c8
Deleted: sha256:42f5c57242c4043f6c23ddb6fff9c1dbfdd79de736770aab61b55e13026fa854
Deleted: sha256:c82648a956b91775fabddbd4866b14c23b382b332020e4d6d8d9afc24100c97c
Deleted: sha256:6f9eed64dfa0dbce61f7c04532c59eeb894d34cb2ac288c90094f4679c8d9180
Deleted: sha256:94634b76d2cff6a2c6a63fa2ae1d572d5a6b33058946893582ba642612d65223
Deleted: sha256:051630fa3b8b7bdbaf80acc1dd62bf2ef172a47bf05ab8f5abcf51e9e0d261c1
Deleted: sha256:7b734bf6062f7d61f14dafc31710cbc819a44a1f53042a963edf997501512020
Deleted: sha256:46bea90fe529c441373f26594885a85650254414bf74a0ff2ee3557d04dfed7e
Deleted: sha256:6ddfbe1fdecab56db3c7dcdbfc4847007a75fea8c5df2b8c84db0352675142e5
Deleted: sha256:a6fac0aabe61754a6cd765ecf1cb401d37d83f6ff92a9e5c73342d923106dd59
Deleted: sha256:1e1290b777f532eef67eb0df1f86f9d1b9c7fffb3d6e8300cba92cb2c37de728
Deleted: sha256:eb52e793a2c98612e077424bcf02a832704b9a7881addd132ac530062eb2a86d
Deleted: sha256:1639921366c1a81d870536ef47950b102ec5ce4821604ec5fb2147394d9069c3
Deleted: sha256:4eca5426906287df5511e165b1a45c06193175eed32332b0a08f7088d229fadc
Deleted: sha256:090c90c7e61f60ed14150728e4badc42c87c7701da62e97f9eb210b197ca2745
Deleted: sha256:9c056d15e9789d0dffd7fe0e48da59257cda989834af7297cf135a8d9249fae1
Deleted: sha256:6fdf072c464124f2b9158cea3d0685a10cb7f6ccfa45d10b7763677e1fc08b40
Deleted: sha256:1f2baf1aa39ec7a006d4fc19b6a29b216fd91eed14cae94139e855de8a5944bc
Deleted: sha256:91d42a5f8b47a61090af171e4888aa9a75fbc210b79b74909b432c444c1fad3c
Deleted: sha256:8a4cbdc719bd82b0d1424842a997aa15c7e77f80ce06836da9681d8a944cec4c
Deleted: sha256:adf3a2ccbc278048fd33976962b7b7f19c6870bcc78f6bce3647212690ddd7db
Deleted: sha256:5389e927ec62379c3ccf18d532e43299bc17bc7d29e9bd2d96ba9569facef722
Deleted: sha256:0cc2615084baa78f34300ae6cfbc3bf8466fe7c3de816fa879ab2891f8a58070
Deleted: sha256:612f7e1d68b7baec8ed4fae6caab310c96c1a9e711966029e3ddc9981e8ecbb5
Deleted: sha256:42a3088ed62623b68a6b62672a7312f2a29eb646f2aefa5f1cd36c124bb26bfc
Deleted: sha256:546e0a6b1c2f4e49c564214efdd73d9363f78bf330c5cdb118e6842ec62ba632
Deleted: sha256:c4619e4b0e39ad66c79090a4f8dc77fd22772ae8a890c7c80b097e8c80606fea
Deleted: sha256:9911dba543e72380cd57a49fd5a13c2de89f70e741a12f0ace124bdd34370e1f
Deleted: sha256:45a93140ab8e9796a5f3859e7dd3aaabe2ff169d0f22d6c961d0ab77f5d88109
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220307123036]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:2ea2121aa9be3ea55b9a124d683324c5cd577729dc50c7e67278536f571c46df]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220307123036] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:2ea2121aa9be3ea55b9a124d683324c5cd577729dc50c7e67278536f571c46df])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:2ea2121aa9be3ea55b9a124d683324c5cd577729dc50c7e67278536f571c46df
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:2ea2121aa9be3ea55b9a124d683324c5cd577729dc50c7e67278536f571c46df
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:2ea2121aa9be3ea55b9a124d683324c5cd577729dc50c7e67278536f571c46df].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 33m 4s
109 actionable tasks: 72 executed, 33 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/lpytalftxnyey

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 - Build # 76 - Aborted!

Posted by Apache Jenkins Server <je...@builds.apache.org>.
beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 - Build # 76 - Aborted:

Check console output at https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/76/ to view the results.

Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #75

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/75/display/redirect?page=changes>

Changes:

[rahuliyer573] py: Import beam plugins before starting SdkHarness

[stephen.patel] BEAM-14011 fix s3 filesystem multipart copy

[Valentyn Tymofieiev] Bump numpy bound to include 1.22 and regenerate container deps.

[github-actions] [BEAM-13925] months in date constructor are 0 indexed

[noreply] Merge pull request #16842 from [BEAM-13932][Playground] Container's user

[noreply] Doc updates and blog post for 2.37.0 (#16887)

[noreply] Remove resolved issue in docs + update class path on sample (#17018)

[noreply] [BEAM-14016] Fixed flaky postcommit test (#17009)

[noreply] Remove resolved issue in notebook

[noreply] [BEAM-13947] Add split() and rsplit(), non-deferred column operations on

[noreply] BEAM-14026 - Fixes bug related to Unnesting nested rows in an array


------------------------------------------
[...truncated 932.24 KB...]
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
Mar 05, 2022 4:03:25 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-05T16:03:23.842Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Mar 05, 2022 4:03:25 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-05T16:03:23.902Z: Worker pool stopped.
Mar 05, 2022 4:03:30 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-03-05_04_32_42-774827235346324432 finished with status CANCELLED.
Load test results for test (ID): faafe6e8-722c-4c48-be5a-33c3a1c4f806 and timestamp: 2022-03-05T12:32:37.558000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                 12350.024
dataflow_v2_java17_total_bytes_count             3.67828715E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220305123041
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:bc15099965a6bde5329dde01fde5be7e0a676e96474f44f9c7ac11530be9ac6c
Deleted: sha256:489af3c4c488f8d7161f7ed73bddf90c910bb996ad62f2d2f689f4c4e2834f35
Deleted: sha256:185f53adaf6f5f86f5f490282eaece40a6cc295c297179caccf089f9d00861fa
Deleted: sha256:40c6d1631e3f21981e9c93cbb5595b32ab7b3f487a52dd230cb73c6c5d6b12ea
Deleted: sha256:4fa859ed747c1a3e57c54612fc4119d55bc247f9471bdc8826ebecebc1a06196
Deleted: sha256:587f28971c6bfe5fb1ead6b4bf629508f1588ace27773f34e60cde0298a8e732
Deleted: sha256:16c126b118fecdd19b061e68cf69ceea0be82d39792af4a8332e3a153182dec9
Deleted: sha256:7fc61ba63a07a90687a72ae39f5a9630c73c22ce0128cba7b545ef612b331ed7
Deleted: sha256:14696a23a947e7d814094dddccebbf8d761945cf257d2e814a6736429df970e2
Deleted: sha256:1d7da97a9f2c15bb6013f1b5091e720808b7ca822a811b4900fd4abc612864dd
Deleted: sha256:3b23abe8debfd8f04a4d51cb29891e46350addeece0d1ed25cd587cb658331b8
Deleted: sha256:a19b6d67110f1c84c64940f70141714f129ab04b3cb28ced462f337bcb4674af
Deleted: sha256:82efce9803e3f9ca2de002359e89e1002875b5feadba22d8a5bc552890cd114c
Deleted: sha256:eabb9f4fa81f144ef2292bb955c7b99a0a0609f39aa88c4433e370ac80542dd4
Deleted: sha256:9ea3b67b6810aa80e9ceb0c6f62a8c2bc06b1c1dabac84c8780c73b491c59e48
Deleted: sha256:acf3e2066aa64e35ceb5ff35300765a6d0429a685ef41a25c42f69296ba4fe23
Deleted: sha256:0871ec7925a4d080eacf0540036bcba3eeaad2fba834c13c6306a57ff93a1c74
Deleted: sha256:307a4d39d880e851c5ccc83bf1ff0007a6d10192284c7677e6168a7641884f8e
Deleted: sha256:3a8ea44c2e724195c18ecb483121f946ed6ddeaccb770aaf929286bfcb7d51bb
Deleted: sha256:c67786088dd42ed14b6661a2630ab3852b6dffc730c56dfdfa60f8e4d96d05ab
Deleted: sha256:27d5be10d56180ce94822b1b14728a159965ae4af5ed91f8d64eb759cc48c806
Deleted: sha256:2439a7539b493adc3ec771de6449a947030b8141d9fb3504ebef7751408e9464
Deleted: sha256:70ebcfe87fd4076a39c8c5249a9059b3ee59e12c823287cd53af43d03e36c4e0
Deleted: sha256:6297924ea911d496f223c36582e6d84abbaf4bc633db6acc3ee1f9956c91780c
Deleted: sha256:8e93c5aaa8ae37fe38bd0f3d9b3afb26f824f7b7ea0d2be568434449e1cb7ab9
Deleted: sha256:e50382346fc5dd4b5f52ece7c09ac8b4c2e3dfc3eb8976e5b67eb0bb025c5db9
Deleted: sha256:0513dfb6338d6ca0b017f7c4efa776afb14c48ad1601434326796c80ae47693d
Deleted: sha256:b4b9a0b1419f49663ed7e8bf9f30ba5cad0b2580ceb330fbbabed4cdd9022574
Deleted: sha256:d3a24da1fec874f2ed40b37ede58bfcbc616744ab315d9d3fad293e7d995e58b
Deleted: sha256:011d6e13098f2f80c106bd98796843690ce936e037be092114d0034fb274956c
Deleted: sha256:469e1fff2e2cfe716009361c1e17b7b2bf770ca75a642e9b90cd08e2ee8cb117
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220305123041]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:bc15099965a6bde5329dde01fde5be7e0a676e96474f44f9c7ac11530be9ac6c]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220305123041] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:bc15099965a6bde5329dde01fde5be7e0a676e96474f44f9c7ac11530be9ac6c])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:bc15099965a6bde5329dde01fde5be7e0a676e96474f44f9c7ac11530be9ac6c
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:bc15099965a6bde5329dde01fde5be7e0a676e96474f44f9c7ac11530be9ac6c
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:bc15099965a6bde5329dde01fde5be7e0a676e96474f44f9c7ac11530be9ac6c].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 33m 11s
109 actionable tasks: 72 executed, 33 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/4ib4fyyab3thw

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #74

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/74/display/redirect?page=changes>

Changes:

[stranniknm] [BEAM-13999] playground - support vertical orientation for graph

[noreply] [adhoc] Prepare aws2 ClientConfiguration for json serialization and

[noreply] Merge pull request #16879 from [BEAM-12164] Add javadocs to

[noreply] [Cleanup] Update pre-v2 go package references (#17002)

[noreply] [BEAM-13885] Add unit tests to window package (#16971)

[noreply] Merge pull request #16891 from [BEAM-13872] [Playground] Increase test

[noreply] Merge pull request #16912 from [BEAM-13878] [Playground] Increase test

[noreply] Merge pull request #16946 from [BEAM-13873] [Playground] Increase test

[noreply] [BEAM-13951] Update mass_comment.py list of Run commands (#16889)

[noreply] [BEAM-10652] Allow Clustering without Partition in BigQuery (#16578)

[noreply] [BEAM-13857] Add K:V flags for expansion service jars and addresses to


------------------------------------------
[...truncated 290.62 KB...]
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
Mar 04, 2022 4:00:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-04T16:00:41.017Z: Cancel request is committed for workflow job: 2022-03-04_04_36_28-11688097497522893748.
Mar 04, 2022 4:00:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-04T16:00:41.049Z: Cleaning up.
Mar 04, 2022 4:00:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-04T16:00:41.108Z: Stopping **** pool...
Mar 04, 2022 4:00:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-04T16:00:41.152Z: Stopping **** pool...
Mar 04, 2022 4:03:14 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-04T16:03:13.330Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Mar 04, 2022 4:03:14 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-04T16:03:13.361Z: Worker pool stopped.
Mar 04, 2022 4:03:20 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-03-04_04_36_28-11688097497522893748 finished with status CANCELLED.
Load test results for test (ID): 19de2143-150c-4ca7-9730-91a6f344b079 and timestamp: 2022-03-04T12:36:23.352000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                 12092.635
dataflow_v2_java17_total_bytes_count             2.90951265E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220304123414
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:54e699f23dd25da2a5e5a7617f4a3e821090856c90221ed048d5760d3c05c294
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220304123414]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:54e699f23dd25da2a5e5a7617f4a3e821090856c90221ed048d5760d3c05c294]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220304123414] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:54e699f23dd25da2a5e5a7617f4a3e821090856c90221ed048d5760d3c05c294])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:54e699f23dd25da2a5e5a7617f4a3e821090856c90221ed048d5760d3c05c294
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:54e699f23dd25da2a5e5a7617f4a3e821090856c90221ed048d5760d3c05c294
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:54e699f23dd25da2a5e5a7617f4a3e821090856c90221ed048d5760d3c05c294].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 29m 26s
109 actionable tasks: 72 executed, 33 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/xcdwt2viqnh7k

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #73

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/73/display/redirect?page=changes>

Changes:

[Alexey Romanenko] Bump org.mongodb:mongo-java-driver to 3.12.10

[noreply] [BEAM-13973] Link Dataproc Flink master URLs to the InteractiveRunner

[noreply] [BEAM-13925] Turn pr bot on for go prs (#16984)

[Pablo Estrada] Skipping flaky sad-path tests for Spanner changestreams

[noreply] [BEAM-13964] Bump kotlin to 1.6.x (#16882)

[noreply] Merge pull request #16906: [BEAM-13974] Handle idle Storage Api streams

[noreply] Merge pull request #16562 from [BEAM-13051][D] Enable pylint warnings

[noreply] [BEAM-13925] A couple small pr-bot bug fixes (#16996)

[noreply] [BEAM-14029] Add getter, setter for target maven repo (#16995)

[noreply] [BEAM-13903] Improve coverage of metricsx package (#16994)

[noreply] [BEAM-13892] Improve coverage of avroio package (#16990)


------------------------------------------
[...truncated 1.05 MB...]
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: Th
Mar 03, 2022 4:00:48 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-03T16:00:47.957Z: Cancel request is committed for workflow job: 2022-03-03_04_40_15-18236648418207913332.
Mar 03, 2022 4:00:48 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-03T16:00:48.039Z: Cleaning up.
Mar 03, 2022 4:00:48 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-03T16:00:48.119Z: Stopping **** pool...
Mar 03, 2022 4:00:49 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-03T16:00:48.185Z: Stopping **** pool...
Mar 03, 2022 4:03:10 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-03T16:03:09.892Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Mar 03, 2022 4:03:10 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-03T16:03:09.966Z: Worker pool stopped.
Mar 03, 2022 4:03:20 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-03-03_04_40_15-18236648418207913332 finished with status CANCELLED.
Load test results for test (ID): d06f5044-823f-4853-b8e0-e5020c8f9fd2 and timestamp: 2022-03-03T12:40:10.193000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                 11884.404
dataflow_v2_java17_total_bytes_count             3.01970633E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220303123806
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:f0521903fd7629c26c43e3c67ea5ca699f2c22836341b04da890bd77aff3367f
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220303123806]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:f0521903fd7629c26c43e3c67ea5ca699f2c22836341b04da890bd77aff3367f]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220303123806] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:f0521903fd7629c26c43e3c67ea5ca699f2c22836341b04da890bd77aff3367f])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:f0521903fd7629c26c43e3c67ea5ca699f2c22836341b04da890bd77aff3367f
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:f0521903fd7629c26c43e3c67ea5ca699f2c22836341b04da890bd77aff3367f
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:f0521903fd7629c26c43e3c67ea5ca699f2c22836341b04da890bd77aff3367f].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 25m 33s
109 actionable tasks: 72 executed, 33 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/gzqqqa665z73q

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #72

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/72/display/redirect?page=changes>

Changes:

[egalpin] Use default context output rather than outputWithTimestamp for

[stranniknm] Palo Alto case study - fix link

[rogelio.hernandez] [BEAM-12777] Removed current docs version redirect

[noreply] Merge pull request #16850: [BEAM-11205] Upgrade Libraries BOM

[noreply] Merge pull request #16484 from [BEAM-13633] [Playground] Implement

[noreply] Add 2022 events blog post (#16975)

[noreply] Clean up Go formatter suggestions (#16973)

[noreply] [BEAM-14012] Add go fmt to Github Actions (#16978)

[noreply] [BEAM-13911] Add basic tests to Go direct runner. (#16979)

[noreply] [BEAM-13960] Add support for more types when converting from between row


------------------------------------------
[...truncated 774.19 KB...]
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmi
Mar 02, 2022 4:03:28 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-02T16:03:28.556Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Mar 02, 2022 4:03:28 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-02T16:03:28.601Z: Worker pool stopped.
Mar 02, 2022 4:03:36 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-03-02_04_39_58-14157234965447229207 finished with status CANCELLED.
Load test results for test (ID): c8be7bac-0c31-448b-82b7-fbf8e213e04c and timestamp: 2022-03-02T12:39:52.927000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                 11907.599
dataflow_v2_java17_total_bytes_count              3.7305674E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220302123746
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:696b0c18a8c3e5416fa9de4e456e9fa73e601dd1063b73978a2e1733d5ef82b0
Deleted: sha256:bd2231812ea39dbbe935087819774756a0f0458904ce6d4d08eef3d81f0ba9c9
Deleted: sha256:a17918d6f3b5b4fd0e8616b007a01f946d86a5ccaf1fb0b4c61c2e8837a02991
Deleted: sha256:399e0cadd4dcae5e076613b56258bb0df1ba1580abef3c2014a0b6910b38f074
Deleted: sha256:1c7bde8db26720434da223e0644b9fc4c305acbbfb8d985d36c299ab4ca70e40
Deleted: sha256:17a44b4d11d8e7ec3327d872f69d53198533e10e0bb8833028a8591af58f00b5
Deleted: sha256:5452ffac731a4f3ad6e18cc591d04209525241c8884f44bc7185d603c241b90b
Deleted: sha256:f5fe5642e3246097e84d197e1928ba94eeb2b2a89d1f24119714e2ef8a9359b2
Deleted: sha256:4fa7afe238b3dc6b2cb03f3754fcd88219704691476388696523f77496d0a0d7
Deleted: sha256:03ff7099e9b07eac2f3db7bbf5c71d0e57483cb57fb4ebd8e5b034ff9555c150
Deleted: sha256:c87401a423863d07b1c76625adfeb6fc8d82fc4b1fb92a601e1c09c91cdc5112
Deleted: sha256:e57610f2be566d722b09eb290eb4ad7ba63e9f31185620348bd5caf56756d315
Deleted: sha256:fa36a3ab397a54511783bcca1b167f983311a13b6a8669e7931dc18cead17a06
Deleted: sha256:5a97d1daa974873c1fdc282a31af60040ce7e24050f9899ae76363992279397b
Deleted: sha256:2014d7bc03daafa79d51030200158b429b78f1b62efbc0d176bbaec919ba44e5
Deleted: sha256:c6b34fe5b40a7c7e7f22b85b84ae57f5f6f1376895d5d597e0258e7129b9e4dd
Deleted: sha256:bd6a5d41952b5320f610540883ad5ef55a82da21c41176a1ccd9088a7765868a
Deleted: sha256:0fb6d2176ff041ec9015b5dcab97cf92ed1662aa47ad7725d2e09e4f2dd990b1
Deleted: sha256:75d5a482125f68c274774995dd1d5c69fb7bc9fa593488bfae230681bb9ccbfd
Deleted: sha256:0dec6be1e8f549d1065d72c83d574568df55453f13ca634f041f252a5b5f0894
Deleted: sha256:3f065c71c23f117ec772509eb068fa8c9bf0a468c3a5d1ba40862c959fbd6c1c
Deleted: sha256:33975ac00f0a543e262526909718fd914062e7853b666e559ebc4eecff51941a
Deleted: sha256:1f6a11373d6b7a749bd96f5393c62d60bfd63b1bd79aea17f276c2675b00555e
Deleted: sha256:ff7a3f7ea865088ef753b738d4b8211bce14634bf50f7e268394d63ec9232641
Deleted: sha256:c9a0e64b5c83d4f4b0132b52a9ef943d01e7161841f9f438214d7b4b4b14f1aa
Deleted: sha256:15e86cedf0ba7b31058553b009ab65b2568f51ae2c9edb488dbb8f6e1493cd8f
Deleted: sha256:aefd58cb364b251a31a432ede9a6c54639527756f74b6e24e59c2449dc626a66
Deleted: sha256:16f8d12b20dbaa5a383a339f5d7fbda569794ec54300acd00afb556750471fd2
Deleted: sha256:401bef75155e37ef0679e3480b71fe21cb0e73326042b8424cd1dfa40dd16ac1
Deleted: sha256:1de44c5fe8807a52445950d5431c2cfb2f6b89a71d843d03a67d0c8f148f4a7c
Deleted: sha256:090db9e705564073beb65a7f5ed2a6135e680c1d0580fa8529403b5128153edd
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220302123746]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:696b0c18a8c3e5416fa9de4e456e9fa73e601dd1063b73978a2e1733d5ef82b0]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220302123746] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:696b0c18a8c3e5416fa9de4e456e9fa73e601dd1063b73978a2e1733d5ef82b0])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:696b0c18a8c3e5416fa9de4e456e9fa73e601dd1063b73978a2e1733d5ef82b0
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:696b0c18a8c3e5416fa9de4e456e9fa73e601dd1063b73978a2e1733d5ef82b0
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:696b0c18a8c3e5416fa9de4e456e9fa73e601dd1063b73978a2e1733d5ef82b0].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 26m 13s
109 actionable tasks: 72 executed, 33 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/b7fwb6lbw4ipw

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #71

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/71/display/redirect?page=changes>

Changes:

[noreply] Build wheels for Python 3.9

[noreply] Merge pull request #16892 from [BEAM-13755] [Playground] Scroll the

[noreply] Merge pull request #16880 from [BEAM-13963][Playground] Get bucket name

[noreply] Merge pull request #16870 from [BEAM-13874][Playground] Tag multifile

[noreply] Merge pull request #16910 from [BEAM-13724] [Playground] Get the default

[noreply] [BEAM-14008] Fix incorrect guava import (#16966)

[noreply] Fix ignored exception in BatchSpannerRead. (#16960)

[noreply] [BEAM-13917] Improve coverage of databaseio package (#16956)

[noreply] [BEAM-13925] Add entry files to process new prs and pr updates for PR

[noreply] [BEAM-13899] Improve coverage of debug package (#16951)

[noreply] [BEAM-13907] Improve coverage of textio package (#16937)

[noreply] [BEAM-9150] Fix beam_PostRelease_Python_Candidate (python RC validation


------------------------------------------
[...truncated 653.07 KB...]
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    di
Mar 01, 2022 4:00:47 PM org.apache.beam.sdk.metrics.MetricsEnvironment getCurrentContainer
WARNING: Reporting metrics are not supported in the current execution environment.
Mar 01, 2022 4:00:47 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-01T16:00:43.740Z: Cancel request is committed for workflow job: 2022-03-01_04_38_30-13844224834732571417.
Mar 01, 2022 4:00:47 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-01T16:00:43.779Z: Cleaning up.
Mar 01, 2022 4:00:47 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-01T16:00:43.871Z: Stopping **** pool...
Mar 01, 2022 4:00:47 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-01T16:00:43.915Z: Stopping **** pool...
Mar 01, 2022 4:03:16 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-01T16:03:15.500Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Mar 01, 2022 4:03:16 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-03-01T16:03:15.551Z: Worker pool stopped.
Mar 01, 2022 4:03:24 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-03-01_04_38_30-13844224834732571417 finished with status CANCELLED.
Load test results for test (ID): 21c09062-7d49-4893-9a3f-fff6df2a07d0 and timestamp: 2022-03-01T12:38:25.128000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                 12006.945
dataflow_v2_java17_total_bytes_count             3.14987614E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220301123604
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:7601b0b9fe14611d3e0cccc40c0e257d8552e91e45f79abe1ee866bd3f788cd6
Deleted: sha256:7c5d44c8736dc53aee22edf6ecb51e28731ab363c6f80f81e84399ec5f28800e
Deleted: sha256:acfbd54b31857ba1a382a0c7b6a03e472cf4872cff1a3501f1a3cbb94ea52cf2
Deleted: sha256:ddc08bd996f3ce82ee09b8f860d608cb957f53b9ae5e5424a0e4887875cfcc7f
Deleted: sha256:a964cb004ea3e12fed2c7ff8a692a4ba3e8f5c6bfef8eecc81af7abcdcd177f2
Deleted: sha256:bb293d81525a6cd3551b1d3b4ef8d4886d77a8160ae102561788c8584ca735b9
Deleted: sha256:216e122f3f133c5cb55a828008895f3c5ee67ce8920830a3f857fa4dcbc9b0f6
Deleted: sha256:bc06d46dcc7d82479b4f44d7a1c24f6706a40873c6a255f28ca2a5620eef01fe
Deleted: sha256:1abf31306df482035062446375ae9a5b4cf6502233da2c6a9a660894e9b2ec17
Deleted: sha256:0cc73153d897110f23a08fa9ba428f5e59042dd40034198125ba431807071973
Deleted: sha256:0b3e5688133e222d134411f9e52cf6809062823f78aa40d824100146897c0ae4
Deleted: sha256:adb03f09c560f729f3cc5a4f520580955e1cb5d2c4a096111a342ae572ab92cc
Deleted: sha256:85b861fd42dc6b205142cef4d90e10c300150efdecf37b2bda5d6b1caceedb82
Deleted: sha256:6d518b6cbd1ffb3adf8710aa32717f4ad72dbc67caf57ba6eb27f786aa1ea3ee
Deleted: sha256:9adef3877bc4fbb34024f95344d5b83cfc494537ea0ac5b101c62d5d14f41988
Deleted: sha256:b3815c9f077612fdbd2227f144c42d6a453efdcf8a0ecea3608714bc1c5ab261
Deleted: sha256:cc8e7382e9d1d9047e80ac2bbbda0ac0441d641ad7c338a9e225fcb0eaa1ca31
Deleted: sha256:c37c2bd9a1a9922c399941ce94de432e44d706eafa2bf7060dbae1a9cfbc943f
Deleted: sha256:96e1fad14081fb2eb5045f9d5f0dd13534921d27b1d3dddc1e4c8071fca545ea
Deleted: sha256:1591b11c7a2aeb2c0c3de66e2f68c2ef693baba9824ecc8e96e4a0a0201c82cc
Deleted: sha256:6e572eacf6ea0beddba0d0824706de666c37784db670364d960734777e5b1b7e
Deleted: sha256:d2b96830acbcc3a42fa5c02f36d80b64d9ec63c6b8e2c6e4a050bc80a8e93a6c
Deleted: sha256:84b809ebd1ca0656b38a0275c86a9aa0c900d9e57ac329663e22a15094977299
Deleted: sha256:29506ed791245ff40bbb3410f2d5d217b8ec74e5360568743c8370825c9c4f5a
Deleted: sha256:f05b746271bb760bf8b2c5697f46cc8b7de6160c41da8c9f66bc682c1993f724
Deleted: sha256:8a2d1c161396583aa63e3806ef2c5250288c0f2bf07f00d9c72ba62e5f317405
Deleted: sha256:2b84efdaa7478c87f14f98eaefd16c037704f58420d54978a8f63ee8774eab3b
Deleted: sha256:e2eb13a6fc0a8bdb0bd55b1895c12e0cdec47b39dc41d18e567ca2f38cac4b0b
Deleted: sha256:84e0b1d8efdfc06bebee3af9babfd8650c882b09706962ed59ac8c8f02101f3a
Deleted: sha256:4bb28389c9bb56331b2d0e6cf885be131e2487a606ddb926ad2f414164589043
Deleted: sha256:a96ea783fd1c04bbaae4ac86d384c056a838081558d43301e86b001a63e4b1a5
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220301123604]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:7601b0b9fe14611d3e0cccc40c0e257d8552e91e45f79abe1ee866bd3f788cd6]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220301123604] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:7601b0b9fe14611d3e0cccc40c0e257d8552e91e45f79abe1ee866bd3f788cd6])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:7601b0b9fe14611d3e0cccc40c0e257d8552e91e45f79abe1ee866bd3f788cd6
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:7601b0b9fe14611d3e0cccc40c0e257d8552e91e45f79abe1ee866bd3f788cd6
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:7601b0b9fe14611d3e0cccc40c0e257d8552e91e45f79abe1ee866bd3f788cd6].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 27m 42s
109 actionable tasks: 72 executed, 33 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/zsgrrnvnx55ae

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #70

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/70/display/redirect>

Changes:


------------------------------------------
[...truncated 753.63 KB...]
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/windmill_cache_access.cc:40
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no
Feb 28, 2022 4:00:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-28T16:00:54.001Z: Cancel request is committed for workflow job: 2022-02-28_04_33_18-12693317070538660066.
Feb 28, 2022 4:00:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-28T16:00:54.055Z: Cleaning up.
Feb 28, 2022 4:00:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-28T16:00:54.127Z: Stopping **** pool...
Feb 28, 2022 4:00:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-28T16:00:54.172Z: Stopping **** pool...
Feb 28, 2022 4:03:17 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-28T16:03:17.662Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Feb 28, 2022 4:03:17 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-28T16:03:17.700Z: Worker pool stopped.
Feb 28, 2022 4:03:25 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-02-28_04_33_18-12693317070538660066 finished with status CANCELLED.
Load test results for test (ID): b4a831c3-af77-4cb3-90cf-b1f44a475058 and timestamp: 2022-02-28T12:33:13.439000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                  12301.01
dataflow_v2_java17_total_bytes_count             2.51162037E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220228123125
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:f11e4cc40d92ec9665ac6a6c5830e9bd0e7a90f749156f3e76d46bc1ba9a2759
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220228123125]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:f11e4cc40d92ec9665ac6a6c5830e9bd0e7a90f749156f3e76d46bc1ba9a2759]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220228123125] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:f11e4cc40d92ec9665ac6a6c5830e9bd0e7a90f749156f3e76d46bc1ba9a2759])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:f11e4cc40d92ec9665ac6a6c5830e9bd0e7a90f749156f3e76d46bc1ba9a2759
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:f11e4cc40d92ec9665ac6a6c5830e9bd0e7a90f749156f3e76d46bc1ba9a2759
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:f11e4cc40d92ec9665ac6a6c5830e9bd0e7a90f749156f3e76d46bc1ba9a2759].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 32m 19s
109 actionable tasks: 72 executed, 33 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/3juq4ae25px5u

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #69

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/69/display/redirect>

Changes:


------------------------------------------
[...truncated 474.73 KB...]
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmi
Feb 27, 2022 4:03:11 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-27T16:03:11.132Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Feb 27, 2022 4:03:11 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-27T16:03:11.200Z: Worker pool stopped.
Feb 27, 2022 4:03:17 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-02-27_04_32_54-4589397634042183329 finished with status CANCELLED.
Load test results for test (ID): 66ab286c-5d85-4640-99b0-5965f5444b7d and timestamp: 2022-02-27T12:32:46.805000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                 12335.482
dataflow_v2_java17_total_bytes_count             4.42431896E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220227123037
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:4b816538dee6070502b18495d0e57973802ec0e3d9046543d6fb1f42ba8fc398
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220227123037]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:4b816538dee6070502b18495d0e57973802ec0e3d9046543d6fb1f42ba8fc398]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220227123037] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:4b816538dee6070502b18495d0e57973802ec0e3d9046543d6fb1f42ba8fc398])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:4b816538dee6070502b18495d0e57973802ec0e3d9046543d6fb1f42ba8fc398
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:4b816538dee6070502b18495d0e57973802ec0e3d9046543d6fb1f42ba8fc398
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:4b816538dee6070502b18495d0e57973802ec0e3d9046543d6fb1f42ba8fc398].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 33m 3s
109 actionable tasks: 72 executed, 33 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/cd7x2rkumccrs

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #68

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/68/display/redirect?page=changes>

Changes:

[Ankur Goenka] [BEAM-13952] Sickbaying

[noreply] Memoize some objects for timer processing to reduce overhead. (#16207)

[noreply] [BEAM-13965] Use TypeDeserializer if type information is available to

[noreply] [BEAM-13912] Add more coverage for dataflow.go (#16903)

[noreply] [BEAM-12563] swaplevel general function for dataframe and series

[noreply] [BEAM-14001] Update coder.go unit tests (#16952)

[noreply] [BEAM-13910] Improve coverage of gcsx package (#16942)

[noreply] [BEAM-13015] Use a DirectExecutor for state since we are just completing


------------------------------------------
[...truncated 141.31 KB...]
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
Feb 26, 2022 4:00:50 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-26T16:00:49.154Z: Cancel request is committed for workflow job: 2022-02-26_04_42_10-4813651441616755827.
Feb 26, 2022 4:00:50 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-26T16:00:49.220Z: Cleaning up.
Feb 26, 2022 4:00:50 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-26T16:00:49.285Z: Stopping **** pool...
Feb 26, 2022 4:00:50 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-26T16:00:49.335Z: Stopping **** pool...
Feb 26, 2022 4:03:11 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-26T16:03:10.586Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Feb 26, 2022 4:03:11 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-26T16:03:10.627Z: Worker pool stopped.
Feb 26, 2022 4:03:20 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-02-26_04_42_10-4813651441616755827 finished with status CANCELLED.
Load test results for test (ID): 7cd11dfd-cd29-4930-aa38-2c247c76f56e and timestamp: 2022-02-26T12:42:03.570000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                 11770.953
dataflow_v2_java17_total_bytes_count             4.05083241E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220226123950
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:c819e3ffd7d5e7ef0f7b8f1b3eba24a3217c96e85d3b4df3a989d7adf9695d5f
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220226123950]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:c819e3ffd7d5e7ef0f7b8f1b3eba24a3217c96e85d3b4df3a989d7adf9695d5f]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220226123950] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:c819e3ffd7d5e7ef0f7b8f1b3eba24a3217c96e85d3b4df3a989d7adf9695d5f])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:c819e3ffd7d5e7ef0f7b8f1b3eba24a3217c96e85d3b4df3a989d7adf9695d5f
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:c819e3ffd7d5e7ef0f7b8f1b3eba24a3217c96e85d3b4df3a989d7adf9695d5f
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:c819e3ffd7d5e7ef0f7b8f1b3eba24a3217c96e85d3b4df3a989d7adf9695d5f].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 23m 49s
109 actionable tasks: 72 executed, 33 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/xjr536q56aw22

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #67

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/67/display/redirect?page=changes>

Changes:

[benjamin.gonzalez] Revert PR#16253 due errors with plugin flaky-test-handler

[noreply] Fix BoundedQueueExecutor and StreamingDataflowWorker to actually limit

[noreply] [BEAM-1857] Add Neo4jIO (#15916)

[noreply] [BEAM-13767] Migrate serveral portable runner tasks to use configuration

[noreply] [BEAM-13996] Removing 'No cluster_manager is associated with the

[noreply] [BEAM-13906] Improve coverage of errors package (#16934)

[noreply] [BEAM-13886] unit tests for trigger package (#16935)

[noreply] [BEAM-4767] Remove beam- prefix from release script tags (#16899)

[noreply] [BEAM-13866] Add small unit tests to errorx, make boolean assignment

[noreply] [BEAM-13925] Add most of the supporting files for the pr management

[noreply] Merge pull request #16846 from [BEAM-12164]: Add sad path tests for


------------------------------------------
[...truncated 656.32 KB...]
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmi
Feb 25, 2022 3:54:12 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
SEVERE: 2022-02-25T15:54:09.994Z: Staged package animal-sniffer-annotations-1.20-vt1E38otwrj1wIzR1vDgznQJTsZ3gSYJaOA_wOd1Iqw.jar at location 'gs://temp-storage-for-perf-tests/loadtests/staging/animal-sniffer-annotations-1.20-vt1E38otwrj1wIzR1vDgznQJTsZ3gSYJaOA_wOd1Iqw.jar' is inaccessible.
Feb 25, 2022 3:54:12 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
SEVERE: 2022-02-25T15:54:10.554Z: Staged package beam-runners-core-java-2.38.0-SNAPSHOT-g1gf2aIDSblOWLSzcUcynwLjKg419MgwyCpDR4jXU8Q.jar at location 'gs://temp-storage-for-perf-tests/loadtests/staging/beam-runners-core-java-2.38.0-SNAPSHOT-g1gf2aIDSblOWLSzcUcynwLjKg419MgwyCpDR4jXU8Q.jar' is inaccessible.
Feb 25, 2022 3:54:13 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
SEVERE: 2022-02-25T15:54:13.600Z: Staged package opencensus-contrib-http-util-0.28.0-ScPbKinx_bL3OSjL6pab0dQPq3zFu2JzAiur2W96eJs.jar at location 'gs://temp-storage-for-perf-tests/loadtests/staging/opencensus-contrib-http-util-0.28.0-ScPbKinx_bL3OSjL6pab0dQPq3zFu2JzAiur2W96eJs.jar' is inaccessible.
Feb 25, 2022 3:54:15 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
WARNING: 2022-02-25T15:54:14.221Z: One or more access checks for temp location or staged files failed. Please refer to other error messages for details. For more information on security and permissions, please see https://cloud.google.com/dataflow/security-and-permissions.
Feb 25, 2022 3:57:16 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
WARNING: 2022-02-25T15:57:15.145Z: One or more access checks for temp location or staged files failed. Please refer to other error messages for details. For more information on security and permissions, please see https://cloud.google.com/dataflow/security-and-permissions.
Feb 25, 2022 4:00:11 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
SEVERE: 2022-02-25T16:00:10.019Z: Staged package animal-sniffer-annotations-1.20-vt1E38otwrj1wIzR1vDgznQJTsZ3gSYJaOA_wOd1Iqw.jar at location 'gs://temp-storage-for-perf-tests/loadtests/staging/animal-sniffer-annotations-1.20-vt1E38otwrj1wIzR1vDgznQJTsZ3gSYJaOA_wOd1Iqw.jar' is inaccessible.
Feb 25, 2022 4:00:11 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
SEVERE: 2022-02-25T16:00:10.564Z: Staged package beam-runners-core-java-2.38.0-SNAPSHOT-g1gf2aIDSblOWLSzcUcynwLjKg419MgwyCpDR4jXU8Q.jar at location 'gs://temp-storage-for-perf-tests/loadtests/staging/beam-runners-core-java-2.38.0-SNAPSHOT-g1gf2aIDSblOWLSzcUcynwLjKg419MgwyCpDR4jXU8Q.jar' is inaccessible.
Feb 25, 2022 4:00:18 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
SEVERE: 2022-02-25T16:00:16.877Z: Staged package opencensus-contrib-http-util-0.28.0-ScPbKinx_bL3OSjL6pab0dQPq3zFu2JzAiur2W96eJs.jar at location 'gs://temp-storage-for-perf-tests/loadtests/staging/opencensus-contrib-http-util-0.28.0-ScPbKinx_bL3OSjL6pab0dQPq3zFu2JzAiur2W96eJs.jar' is inaccessible.
Feb 25, 2022 4:00:18 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
WARNING: 2022-02-25T16:00:17.999Z: One or more access checks for temp location or staged files failed. Please refer to other error messages for details. For more information on security and permissions, please see https://cloud.google.com/dataflow/security-and-permissions.
Feb 25, 2022 4:01:05 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-25T16:01:05.181Z: Cancel request is committed for workflow job: 2022-02-25_04_32_45-12207937927914771831.
Feb 25, 2022 4:01:05 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-25T16:01:05.241Z: Cleaning up.
Feb 25, 2022 4:01:06 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-25T16:01:05.549Z: Stopping **** pool...
Feb 25, 2022 4:01:06 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-25T16:01:05.619Z: Stopping **** pool...
Feb 25, 2022 4:03:30 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-25T16:03:29.965Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Feb 25, 2022 4:03:30 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-25T16:03:30.007Z: Worker pool stopped.
Feb 25, 2022 4:03:36 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-02-25_04_32_45-12207937927914771831 finished with status CANCELLED.
Load test results for test (ID): 52b2b53f-6530-4526-825e-051975721ddc and timestamp: 2022-02-25T12:32:40.120000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                  12218.91
dataflow_v2_java17_total_bytes_count             3.53406182E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220225123038
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:0c16c549805315a1fd45a981f9490cf1860d4a2925cda18fa2c1c091105773e0
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220225123038]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:0c16c549805315a1fd45a981f9490cf1860d4a2925cda18fa2c1c091105773e0]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220225123038] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:0c16c549805315a1fd45a981f9490cf1860d4a2925cda18fa2c1c091105773e0])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:0c16c549805315a1fd45a981f9490cf1860d4a2925cda18fa2c1c091105773e0
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:0c16c549805315a1fd45a981f9490cf1860d4a2925cda18fa2c1c091105773e0
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:0c16c549805315a1fd45a981f9490cf1860d4a2925cda18fa2c1c091105773e0].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 33m 19s
109 actionable tasks: 72 executed, 33 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/6oxj7mvyjcxjm

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #66

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/66/display/redirect?page=changes>

Changes:

[noreply] [BEAM-12645] Fix code-cov flakes due to monorepo. (#16925)

[noreply] [BEAM-13969] Deprecate stringx package (#16884)

[noreply] Add Go badge to ReadMe (#16897)

[noreply] [BEAM-13980] Re-add method gone missing in af2f8ee6 (#16918)

[noreply] [BEAM-13884] Improve mtime package (#16924)

[noreply] Minor: Update Go API doc links (#16932)

[noreply] [BEAM-13218] Re-enable

[noreply] Merge pull request #16857 from [BEAM-13662] [Playground] Support

[noreply] Merge pull request #16826 from [BEAM-13870] [Playground] Increase test


------------------------------------------
[...truncated 654.18 KB...]
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmi
Feb 24, 2022 4:00:46 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-24T16:00:45.414Z: Cancel request is committed for workflow job: 2022-02-24_04_32_33-9861567060511479903.
Feb 24, 2022 4:00:46 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-24T16:00:45.460Z: Cleaning up.
Feb 24, 2022 4:00:46 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-24T16:00:45.543Z: Stopping **** pool...
Feb 24, 2022 4:00:46 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-24T16:00:45.582Z: Stopping **** pool...
Feb 24, 2022 4:03:05 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-24T16:03:04.960Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Feb 24, 2022 4:03:05 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-24T16:03:05.006Z: Worker pool stopped.
Feb 24, 2022 4:03:10 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-02-24_04_32_33-9861567060511479903 finished with status CANCELLED.
Load test results for test (ID): 4fedf486-70a1-4945-9983-5f63d7ada99b and timestamp: 2022-02-24T12:32:27.711000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                 12335.382
dataflow_v2_java17_total_bytes_count             3.09057116E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220224123032
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:b0db093dd54415828de5bc42a37c455a9ad54a28233c8eef5d19027f417d7bfc
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220224123032]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:b0db093dd54415828de5bc42a37c455a9ad54a28233c8eef5d19027f417d7bfc]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220224123032] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:b0db093dd54415828de5bc42a37c455a9ad54a28233c8eef5d19027f417d7bfc])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:b0db093dd54415828de5bc42a37c455a9ad54a28233c8eef5d19027f417d7bfc
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:b0db093dd54415828de5bc42a37c455a9ad54a28233c8eef5d19027f417d7bfc
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:b0db093dd54415828de5bc42a37c455a9ad54a28233c8eef5d19027f417d7bfc].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 32m 58s
109 actionable tasks: 72 executed, 33 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/hdbzyuvddvpeu

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #65

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/65/display/redirect?page=changes>

Changes:

[Kyle Weaver] [BEAM-13796] projection pushdown in BQ IO

[Kyle Weaver] [BEAM-13796] Move test to ReadTest class and correct javadoc for

[Kyle Weaver] [BEAM-13796] Pushdown is not supported on TypedRead#fromQuery.

[noreply] [BEAM-13738] Reenable ignored SQS test after bumping elasticmq for fixed

[noreply] fix build status link (#16907)

[noreply] Merge pull request #16549 from [BEAM-13681][Playground] Refactoring

[noreply] Merge pull request #16732 from [BEAM-13825] [Playground] updated

[noreply] Merge pull request #16683 from [BEAM-13713][Playground] Java graph

[noreply] case study pages - improvements and fixes (#16896)

[noreply] Palo Alto case study (#16915)


------------------------------------------
[...truncated 100.17 KB...]
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
Feb 23, 2022 12:53:34 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
SEVERE: 2022-02-23T12:53:32.542Z: generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
Feb 23, 2022 12:53:38 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
SEVERE: 2022-02-23T12:53:35.842Z: generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/windmill/client/streaming_rpc_client.cc:697
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:230
FATAL: command execution failed
java.io.IOException: Backing channel 'apache-beam-jenkins-11' is disconnected.
	at hudson.remoting.RemoteInvocationHandler.channelOrFail(RemoteInvocationHandler.java:216)
	at hudson.remoting.RemoteInvocationHandler.invoke(RemoteInvocationHandler.java:286)
	at com.sun.proxy.$Proxy126.isAlive(Unknown Source)
	at hudson.Launcher$RemoteLauncher$ProcImpl.isAlive(Launcher.java:1213)
	at hudson.Launcher$RemoteLauncher$ProcImpl.join(Launcher.java:1205)
	at hudson.Launcher$ProcStarter.join(Launcher.java:522)
	at hudson.plugins.gradle.Gradle.perform(Gradle.java:317)
	at hudson.tasks.BuildStepMonitor$1.perform(BuildStepMonitor.java:20)
	at hudson.model.AbstractBuild$AbstractBuildExecution.perform(AbstractBuild.java:806)
	at hudson.model.Build$BuildExecution.build(Build.java:198)
	at hudson.model.Build$BuildExecution.doRun(Build.java:163)
	at hudson.model.AbstractBuild$AbstractBuildExecution.run(AbstractBuild.java:514)
	at hudson.model.Run.execute(Run.java:1888)
	at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:43)
	at hudson.model.ResourceController.execute(ResourceController.java:99)
	at hudson.model.Executor.run(Executor.java:432)
Caused by: java.io.IOException: Unexpected termination of the channel
	at hudson.remoting.SynchronousCommandTransport$ReaderThread.run(SynchronousCommandTransport.java:75)
Caused by: java.io.EOFException
	at java.io.ObjectInputStream$PeekInputStream.readFully(ObjectInputStream.java:2799)
	at java.io.ObjectInputStream$BlockDataInputStream.readShort(ObjectInputStream.java:3274)
	at java.io.ObjectInputStream.readStreamHeader(ObjectInputStream.java:934)
	at java.io.ObjectInputStream.<init>(ObjectInputStream.java:396)
	at hudson.remoting.ObjectInputStreamEx.<init>(ObjectInputStreamEx.java:49)
	at hudson.remoting.Command.readFrom(Command.java:142)
	at hudson.remoting.Command.readFrom(Command.java:128)
	at hudson.remoting.AbstractSynchronousByteArrayCommandTransport.read(AbstractSynchronousByteArrayCommandTransport.java:35)
	at hudson.remoting.SynchronousCommandTransport$ReaderThread.run(SynchronousCommandTransport.java:61)
Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure
ERROR: apache-beam-jenkins-11 is offline; cannot locate jdk_1.8_latest

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #64

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/64/display/redirect>

Changes:


------------------------------------------
Failed to access build log

java.io.IOException: Pipe closed after 0 cycles
	at org.apache.sshd.common.channel.ChannelPipedInputStream.read(ChannelPipedInputStream.java:126)
	at org.apache.sshd.common.channel.ChannelPipedInputStream.read(ChannelPipedInputStream.java:105)
	at hudson.remoting.FlightRecorderInputStream.read(FlightRecorderInputStream.java:93)
	at hudson.remoting.ChunkedInputStream.readHeader(ChunkedInputStream.java:74)
	at hudson.remoting.ChunkedInputStream.readUntilBreak(ChunkedInputStream.java:104)
	at hudson.remoting.ChunkedCommandTransport.readBlock(ChunkedCommandTransport.java:39)
	at hudson.remoting.AbstractSynchronousByteArrayCommandTransport.read(AbstractSynchronousByteArrayCommandTransport.java:34)
	at hudson.remoting.SynchronousCommandTransport$ReaderThread.run(SynchronousCommandTransport.java:61)
Caused: hudson.remoting.ChannelClosedException: Channel "hudson.remoting.Channel@21f48f07:apache-beam-jenkins-14": channel is already closed
	at hudson.remoting.Request.checkIfCanBeExecutedOnChannel(Request.java:130)
	at hudson.remoting.UserRequest.checkIfCanBeExecutedOnChannel(UserRequest.java:114)
	at hudson.remoting.Request.call(Request.java:155)
	at hudson.remoting.Channel.call(Channel.java:1000)
	at hudson.FilePath.act(FilePath.java:1165)
	at hudson.FilePath.act(FilePath.java:1154)
	at hudson.FilePath.toURI(FilePath.java:1300)
	at hudson.tasks.MailSender.createFailureMail(MailSender.java:325)
	at hudson.tasks.MailSender.createMail(MailSender.java:182)
	at hudson.tasks.MailSender.run(MailSender.java:112)
	at hudson.tasks.Mailer.perform(Mailer.java:179)
	at hudson.tasks.Mailer.perform(Mailer.java:142)
	at hudson.tasks.BuildStepMonitor$1.perform(BuildStepMonitor.java:20)
	at hudson.model.AbstractBuild$AbstractBuildExecution.perform(AbstractBuild.java:806)
	at hudson.model.AbstractBuild$AbstractBuildExecution.performAllBuildSteps(AbstractBuild.java:755)
	at hudson.model.Build$BuildExecution.post2(Build.java:178)
	at hudson.model.AbstractBuild$AbstractBuildExecution.post(AbstractBuild.java:699)
	at hudson.model.Run.execute(Run.java:1913)
	at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:43)
	at hudson.model.ResourceController.execute(ResourceController.java:99)
	at hudson.model.Executor.run(Executor.java:432)

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #63

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/63/display/redirect>

Changes:


------------------------------------------
[...truncated 514.66 KB...]
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
Feb 21, 2022 7:40:32 PM org.apache.beam.sdk.metrics.MetricsEnvironment getCurrentContainer
WARNING: Reporting metrics are not supported in the current execution environment.
FATAL: command execution failed
java.io.IOException: Backing channel 'apache-beam-jenkins-5' is disconnected.
	at hudson.remoting.RemoteInvocationHandler.channelOrFail(RemoteInvocationHandler.java:216)
	at hudson.remoting.RemoteInvocationHandler.invoke(RemoteInvocationHandler.java:286)
	at com.sun.proxy.$Proxy121.isAlive(Unknown Source)
	at hudson.Launcher$RemoteLauncher$ProcImpl.isAlive(Launcher.java:1213)
	at hudson.Launcher$RemoteLauncher$ProcImpl.join(Launcher.java:1205)
	at hudson.Launcher$ProcStarter.join(Launcher.java:522)
	at hudson.plugins.gradle.Gradle.perform(Gradle.java:317)
	at hudson.tasks.BuildStepMonitor$1.perform(BuildStepMonitor.java:20)
	at hudson.model.AbstractBuild$AbstractBuildExecution.perform(AbstractBuild.java:806)
	at hudson.model.Build$BuildExecution.build(Build.java:198)
	at hudson.model.Build$BuildExecution.doRun(Build.java:163)
	at hudson.model.AbstractBuild$AbstractBuildExecution.run(AbstractBuild.java:514)
	at hudson.model.Run.execute(Run.java:1888)
	at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:43)
	at hudson.model.ResourceController.execute(ResourceController.java:99)
	at hudson.model.Executor.run(Executor.java:432)
Caused by: java.io.IOException: Pipe closed after 0 cycles
	at org.apache.sshd.common.channel.ChannelPipedInputStream.read(ChannelPipedInputStream.java:126)
	at org.apache.sshd.common.channel.ChannelPipedInputStream.read(ChannelPipedInputStream.java:105)
	at hudson.remoting.FlightRecorderInputStream.read(FlightRecorderInputStream.java:93)
	at hudson.remoting.ChunkedInputStream.readHeader(ChunkedInputStream.java:74)
	at hudson.remoting.ChunkedInputStream.readUntilBreak(ChunkedInputStream.java:104)
	at hudson.remoting.ChunkedCommandTransport.readBlock(ChunkedCommandTransport.java:39)
	at hudson.remoting.AbstractSynchronousByteArrayCommandTransport.read(AbstractSynchronousByteArrayCommandTransport.java:34)
	at hudson.remoting.SynchronousCommandTransport$ReaderThread.run(SynchronousCommandTransport.java:61)
Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure
ERROR: apache-beam-jenkins-5 is offline; cannot locate jdk_1.8_latest

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #62

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/62/display/redirect?page=changes>

Changes:

[rogelio.hernandez] [BEAM-13051] Pylint misplaced-bare-raise warning enabled

[Jeff Tapper] Update Java LTS roadmap info on website for Java 17

[Pablo Estrada] Simplify README for new users

[Kyle Weaver] [BEAM-13106] Support Flink 1.14.

[Kyle Weaver] [BEAM-13106] Reuse executor instead of shutting it down mid-test.

[Kyle Weaver] [BEAM-13106] Prevent infinite wait in Flink savepoint test.

[Ismaël Mejía] [BEAM-13202] Fix typos on tests names for VarianceFnTest

[Kenneth Knowles] Disable AfterSynchronizedProcessingTime test on Dataflow

[Ismaël Mejía] [BEAM-13202] Add Coder to CountIfFn.Accum

[Ismaël Mejía] [BEAM-13202] Reuse Count transform code since CountIf is a specific case

[Kenneth Knowles] Add test category UsesProcessingTimeTimers

[Kenneth Knowles] Label tests that need UsesProcessingTimeTimers

[Kenneth Knowles] Exclude UsesProcessingTimeTimers from SamzaRunner tests

[Kyle Weaver] [BEAM-13106] A couple additional fixes to FlinkSavepointTest.

[mmack] [adhoc] Migrate KinesisIOIT to use ITEnvironment for Localstack based IT

[rogelio.hernandez] [BEAM-13051] Added descriptions to Kinesis and PortableRunner exceptions

[noreply] [BEAM-13955] Fix pylint breakage from #16836 (#16867)

[relax] Fix TableRow conversion for the case of fields named "f"

[noreply] Bump dataflow.fnapi_container_version (#16874)

[mmack] [BEAM-13563] Introducing common AWS ClientBuilderFactory to unify

[laraschmidt] Fix final allowskew error to properly handle a large allowedSkew

[noreply] Case studies page improvements (#16702)

[noreply] [BEAM-13946] Add get_dummies(), a non-deferred column operation on

[noreply] [release-2.36.0] Fix pickler argument for 2.36 blog (#16774)

[thiagotnunes] fix: fix bug when retrieving either string or json

[noreply] [adhoc] Avoid using SerializablePipelineOptions for testing to minimize

[noreply] [BEAM-13812] Integrate DataprocClusterManager into Interactive

[noreply] [BEAM-12572] Fix failing python examples tests in Dataflow runner

[noreply] Remove build status from PR (#16902)


------------------------------------------
[...truncated 455.93 KB...]
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    ./dist_proc/dax/workflow/****/streaming/windowing_api_delegate.h:67
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    ./dist_proc/dax/workflow/****/streaming/windowing_api_delegate.h:67
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    ./dist_proc/dax/workflow/****/streaming/windowing_api_delegate.h:67
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    ./dist_proc/dax/workflow/****/streaming/windowing_api_delegate.h:67
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    ./dist_proc/dax/workflow/****/streaming/windowing_api_delegate.h:67
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. T
Feb 19, 2022 2:23:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
SEVERE: 2022-02-19T14:23:53.629Z: Staged package hamcrest-2.1-upOy46ViMiukMvChtTrdzFXLGIJTMZoCDtd_gk5pIFA.jar at location 'gs://temp-storage-for-perf-tests/loadtests/staging/hamcrest-2.1-upOy46ViMiukMvChtTrdzFXLGIJTMZoCDtd_gk5pIFA.jar' is inaccessible.
Feb 19, 2022 2:23:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
SEVERE: 2022-02-19T14:23:53.673Z: Staged package hamcrest-core-2.1-4JEJ5UoonYhQa5v-yYfd0Zn0IXyUZBMmaDUbmk8Avuk.jar at location 'gs://temp-storage-for-perf-tests/loadtests/staging/hamcrest-core-2.1-4JEJ5UoonYhQa5v-yYfd0Zn0IXyUZBMmaDUbmk8Avuk.jar' is inaccessible.
Feb 19, 2022 2:23:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
SEVERE: 2022-02-19T14:23:53.782Z: Staged package j2objc-annotations-1.3-Ia8wySJnvWEiwOC00gzMtmQaN-r5VsZUDsRx1YTmSns.jar at location 'gs://temp-storage-for-perf-tests/loadtests/staging/j2objc-annotations-1.3-Ia8wySJnvWEiwOC00gzMtmQaN-r5VsZUDsRx1YTmSns.jar' is inaccessible.
Feb 19, 2022 2:23:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
SEVERE: 2022-02-19T14:23:53.925Z: Staged package jackson-core-asl-1.9.13-RAqctcqVshX5U9OiCmsaENofCbUpqd3qX4pJBd2rT1o.jar at location 'gs://temp-storage-for-perf-tests/loadtests/staging/jackson-core-asl-1.9.13-RAqctcqVshX5U9OiCmsaENofCbUpqd3qX4pJBd2rT1o.jar' is inaccessible.
Feb 19, 2022 2:23:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
SEVERE: 2022-02-19T14:23:54.056Z: Staged package jackson-mapper-asl-1.9.13-dOegenby7breKTEqWi68z6AZEovAIezjhW12GX6b4MI.jar at location 'gs://temp-storage-for-perf-tests/loadtests/staging/jackson-mapper-asl-1.9.13-dOegenby7breKTEqWi68z6AZEovAIezjhW12GX6b4MI.jar' is inaccessible.
Feb 19, 2022 2:23:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
SEVERE: 2022-02-19T14:23:54.165Z: Staged package javax.annotation-api-1.3.2-4EulGVvNVV3JVlD3zGFNFR5LzVLSmhC4qiGX86uJq5s.jar at location 'gs://temp-storage-for-perf-tests/loadtests/staging/javax.annotation-api-1.3.2-4EulGVvNVV3JVlD3zGFNFR5LzVLSmhC4qiGX86uJq5s.jar' is inaccessible.
Feb 19, 2022 2:23:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
SEVERE: 2022-02-19T14:23:54.237Z: Staged package json-20200518-D_MObagvFa-vDna9bplqerVWaFEWaoP5c8ZKmg-UbS0.jar at location 'gs://temp-storage-for-perf-tests/loadtests/staging/json-20200518-D_MObagvFa-vDna9bplqerVWaFEWaoP5c8ZKmg-UbS0.jar' is inaccessible.
Feb 19, 2022 2:23:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
WARNING: 2022-02-19T14:23:55.298Z: One or more access checks for temp location or staged files failed. Please refer to other error messages for details. For more information on security and permissions, please see https://cloud.google.com/dataflow/security-and-permissions.
Feb 19, 2022 3:40:41 PM org.apache.beam.sdk.metrics.MetricsEnvironment getCurrentContainer
WARNING: Reporting metrics are not supported in the current execution environment.
Feb 19, 2022 4:00:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-19T16:00:43.100Z: Cancel request is committed for workflow job: 2022-02-19_04_38_38-4711337483547601338.
Feb 19, 2022 4:00:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-19T16:00:43.134Z: Cleaning up.
Feb 19, 2022 4:00:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-19T16:00:43.190Z: Stopping **** pool...
Feb 19, 2022 4:00:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-19T16:00:43.236Z: Stopping **** pool...
Feb 19, 2022 4:03:06 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-19T16:03:04.605Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Feb 19, 2022 4:03:06 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-19T16:03:04.639Z: Worker pool stopped.
Feb 19, 2022 4:03:11 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-02-19_04_38_38-4711337483547601338 finished with status CANCELLED.
Load test results for test (ID): a661d839-1906-49f5-8968-83235c00983c and timestamp: 2022-02-19T12:38:32.929000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                 11991.314
dataflow_v2_java17_total_bytes_count             3.54713284E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220219123614
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:1293f3c1f5c749373548c4fedf4150758d7d60d2d7845b2b7fb2524854eefd81
Deleted: sha256:3c423726ab9363e4d00408d34380976b419030ef1d5b24721111c53fa66de2c4
Deleted: sha256:75411ec94b698ca7db00419fe5fef5ae8b9eb67912f1dad43163bc0066f66b87
Deleted: sha256:b837c2438bd11ea3ed3a0fbc2185b257d80862ab5005fd53d7dd52577ad1a1f6
Deleted: sha256:126a3b5e3e5d30f6f1f798b00ff4ea7bdb7e7790a13d16632e23ae80cdb90324
Deleted: sha256:97481abd42016c7e93b5841e078b6d0c469f45a67ddd4815f7facd6a27b30271
Deleted: sha256:5e68b61330d91fef3b4c22c908b8e841a7a6f8c95f4d5d32d0717a1115c97037
Deleted: sha256:21a4dbbf7614b54914919ca98eeb2ef5cdf1ac41d19bf63bb0c7ddae0bed06c0
Deleted: sha256:4846de45ce156a7f946bbfe8f463954f317729001274fc4694aa1d1893c3b0e0
Deleted: sha256:bbe767b371f8dcba12de888f099b06500f9aa9b08e0b73ecd93d836af14d093d
Deleted: sha256:d7459e76138ded7944e88fe82a0fa6e76e6ef104315ace45778ee4777872cad1
Deleted: sha256:2563084365745696356ceaf0fbc056076e4c418a7431a395d2e46f10ae3b0c90
Deleted: sha256:1ab8283bb118ec225ad0a73d7c60ba9afe9e1eda394517e1870f8dff627b0e3b
Deleted: sha256:5edd13676f82292ac6590afb4521c72d433a3068b5f2fb64b344e44036bfbdd5
Deleted: sha256:58b497267e94725c3e97f20c9d28aaae7191efac8cf6c0fa6448a9dbf7cee9c5
Deleted: sha256:d6524ba6f158924859aeefc3e8778ea997c8e1b86ed378694e7c37feaf24ca90
Deleted: sha256:96ce392f8327872fe9c035229632c49cf1c7aae1a315302c6fa9b6bd4cc18b0d
Deleted: sha256:bc6db1b36044863caa156087444225e75c0dd668fe6d4287d5c90b553bd04a1c
Deleted: sha256:db8083f907bf23bd05a0ced5da72e684eadc18f781a48a633bc39dc7be21747d
Deleted: sha256:564b42f7ab387caeec0292ab455cb7d6767c6e81bf29ad8930323deb26d27c5c
Deleted: sha256:fea99065cb0d8f68f6e35f220096eb9776698737ed808486f2e4c5e869d8bafd
Deleted: sha256:080e7ca1d07ecf8d49463bddcc9a0dbfd45017d5b5d283e54d9584806a879735
Deleted: sha256:3b4552e7d456914d5716e1b4d6b2c8bc04979ecdc651ce238ef21dd459f4078c
Deleted: sha256:08e71de8d981e31c6b8b0bbcecfd71463f38a7d22c36fc7678c2c001cdc6a2ed
Deleted: sha256:e3ba66dabc62713259f986f96730a3b752df8d4134bd43d538731d7fd47ce275
Deleted: sha256:7021395496e04813f8c35a31e60547419ef2f0a731688de2a72ee8ee7b8996d7
Deleted: sha256:d6c18b576f9bcaefb2e2d0e38151df6876967cef9c2a4cf8133e8957ad82a9ce
Deleted: sha256:5222263beed96383880e29c1547e7d49e880313e94418cc087b3dc81306664a1
Deleted: sha256:646ebe7ea944feb146e1ffdb74cab1764b4ccb1782711b9d70ac1457096143b1
Deleted: sha256:dd94fb87a211d54a188763c51ee0e61f9effbc286f1bd057ab70436ddf64e638
Deleted: sha256:cb37765bc16b8d89996dedfc2476105148f9c3381ef35bca6dbd43ca83e31bd2
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220219123614]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:1293f3c1f5c749373548c4fedf4150758d7d60d2d7845b2b7fb2524854eefd81]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220219123614] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:1293f3c1f5c749373548c4fedf4150758d7d60d2d7845b2b7fb2524854eefd81])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:1293f3c1f5c749373548c4fedf4150758d7d60d2d7845b2b7fb2524854eefd81
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:1293f3c1f5c749373548c4fedf4150758d7d60d2d7845b2b7fb2524854eefd81
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:1293f3c1f5c749373548c4fedf4150758d7d60d2d7845b2b7fb2524854eefd81].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 27m 21s
109 actionable tasks: 72 executed, 33 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/7namch4m7i442

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #59

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/59/display/redirect?page=changes>

Changes:

[Kyle Weaver] [BEAM-12712] Spark: Exclude looping timer tests.

[Kyle Weaver] [BEAM-13919] Annotate PerKeyOrderingTest with UsesStatefulParDo.

[noreply] Update 2.36.0 blog post to mention ARM64 support

[stranniknm] [BEAM-13785] playground - enable scio sdk

[noreply] Minor: Disable checker framework in nightly snapshot (#16829)

[artur.khanin] Updated example link

[noreply] [BEAM-13860] Make `DoFn.infer_output_type` return element type (#16788)

[noreply] [BEAM-13894] Unit test utilities in the ptest package (#16830)

[Kenneth Knowles] Add test for processing time continuation trigger

[noreply] [BEAM-13922] [Coverage] Make boot.go more testable and add tests

[noreply] Exclude SpannerChangeStream IT from Dataflow V1 postcommit (#16851)

[noreply] [BEAM-13930] Address StateSpec consistency issue between Runner and Fn

[mattcasters] [BEAM-13854] Document casting trick for Avro value serializer in KafkaIO

[noreply] Merge pull request #16838 from [BEAM-13931] - make sure large rows cause

[noreply] Seznam Case Study (#16825)

[noreply] [Website] Apache Hop Case Study (#16824)

[noreply] [BEAM-13694] Force hadoop-hdfs-client in hadoopVersion tests for hdfs

[noreply] [Website] Ricardo - added case study feedback (#16807)

[noreply] Merge pull request #16735 from [BEAM-13827] - fix medium file size

[noreply] Merge pull request #16753 from [BEAM-13837] [Playground] show graph on


------------------------------------------
[...truncated 796.28 KB...]
    ./dist_proc/dax/workflow/****/streaming/windowing_api_delegate.h:67
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    ./dist_proc/dax/workflow/****/streaming/windowing_api_delegate.h:67
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    ./dist_proc/dax/workflow/****/streaming/windowing_api_delegate.h:67
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    ./dist_proc/dax/workflow/****/streaming/windowing_api_delegate.h:67
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    ./dist_proc/dax/workflow/****/streaming/windowing_api_delegate.h:67
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    ./dist_proc/dax/workflow/****/streaming/windowing_api_delegate.h:67
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    ./dist_proc/dax/workflow/****/streaming/windowing_api_delegate.h:67
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    ./dist_proc/dax/workflow/****/streaming/windowing_api_delegate.h:67
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    ./dist_proc/dax/workflow/****/streaming/windowing_api_delegate.h:67
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    ./dist_proc/dax/workflow/****/streaming/windowing_api_delegate.h:67
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    ./dist_proc/dax/workflow/****/streaming/windowing_api_delegate.h:67
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    ./dist_proc/dax/workflow/****/streaming/windowing_api_delegate.h:67
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    ./dist_proc/dax/workflow/****/streaming/windowing_api_delegate.h:67
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    ./dist_proc/dax/workflow/****/streaming/windowing_api_delegate.h:67
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    ./dist_proc/dax/workflow/****/streaming/windowing_api_delegate.h:67
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    ./dist_proc/dax/workflow/****/streaming/windowing_api_delegate.h:67
==>
    dist_proc/dax/workflow/wor
Feb 16, 2022 4:03:37 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-16T16:03:36.547Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Feb 16, 2022 4:03:37 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-16T16:03:36.585Z: Worker pool stopped.
Feb 16, 2022 4:03:43 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-02-16_04_33_51-2863211749118102678 finished with status CANCELLED.
Load test results for test (ID): 0bb67131-b617-4669-b7c9-1d57006e8c90 and timestamp: 2022-02-16T12:33:44.951000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                 12270.358
dataflow_v2_java17_total_bytes_count              3.4442286E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220216123127
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:c8ede2acb1dd5f9e5d3a6ed6381a1866f89dd0d1cff8bbae002cf2d6f8716ba7
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220216123127]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:c8ede2acb1dd5f9e5d3a6ed6381a1866f89dd0d1cff8bbae002cf2d6f8716ba7]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220216123127] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:c8ede2acb1dd5f9e5d3a6ed6381a1866f89dd0d1cff8bbae002cf2d6f8716ba7])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:c8ede2acb1dd5f9e5d3a6ed6381a1866f89dd0d1cff8bbae002cf2d6f8716ba7
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:c8ede2acb1dd5f9e5d3a6ed6381a1866f89dd0d1cff8bbae002cf2d6f8716ba7
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:c8ede2acb1dd5f9e5d3a6ed6381a1866f89dd0d1cff8bbae002cf2d6f8716ba7].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 32m 36s
109 actionable tasks: 72 executed, 33 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/564qjl5th5j2y

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #58

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/58/display/redirect?page=changes>

Changes:

[Kyle Weaver] [BEAM-12712] Spark: Exclude looping timer tests.

[Kyle Weaver] [BEAM-13919] Annotate PerKeyOrderingTest with UsesStatefulParDo.

[noreply] Update 2.36.0 blog post to mention ARM64 support

[noreply] Minor: Disable checker framework in nightly snapshot (#16829)

[artur.khanin] Updated example link

[noreply] [BEAM-13860] Make `DoFn.infer_output_type` return element type (#16788)

[noreply] [BEAM-13894] Unit test utilities in the ptest package (#16830)

[Kenneth Knowles] Add test for processing time continuation trigger

[noreply] [BEAM-13922] [Coverage] Make boot.go more testable and add tests

[noreply] Exclude SpannerChangeStream IT from Dataflow V1 postcommit (#16851)

[noreply] [BEAM-13930] Address StateSpec consistency issue between Runner and Fn


------------------------------------------
[...truncated 81.96 KB...]
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
Feb 15, 2022 12:46:18 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
SEVERE: 2022-02-15T12:46:17.104Z: generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
Feb 15, 2022 12:46:20 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
SEVERE: 2022-02-15T12:46:18.268Z: generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
Feb 15, 2022 12:46:23 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
SEVERE: 2022-02-15T12:46:20.731Z: generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
FATAL: command execution failed
java.io.IOException: Backing channel 'apache-beam-jenkins-1' is disconnected.
	at hudson.remoting.RemoteInvocationHandler.channelOrFail(RemoteInvocationHandler.java:216)
	at hudson.remoting.RemoteInvocationHandler.invoke(RemoteInvocationHandler.java:286)
	at com.sun.proxy.$Proxy139.isAlive(Unknown Source)
	at hudson.Launcher$RemoteLauncher$ProcImpl.isAlive(Launcher.java:1211)
	at hudson.Launcher$RemoteLauncher$ProcImpl.join(Launcher.java:1203)
	at hudson.Launcher$ProcStarter.join(Launcher.java:523)
	at hudson.plugins.gradle.Gradle.perform(Gradle.java:317)
	at hudson.tasks.BuildStepMonitor$1.perform(BuildStepMonitor.java:20)
	at hudson.model.AbstractBuild$AbstractBuildExecution.perform(AbstractBuild.java:803)
	at hudson.model.Build$BuildExecution.build(Build.java:197)
	at hudson.model.Build$BuildExecution.doRun(Build.java:163)
	at hudson.model.AbstractBuild$AbstractBuildExecution.run(AbstractBuild.java:513)
	at hudson.model.Run.execute(Run.java:1906)
	at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:43)
	at hudson.model.ResourceController.execute(ResourceController.java:97)
	at hudson.model.Executor.run(Executor.java:429)
Caused by: java.io.IOException: Pipe closed after 0 cycles
	at org.apache.sshd.common.channel.ChannelPipedInputStream.read(ChannelPipedInputStream.java:118)
	at org.apache.sshd.common.channel.ChannelPipedInputStream.read(ChannelPipedInputStream.java:101)
	at hudson.remoting.FlightRecorderInputStream.read(FlightRecorderInputStream.java:93)
	at hudson.remoting.ChunkedInputStream.readHeader(ChunkedInputStream.java:74)
	at hudson.remoting.ChunkedInputStream.readUntilBreak(ChunkedInputStream.java:104)
	at hudson.remoting.ChunkedCommandTransport.readBlock(ChunkedCommandTransport.java:39)
	at hudson.remoting.AbstractSynchronousByteArrayCommandTransport.read(AbstractSynchronousByteArrayCommandTransport.java:34)
	at hudson.remoting.SynchronousCommandTransport$ReaderThread.run(SynchronousCommandTransport.java:61)
Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure
ERROR: apache-beam-jenkins-1 is offline; cannot locate jdk_1.8_latest

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #57

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/57/display/redirect?page=changes>

Changes:

[akustov] fix name project id from secreton scio deploy action

[alexander.zhuravlev] [BEAM-13775] Fixed bug with run button

[ihr] [BEAM-13836] Fix the answers placeholders locations in the Python katas

[noreply] Merge pull request #16703 from [BEAM-13804][Playground][Bugfix] Add

[noreply] Merge pull request #16611 from [BEAM-13712][Playground] Add graph for

[noreply] Merge pull request #16757 from [BEAM-13655] [Playground] Persist the


------------------------------------------
[...truncated 94.10 KB...]
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
Feb 14, 2022 12:40:14 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
SEVERE: 2022-02-14T12:40:13.268Z: generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
Feb 14, 2022 12:40:16 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
SEVERE: 2022-02-14T12:40:14.677Z: generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
Feb 14, 2022 12:40:16 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
SEVERE: 2022-02-14T12:40:15.260Z: generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
Feb 14, 2022 12:40:17 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
SEVERE: 2022-02-14T12:40:17.728Z: generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
Feb 14, 2022 12:40:23 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
SEVERE: 2022-02-14T12:40:20.302Z: generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
FATAL: command execution failed
java.io.IOException: Backing channel 'apache-beam-jenkins-11' is disconnected.
	at hudson.remoting.RemoteInvocationHandler.channelOrFail(RemoteInvocationHandler.java:216)
	at hudson.remoting.RemoteInvocationHandler.invoke(RemoteInvocationHandler.java:286)
	at com.sun.proxy.$Proxy135.isAlive(Unknown Source)
	at hudson.Launcher$RemoteLauncher$ProcImpl.isAlive(Launcher.java:1211)
	at hudson.Launcher$RemoteLauncher$ProcImpl.join(Launcher.java:1203)
	at hudson.Launcher$ProcStarter.join(Launcher.java:523)
	at hudson.plugins.gradle.Gradle.perform(Gradle.java:317)
	at hudson.tasks.BuildStepMonitor$1.perform(BuildStepMonitor.java:20)
	at hudson.model.AbstractBuild$AbstractBuildExecution.perform(AbstractBuild.java:803)
	at hudson.model.Build$BuildExecution.build(Build.java:197)
	at hudson.model.Build$BuildExecution.doRun(Build.java:163)
	at hudson.model.AbstractBuild$AbstractBuildExecution.run(AbstractBuild.java:513)
	at hudson.model.Run.execute(Run.java:1906)
	at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:43)
	at hudson.model.ResourceController.execute(ResourceController.java:97)
	at hudson.model.Executor.run(Executor.java:429)
Caused by: java.io.IOException: Unexpected termination of the channel
	at hudson.remoting.SynchronousCommandTransport$ReaderThread.run(SynchronousCommandTransport.java:75)
Caused by: java.io.EOFException
	at java.io.ObjectInputStream$PeekInputStream.readFully(ObjectInputStream.java:2799)
	at java.io.ObjectInputStream$BlockDataInputStream.readShort(ObjectInputStream.java:3274)
	at java.io.ObjectInputStream.readStreamHeader(ObjectInputStream.java:934)
	at java.io.ObjectInputStream.<init>(ObjectInputStream.java:396)
	at hudson.remoting.ObjectInputStreamEx.<init>(ObjectInputStreamEx.java:49)
	at hudson.remoting.Command.readFrom(Command.java:142)
	at hudson.remoting.Command.readFrom(Command.java:128)
	at hudson.remoting.AbstractSynchronousByteArrayCommandTransport.read(AbstractSynchronousByteArrayCommandTransport.java:35)
	at hudson.remoting.SynchronousCommandTransport$ReaderThread.run(SynchronousCommandTransport.java:61)
Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure
ERROR: apache-beam-jenkins-11 is offline; cannot locate jdk_1.8_latest

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #56

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/56/display/redirect?page=changes>

Changes:

[randomstep] [BEAM-9195] Bump org.testcontainers to 1.16.3


------------------------------------------
[...truncated 1.42 MB...]
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    ./dist_proc/dax/workflow/****/streaming/windowing_api_delegate.h:67
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    ./dist_proc/dax/workflow/****/streaming/windowing_api_delegate.h:67
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    ./dist_proc/dax/workflow/****/streaming/windowing_api_delegate.h:67
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    ./dist_proc/dax/workflow/****/streaming/windowing_api_delegate.h:67
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    ./dist_proc/dax/workflow/****/streaming/windowing_api_delegate.h:67
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already c
Feb 13, 2022 3:26:38 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
SEVERE: 2022-02-13T15:26:37.007Z: generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
Feb 13, 2022 4:01:04 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-13T16:01:04.299Z: Cancel request is committed for workflow job: 2022-02-13_04_32_57-17054085345386009290.
Feb 13, 2022 4:01:04 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-13T16:01:04.329Z: Cleaning up.
Feb 13, 2022 4:01:06 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-13T16:01:04.398Z: Stopping **** pool...
Feb 13, 2022 4:01:06 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-13T16:01:04.444Z: Stopping **** pool...
Feb 13, 2022 4:03:28 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-13T16:03:27.772Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Feb 13, 2022 4:03:28 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-13T16:03:27.812Z: Worker pool stopped.
Feb 13, 2022 4:03:35 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-02-13_04_32_57-17054085345386009290 finished with status CANCELLED.
Load test results for test (ID): 89ea341a-d48c-4d4f-a68a-835fd3c80afa and timestamp: 2022-02-13T12:32:51.044000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                 12351.654
dataflow_v2_java17_total_bytes_count             2.45877318E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220213123036
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:c78fa28d22ea2ecabe9ebcf56db510e02ec0ac86d00649e4c76b38678cf50423
Deleted: sha256:b38900ddaa3744c3412364ade4945a28d05f61902c8a65d92e9585be13360d22
Deleted: sha256:116368e088109b3cf43860ef7bd64173f04cb5ec2c0e1a1e4c83fe799cbafa91
Deleted: sha256:88ae632e197f64b6ea5e8857f812e0eb0edc4511077c8b333fc14718c78cbae9
Deleted: sha256:507dc1495395eb9ff915fe353cc52f8e4d16223abdb86d887403d8049edac62c
Deleted: sha256:8e575a9ef573c3fc9b90adf0be557d452e9d558ed85a140b146a1f92113028b8
Deleted: sha256:1c09bc4b47f058ad6d4792b4af872ee16a9721ed551ab79088eb92f6cbf82984
Deleted: sha256:e8aa285ac53e9355c298c239dcf095b12ecbaf900d84c71801e153d0cdc17c46
Deleted: sha256:967a551ecf455a99ad89c1cd90dcd7611cb94e6da9dcf2026aedb97c732c9e90
Deleted: sha256:2a14a5bf1bad8c4325e75c6ecd80a7d16ccfad980dfe082d26bf89faa0eb03fc
Deleted: sha256:a696761ea3a383b28ca46f462a2cdf89199e6477f3db3263456880c4cf54cdf3
Deleted: sha256:69d0d70d94d7ba9a075e52f0467a7b80b06ad6fa25ce3a36d55cb4f173ec0629
Deleted: sha256:1f54fd3e21e08ba50fb9e443d703ea7c1ffdc7940a9ef933fe7d840cb5b2013e
Deleted: sha256:cdaa481df1ba52f7eab9fceda685bdddf54fc60a8ff850a31b581e084c24b3d9
Deleted: sha256:677ee4a7e7cc96b04f7e3c2521038a1aef9cc2bf89135b16db0424f4b1b3455e
Deleted: sha256:1093766a6c348b424083612a6b4c72587fedf82ebff6c1b7e444a835bc91d306
Deleted: sha256:9d2a34412fe0091c943de4ac27ba31e0ce4962dd8988d261ded53bae11450a9e
Deleted: sha256:c38a91e024fe7eed035920c40ec03a6fe0a706fd0d107f2f41600e43880f3bfc
Deleted: sha256:33d786122483441fe491b5b9dc332a6295eb9cc2c0a33e8da950030ceefef37c
Deleted: sha256:63995908f7e73e2442a663afb7d7be641eeea87811a1ec127654974ee9f59af8
Deleted: sha256:649df3ea8ff86884b7d9ee64ddf02af24dd5186ac66501f659413a904ab802e3
Deleted: sha256:fd5aed5a550a1487192af1d38dfccb1c6688d8e91c48b372e935cce297b3fb2a
Deleted: sha256:3ee0d037dc91aefa2701735324d1fad4470c2209d7152e76b54b3761950a89e4
Deleted: sha256:fe8173898264ab1628c85b73ac6c7a41f0409122ad3468d0801d8180d5e6fae8
Deleted: sha256:9975e1a2e920e7c63e3f5e456a2957e8a91d9ab871b7cdfbc522928bf892a4cf
Deleted: sha256:2c910015e66651cd4793fb920fb0c1173a10dc8dbb40adfa9eb20734965117cd
Deleted: sha256:d648e0e5378f38662c0bb68766f9bcfa4e0645e6136dc1b03aa1b59b4576c83d
Deleted: sha256:acc3e8ce99aabc18a0f9fbfb901148e95464b2dc6141e0e9152f5fd9f339d6f7
Deleted: sha256:1642cc3fcf019bed2dd31fc9164e9b1d1130a048d47be8e50e8e7cefb1a470aa
Deleted: sha256:0eb40d3513aa178f72b7aff1865f5a150504f8d81dc156fdce0809c5390dcf9b
Deleted: sha256:249e6fb5cee4fca168ee01915583e5bd51744774bf667cc442d281841e8b2e23
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220213123036]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:c78fa28d22ea2ecabe9ebcf56db510e02ec0ac86d00649e4c76b38678cf50423]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220213123036] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:c78fa28d22ea2ecabe9ebcf56db510e02ec0ac86d00649e4c76b38678cf50423])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:46392edaf7ded0ab88c2474a6b01107b71ec81f34c489bc1030a5d2c3cdc0bab
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:46392edaf7ded0ab88c2474a6b01107b71ec81f34c489bc1030a5d2c3cdc0bab
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:46392edaf7ded0ab88c2474a6b01107b71ec81f34c489bc1030a5d2c3cdc0bab].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:c78fa28d22ea2ecabe9ebcf56db510e02ec0ac86d00649e4c76b38678cf50423
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:c78fa28d22ea2ecabe9ebcf56db510e02ec0ac86d00649e4c76b38678cf50423
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:c78fa28d22ea2ecabe9ebcf56db510e02ec0ac86d00649e4c76b38678cf50423].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 33m 22s
109 actionable tasks: 73 executed, 32 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/axhex3yhbp7fy

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #55

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/55/display/redirect?page=changes>

Changes:

[Carl Yeksigian] Cache bucket matcher regex in GcsPath

[benjamin.gonzalez] [BEAM-12672] Retry flaky tests

[benjamin.gonzalez] [BEAM-12672] Fix spotlessApply

[laraschmidt] Fixing the log line to properly handle a large allowed skew.

[n] BEAM-13159 Update embedded-redis dependency

[n] address comments

[noreply] Minor: Add 2.38.0 section to CHANGES.md (#16804)

[noreply] [BEAM-12000] Fix typo in portable Python job definition (#16812)

[noreply] [BEAM-12164]: Fixes SpannerChangeStreamIT (#16806)

[noreply] [BEAM-12572] Fix failures in python examples tests (#16781)

[noreply] [BEAM-13921] filter out debeziumIO test for spark runner (#16815)

[noreply] [BEAM-13855] Skip SpannerChangeStreamOrderedWithinKeyIT and

[noreply] [BEAM-13679] playground - move quick start category to the top (#16808)

[noreply] Update license_script.sh (#16789)

[noreply] [BEAM-13908] [Coverage] Better testing coverage for gcpopts (#16816)

[noreply] Merge pull request #16809 from [BEAM-12164] Added integration test for

[noreply] [BEAM-4032]Support staging binary distributions of dependency packages

[noreply] [BEAM-13834] Increase influxDB persistent storage. (#16817)

[noreply] Minor: Fix link to nexmark benchmarks (#16803)

[noreply] Regenerate python container base_image_requirements.txt (#16832)


------------------------------------------
[...truncated 367.74 KB...]
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server
Feb 12, 2022 4:00:36 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-12T16:00:35.749Z: Cancel request is committed for workflow job: 2022-02-12_04_39_03-7325153169803150548.
Feb 12, 2022 4:00:36 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-12T16:00:35.838Z: Cleaning up.
Feb 12, 2022 4:00:36 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-12T16:00:35.914Z: Stopping **** pool...
Feb 12, 2022 4:00:36 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-12T16:00:35.972Z: Stopping **** pool...
Feb 12, 2022 4:00:46 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
SEVERE: 2022-02-12T16:00:44.613Z: generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
Feb 12, 2022 4:03:07 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-12T16:03:04.934Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Feb 12, 2022 4:03:07 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-12T16:03:04.975Z: Worker pool stopped.
Feb 12, 2022 4:03:12 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-02-12_04_39_03-7325153169803150548 finished with status CANCELLED.
Load test results for test (ID): a119abca-b425-40dc-9cfe-e9cb33f115fa and timestamp: 2022-02-12T12:38:57.731000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                 11915.867
dataflow_v2_java17_total_bytes_count             3.11247384E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220212123631
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:6868150c55d77b670e18cb7af7716c528a53dc7e9afaf6070ed4ee3dfaf62a95
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220212123631]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:6868150c55d77b670e18cb7af7716c528a53dc7e9afaf6070ed4ee3dfaf62a95]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220212123631] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:6868150c55d77b670e18cb7af7716c528a53dc7e9afaf6070ed4ee3dfaf62a95])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:6868150c55d77b670e18cb7af7716c528a53dc7e9afaf6070ed4ee3dfaf62a95
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:6868150c55d77b670e18cb7af7716c528a53dc7e9afaf6070ed4ee3dfaf62a95
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:6868150c55d77b670e18cb7af7716c528a53dc7e9afaf6070ed4ee3dfaf62a95].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 27m 4s
109 actionable tasks: 72 executed, 33 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/x5zzcq4hc6jr6

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #54

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/54/display/redirect?page=changes>

Changes:

[Carl Yeksigian] Cache bucket matcher regex in GcsPath

[benjamin.gonzalez] [BEAM-12672] Retry flaky tests

[benjamin.gonzalez] [BEAM-12672] Fix spotlessApply

[laraschmidt] Fixing the log line to properly handle a large allowed skew.

[n] BEAM-13159 Update embedded-redis dependency

[n] address comments

[noreply] Minor: Add 2.38.0 section to CHANGES.md (#16804)

[noreply] [BEAM-12000] Fix typo in portable Python job definition (#16812)

[noreply] [BEAM-12164]: Fixes SpannerChangeStreamIT (#16806)

[noreply] [BEAM-12572] Fix failures in python examples tests (#16781)

[noreply] [BEAM-13921] filter out debeziumIO test for spark runner (#16815)

[noreply] [BEAM-13855] Skip SpannerChangeStreamOrderedWithinKeyIT and

[noreply] [BEAM-13679] playground - move quick start category to the top (#16808)

[noreply] Update license_script.sh (#16789)

[noreply] [BEAM-13908] [Coverage] Better testing coverage for gcpopts (#16816)

[noreply] Merge pull request #16809 from [BEAM-12164] Added integration test for

[noreply] [BEAM-4032]Support staging binary distributions of dependency packages


------------------------------------------
[...truncated 169.41 KB...]
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server
Feb 11, 2022 6:24:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
SEVERE: 2022-02-11T18:24:43.383Z: generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
generic::internal: The work item requesting state read is no longer valid on the backend. The work has already completed or will be retried. This is expected during autoscaling events.
passed through:
==>
    dist_proc/dax/workflow/****/streaming/streaming_rpc_windmill_service_server.cc:688
==>
    dist_proc/dax/workflow/****/streaming/merge_windows_fn.cc:214
Build timed out (after 240 minutes). Marking the build as aborted.
FATAL: command execution failed
hudson.remoting.ChannelClosedException: Channel "hudson.remoting.Channel@7b5cc6d:apache-beam-jenkins-3": Remote call on apache-beam-jenkins-3 failed. The channel is closing down or has closed down
	at hudson.remoting.Channel.call(Channel.java:994)
	at hudson.remoting.RemoteInvocationHandler.invoke(RemoteInvocationHandler.java:286)
	at com.sun.proxy.$Proxy128.isAlive(Unknown Source)
	at hudson.Launcher$RemoteLauncher$ProcImpl.isAlive(Launcher.java:1211)
	at hudson.Launcher$RemoteLauncher$ProcImpl.join(Launcher.java:1203)
	at hudson.Launcher$ProcStarter.join(Launcher.java:523)
	at hudson.plugins.gradle.Gradle.perform(Gradle.java:317)
	at hudson.tasks.BuildStepMonitor$1.perform(BuildStepMonitor.java:20)
	at hudson.model.AbstractBuild$AbstractBuildExecution.perform(AbstractBuild.java:803)
	at hudson.model.Build$BuildExecution.build(Build.java:197)
	at hudson.model.Build$BuildExecution.doRun(Build.java:163)
	at hudson.model.AbstractBuild$AbstractBuildExecution.run(AbstractBuild.java:513)
	at hudson.model.Run.execute(Run.java:1906)
	at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:43)
	at hudson.model.ResourceController.execute(ResourceController.java:97)
	at hudson.model.Executor.run(Executor.java:429)
Caused by: java.io.IOException
	at hudson.remoting.Channel.close(Channel.java:1499)
	at hudson.remoting.Channel.close(Channel.java:1455)
	at hudson.slaves.SlaveComputer.closeChannel(SlaveComputer.java:884)
	at hudson.slaves.SlaveComputer.access$100(SlaveComputer.java:110)
	at hudson.slaves.SlaveComputer$2.run(SlaveComputer.java:765)
	at jenkins.util.ContextResettingExecutorService$1.run(ContextResettingExecutorService.java:28)
	at jenkins.security.ImpersonatingExecutorService$1.run(ImpersonatingExecutorService.java:68)
	at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
	at java.util.concurrent.FutureTask.run(FutureTask.java:266)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
	at java.lang.Thread.run(Thread.java:748)
Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure
ERROR: apache-beam-jenkins-3 is offline; cannot locate jdk_1.8_latest

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #53

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/53/display/redirect?page=changes>

Changes:

[david.prieto.rivera] Missing contribution

[noreply] [BEAM-13803] Add support for native iterable side inputs to the Go SDK

[noreply] [BEAM-11095] Better error handling for illegal emit functions (#16776)

[noreply] Merge pull request #16613 from Supporting JdbcIO driver in classpath for

[noreply] Merge pull request #15848 from [BEAM-13835] An any-type implementation

[Valentyn Tymofieiev] [BEAM-12920] Assume that bare generators types define simple generators.

[Valentyn Tymofieiev] Add a container for Python 3.9.

[Valentyn Tymofieiev] Allow job submission with Python 3.9 on Dataflow runner

[Valentyn Tymofieiev] Add Python 3.9 test suites. Keep Dataflow V1 suites unchanged for now.

[Valentyn Tymofieiev] Add py3.9 Github actions suites.

[Valentyn Tymofieiev] Py39 Doc updates.

[Valentyn Tymofieiev] [BEAM-9980] Simplify run_validates_container.sh to avoid branching.

[Valentyn Tymofieiev] Update Cython to a new version that has py39 wheels.

[Valentyn Tymofieiev] [BEAM-13845] Fix comparison with potentially incomparable default

[Valentyn Tymofieiev] [BEAM-12920] Assume that bare generators types define simple generators.

[Valentyn Tymofieiev] Mark Python 3.9 as supported version.

[noreply] [release-2.36.0][website] Fix github release notes script, header for

[noreply] Use shell to run python for setupVirtualenv (#16796)

[Daniel Oliveira] [BEAM-13830] Properly shut down Debezium expansion service in IT script.

[noreply] Merge pull request #16659 from [BEAM-13774][Playground] Add user to

[Valentyn Tymofieiev] [BEAM-13868] Remove gsutil dep from hdfs IT test.

[noreply] [BEAM-13776][Playground] (#16731)

[noreply] [BEAM-13867] Drop NaNs returned by nlargest in flight_delays example

[noreply] Announce Python 3.9 in CHANGES.md (#16802)

[Brian Hulette] Moving to 2.38.0-SNAPSHOT on master branch.

[noreply] [BEAM-11095] Better error handling for iter/reiter/multimap (#16794)


------------------------------------------
[...truncated 49.52 KB...]
95d3511f4a11: Pushed
7741f2d71045: Pushed
c617e58f0c50: Pushed
bfbffcc0b02d: Pushed
92715a35625b: Pushed
4bfcea871a23: Pushed
5331ce361341: Layer already exists
babf5ef2e3ab: Layer already exists
613ab28cf833: Layer already exists
bed676ceab7a: Layer already exists
90eacd4783d8: Pushed
6398d5cccd2c: Layer already exists
0b0f2f2f5279: Layer already exists
9bc4e1e02f96: Pushed
e08ee0196508: Pushed
20220210123653: digest: sha256:9ad812c02dd245785430c732fa9769e7eaac5982fbeb695b53b4928449dfd98f size: 4314

> Task :sdks:java:testing:load-tests:run
Feb 10, 2022 12:38:57 PM org.apache.beam.runners.dataflow.options.DataflowPipelineOptions$StagingLocationFactory create
INFO: No stagingLocation provided, falling back to gcpTempLocation
Feb 10, 2022 12:38:57 PM org.apache.beam.runners.dataflow.DataflowRunner fromOptions
INFO: PipelineOptions.filesToStage was not specified. Defaulting to files from the classpath: will stage 204 files. Enable logging at DEBUG level to see which files will be staged.
Feb 10, 2022 12:38:58 PM org.apache.beam.sdk.Pipeline validate
WARNING: The following transforms do not have stable unique names: Window.Into()
Feb 10, 2022 12:38:58 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Executing pipeline on the Dataflow Service, which will have billing implications related to Google Compute Engine usage and other Google Cloud Services.
Feb 10, 2022 12:39:00 PM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Uploading 204 files from PipelineOptions.filesToStage to staging location to prepare for execution.
Feb 10, 2022 12:39:00 PM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Staging files complete: 204 files cached, 0 files newly uploaded in 0 seconds
Feb 10, 2022 12:39:00 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Staging portable pipeline proto to gs://temp-storage-for-perf-tests/loadtests/staging/
Feb 10, 2022 12:39:00 PM org.apache.beam.runners.dataflow.util.PackageUtil tryStagePackage
INFO: Uploading <114077 bytes, hash dba166edc882bcda34ba22289bdebe8befaa746e3bb79572e112f34c2d80cea3> to gs://temp-storage-for-perf-tests/loadtests/staging/pipeline-26Fm7ciCvNo0uiIom96-i--qdG47t5Vy4RLzTC2AzqM.pb
Feb 10, 2022 12:39:02 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read input/DataflowRunner.StreamingUnboundedRead.ReadWithIds as step s1
Feb 10, 2022 12:39:02 PM org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource split
INFO: Split into 20 bundles of sizes: [org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3dd31157, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@31c628e7, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3240b2a4, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@58434b19, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7d3fb0ef, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7dbe2ebf, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4adc663e, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@885e7ff, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@8bd86c8, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4fa9ab6, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2d3ef181, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@a2341c6, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6e4c0d8c, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3e3315d9, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@64db4967, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@74e6094b, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7a485a36, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5cf3157b, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@625dfff3, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@26350ea2]
Feb 10, 2022 12:39:02 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read input/StripIds as step s2
Feb 10, 2022 12:39:02 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect start time metrics (input) as step s3
Feb 10, 2022 12:39:02 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Window.Into()/Window.Assign as step s4
Feb 10, 2022 12:39:02 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read co-input/DataflowRunner.StreamingUnboundedRead.ReadWithIds as step s5
Feb 10, 2022 12:39:02 PM org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource split
INFO: Split into 20 bundles of sizes: [org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@806996, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@78b612c6, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@257e0827, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@22752544, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@21ba2445, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@69d23296, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3c3820bb, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@376c7d7d, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4784efd9, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3fba233d, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@427ae189, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@16a9eb2e, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@76332405, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@187e5235, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@d1d8e1a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5434e40c, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3b48e183, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@514de325, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@30c1da48, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@43a65cd8]
Feb 10, 2022 12:39:02 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read co-input/StripIds as step s6
Feb 10, 2022 12:39:02 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect start time metrics (co-input) as step s7
Feb 10, 2022 12:39:02 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Window.Into()2/Window.Assign as step s8
Feb 10, 2022 12:39:02 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/MakeUnionTable0 as step s9
Feb 10, 2022 12:39:02 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/MakeUnionTable1 as step s10
Feb 10, 2022 12:39:02 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/Flatten as step s11
Feb 10, 2022 12:39:02 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/GBK as step s12
Feb 10, 2022 12:39:02 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/ConstructCoGbkResultFn as step s13
Feb 10, 2022 12:39:02 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Ungroup and reiterate as step s14
Feb 10, 2022 12:39:02 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect total bytes as step s15
Feb 10, 2022 12:39:02 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect end time metrics as step s16
Feb 10, 2022 12:39:02 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Dataflow SDK version: 2.38.0-SNAPSHOT
Feb 10, 2022 12:39:04 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To access the Dataflow monitoring console, please navigate to https://console.cloud.google.com/dataflow/jobs/us-central1/2022-02-10_04_39_02-7275390966568846868?project=apache-beam-testing
Feb 10, 2022 12:39:04 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Submitted job: 2022-02-10_04_39_02-7275390966568846868
Feb 10, 2022 12:39:04 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To cancel the job using the 'gcloud' tool, run:
> gcloud dataflow jobs --project=apache-beam-testing cancel --region=us-central1 2022-02-10_04_39_02-7275390966568846868
Feb 10, 2022 12:39:19 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
WARNING: 2022-02-10T12:39:17.775Z: The workflow name is not a valid Cloud Label. Labels applied to Cloud resources (such as GCE Instances) for monitoring will be labeled with this modified job name: load0tests0java170dataflow0v20streaming0cogbk01-jenkins-02-ci9d. For the best monitoring experience, please name your job with a valid Cloud Label. For details, see: https://cloud.google.com/compute/docs/labeling-resources#restrictions
Feb 10, 2022 12:39:24 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-10T12:39:24.024Z: Worker configuration: e2-standard-2 in us-central1-b.
Feb 10, 2022 12:39:24 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-10T12:39:24.566Z: Expanding SplittableParDo operations into optimizable parts.
Feb 10, 2022 12:39:24 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-10T12:39:24.609Z: Expanding CollectionToSingleton operations into optimizable parts.
Feb 10, 2022 12:39:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-10T12:39:24.685Z: Expanding CoGroupByKey operations into optimizable parts.
Feb 10, 2022 12:39:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-10T12:39:24.790Z: Expanding SplittableProcessKeyed operations into optimizable parts.
Feb 10, 2022 12:39:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-10T12:39:24.823Z: Expanding GroupByKey operations into streaming Read/Write steps
Feb 10, 2022 12:39:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-10T12:39:24.901Z: Lifting ValueCombiningMappingFns into MergeBucketsMappingFns
Feb 10, 2022 12:39:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-10T12:39:25.037Z: Fusing adjacent ParDo, Read, Write, and Flatten operations
Feb 10, 2022 12:39:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-10T12:39:25.077Z: Unzipping flatten CoGroupByKey-Flatten for input CoGroupByKey-MakeUnionTable0-ParMultiDo-ConstructUnionTable-.output
Feb 10, 2022 12:39:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-10T12:39:25.109Z: Fusing unzipped copy of CoGroupByKey/GBK/WriteStream, through flatten CoGroupByKey/Flatten, into producer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable)
Feb 10, 2022 12:39:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-10T12:39:25.148Z: Fusing consumer CoGroupByKey/GBK/WriteStream into CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable)
Feb 10, 2022 12:39:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-10T12:39:25.172Z: Fusing consumer Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read input/Impulse
Feb 10, 2022 12:39:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-10T12:39:25.207Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
Feb 10, 2022 12:39:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-10T12:39:25.229Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
Feb 10, 2022 12:39:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-10T12:39:25.268Z: Fusing consumer Read input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
Feb 10, 2022 12:39:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-10T12:39:25.325Z: Fusing consumer Collect start time metrics (input)/ParMultiDo(TimeMonitor) into Read input/ParDo(StripIds)/ParMultiDo(StripIds)
Feb 10, 2022 12:39:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-10T12:39:25.363Z: Fusing consumer Window.Into()/Window.Assign into Collect start time metrics (input)/ParMultiDo(TimeMonitor)
Feb 10, 2022 12:39:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-10T12:39:25.398Z: Fusing consumer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable) into Window.Into()/Window.Assign
Feb 10, 2022 12:39:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-10T12:39:25.433Z: Fusing consumer Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read co-input/Impulse
Feb 10, 2022 12:39:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-10T12:39:25.465Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
Feb 10, 2022 12:39:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-10T12:39:25.499Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
Feb 10, 2022 12:39:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-10T12:39:25.534Z: Fusing consumer Read co-input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
Feb 10, 2022 12:39:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-10T12:39:25.568Z: Fusing consumer Collect start time metrics (co-input)/ParMultiDo(TimeMonitor) into Read co-input/ParDo(StripIds)/ParMultiDo(StripIds)
Feb 10, 2022 12:39:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-10T12:39:25.600Z: Fusing consumer Window.Into()2/Window.Assign into Collect start time metrics (co-input)/ParMultiDo(TimeMonitor)
Feb 10, 2022 12:39:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-10T12:39:25.638Z: Fusing consumer CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable) into Window.Into()2/Window.Assign
Feb 10, 2022 12:39:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-10T12:39:25.667Z: Fusing consumer CoGroupByKey/GBK/MergeBuckets into CoGroupByKey/GBK/ReadStream
Feb 10, 2022 12:39:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-10T12:39:25.700Z: Fusing consumer CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult) into CoGroupByKey/GBK/MergeBuckets
Feb 10, 2022 12:39:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-10T12:39:25.735Z: Fusing consumer Ungroup and reiterate/ParMultiDo(UngroupAndReiterate) into CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult)
Feb 10, 2022 12:39:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-10T12:39:25.758Z: Fusing consumer Collect total bytes/ParMultiDo(ByteMonitor) into Ungroup and reiterate/ParMultiDo(UngroupAndReiterate)
Feb 10, 2022 12:39:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-10T12:39:25.801Z: Fusing consumer Collect end time metrics/ParMultiDo(TimeMonitor) into Collect total bytes/ParMultiDo(ByteMonitor)
Feb 10, 2022 12:39:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-10T12:39:26.369Z: Starting 5 ****s in us-central1-b...
Feb 10, 2022 12:39:33 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-10T12:39:31.592Z: Your project already contains 100 Dataflow-created metric descriptors, so new user metrics of the form custom.googleapis.com/* will not be created. However, all user metrics are also available in the metric dataflow.googleapis.com/job/user_counter. If you rely on the custom metrics, you can delete old / unused metric descriptors. See https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.list and https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.delete
Feb 10, 2022 12:40:14 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-10T12:40:13.332Z: Autoscaling: Raised the number of ****s to 5 so that the pipeline can catch up with its backlog and keep up with its input rate.
Feb 10, 2022 12:41:16 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-10T12:41:13.675Z: Workers have started successfully.
Feb 10, 2022 12:41:16 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-10T12:41:13.702Z: Workers have started successfully.
Feb 10, 2022 4:00:38 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-10T16:00:36.603Z: Cancel request is committed for workflow job: 2022-02-10_04_39_02-7275390966568846868.
Feb 10, 2022 4:00:38 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-10T16:00:36.695Z: Cleaning up.
Feb 10, 2022 4:00:38 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-10T16:00:36.849Z: Stopping **** pool...
Feb 10, 2022 4:00:38 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-10T16:00:36.959Z: Stopping **** pool...
Feb 10, 2022 4:02:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-10T16:02:52.891Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Feb 10, 2022 4:02:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-10T16:02:52.971Z: Worker pool stopped.
Feb 10, 2022 4:02:59 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-02-10_04_39_02-7275390966568846868 finished with status CANCELLED.
Load test results for test (ID): 37f523ed-881d-4b38-8fdc-b28ad35809ff and timestamp: 2022-02-10T12:38:57.891000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                 11877.094
dataflow_v2_java17_total_bytes_count              1.7028817E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220210123653
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:9ad812c02dd245785430c732fa9769e7eaac5982fbeb695b53b4928449dfd98f
Deleted: sha256:590a9b1989561eb21a60322610b6b7efa157611ed871ff93ba20b093b1dcc161
Deleted: sha256:78aaa23d73716cd2f86ef9d225fc7d8bbe03c076f47b60788a25a53606291778
Deleted: sha256:2d46e8e66b27e127818d859abd437417872f0bf1066c8c270810b0ee1b6bd6ad
Deleted: sha256:84f64073edef8646dea0311cc59efae8b3838fd4d673e1db3120c093e4dbcfaa
Deleted: sha256:d05d03b81a5592c9a266a8fb1de0c7eeb4f7da2f31b81eb4cf0b9c0af381b612
Deleted: sha256:72688c1305690a6a7fe6f9410cc113d2bad3950676675c4519a0d6198d778b51
Deleted: sha256:9c4b86f21f91a30a7320b8c350be5a2f37bd6986617f8c01cf6dc15d2eb6f7bf
Deleted: sha256:2bf00ea8bc516e7130b5b2e0d62f6ee063b271916a2abb8158f8a937d1f4bd85
Deleted: sha256:5388bd28cc36d4a143be60229b3fb503646b640a0a75f8acab5c976079aafa3d
Deleted: sha256:146e0ce63064adeedc8d8a5a110eb32e42cd874f679f3f4d52eba6fbd6a1e9a8
Deleted: sha256:c37fe2fb6fad40e3c8a91d8ab19f168952f6536a2e3a3e2b898250aa32db18cc
Deleted: sha256:b7631bbea9cd3072deb35aed934a35890f9e816fdb41e5f23a6a4df2c0aa8431
Deleted: sha256:20fa2bd9b7c8279498fd3098c2afa1223fd32b227b2d59c69a8b227e7d8dc6e3
Deleted: sha256:84ecea060e8bfc0ed3594c3aeadb6104066da70633997f8dbf69ad3f402b69ea
Deleted: sha256:c3e665845587b2b2c0bb6bfb3199e0601863e5b8d43977a16cbc4b8f7fadaf19
Deleted: sha256:5ba672382051ead59116f9538d1675b8cac51ec82204353b7090963498509510
Deleted: sha256:869c21b852fd40bceaa678cd86cd2924809bc2e3fabdb64a9accb61b72ee59b2
Deleted: sha256:dbcdba2d7363e815574bb5ff300c726872f88e027f7f2c2462550b74c4da160d
Deleted: sha256:66dd1df72475d69d0046dea51f566fc7a0f5d097d5fa86b36aaf4b64c7de80e7
Deleted: sha256:8d3d431337cbc6760d0af229f42003ccc833dac7c13a871bc9c59fec76d589f0
Deleted: sha256:c14f3713d224865943e32ed14e8e728a3a36f5682e6f4b5230968054bbe3871f
Deleted: sha256:e3c4347a27d74461b2195332ddb736bd609a3bb474c0f33486f861a07e0d61bb
Deleted: sha256:bf914c83fc000ec536968dedd8a96daf074a1bf2d3b5990798b000e34e1e78b7
Deleted: sha256:05aeec9e1db395000cb4e0eaa8d761c1ea7642a43e96f23cb868b1e34e076049
Deleted: sha256:bf69a846088b0b80427c47df88001cea36997cdb00111ed821ee0943b8ca88b1
Deleted: sha256:10b8e1c2fb4dbcc193c41295f75674e2292ee3b7868892ee7b95116c02fa6907
Deleted: sha256:0626ded6acc614373905edb8d597e5912cbf67682f59c8c511a8c41e08583ceb
Deleted: sha256:13e77d809f354ea52ddd0fc82f7c197121897b5fc55cf7a5af06d91a7ed69e58
Deleted: sha256:e8e2950d7b2032819bc64f1afe4e6383e7d3b55b8a340da8eeee28a4fca803f5
Deleted: sha256:e979f7620fc2bcb1b609435a5fbc0ae4b65f3e82cf73e7457f3188eb43913bdd
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220210123653]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:9ad812c02dd245785430c732fa9769e7eaac5982fbeb695b53b4928449dfd98f]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220210123653] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:9ad812c02dd245785430c732fa9769e7eaac5982fbeb695b53b4928449dfd98f])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:9ad812c02dd245785430c732fa9769e7eaac5982fbeb695b53b4928449dfd98f
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:9ad812c02dd245785430c732fa9769e7eaac5982fbeb695b53b4928449dfd98f
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:9ad812c02dd245785430c732fa9769e7eaac5982fbeb695b53b4928449dfd98f].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 26m 29s
109 actionable tasks: 73 executed, 32 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/5fzbbbyzamqro

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #52

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/52/display/redirect?page=changes>

Changes:

[noreply] Update README.md

[marco.robles] Update README with latest PreCommit Jobs

[marco.robles] Update Postcommit jobs with latest jobs

[marco.robles] Update Performace job tests in readme

[marco.robles] update load job tests with latest updates

[marco.robles] update other jobs test with latest updates

[marco.robles] mismatch links fix

[marco.robles] update trigger phrase for some postCommit jobs

[marco.robles] correct trigger phrases in readme

[marco.robles] add pending jobs to readme

[noreply] Update README.md

[mmack] [BEAM-13246] Add support for S3 Bucket Key at the object level (AWS Sdk

[Pablo Estrada] Output successful rows from BQ Streaming Inserts

[schapman] BEAM-13439 Type annotation for ptransform_fn

[noreply] [BEAM-13606] Fail bundles with failed BigTable mutations (#16751)

[mmack] [adhoc] Remove remaining usage of Powermock from aws2.

[marco.robles] fix broken links in jobs & remove the invalid ones

[Kyle Weaver] Update Dataflow Python dev container images.

[Kiley Sok] Add java 17 to changes

[noreply] [BEAM-12914] Add missing 3.9 opcodes to type inference. (#16761)

[noreply] [BEAM-13321] Initial BigQueryIO externalization. (#16489)

[noreply] [BEAM-13193] Enable process bundle response elements embedding in Java

[noreply] [BEAM-13830] added a debeziumio_expansion_addr flag to GoSDK (#16780)

[noreply] Apply spotless. (#16783)

[Daniel Oliveira] [BEAM-13732] Switch x-lang BigQueryIO expansion service to GCP one.

[noreply] [BEAM-13858] Fix broken github action on :sdks:go:examples:wordCount

[Kiley Sok] add jira for runner v2

[noreply] [BEAM-13732] Go SDK BigQuery IO wrapper. Initial implementation.

[noreply] [BEAM-13732] Add example for Go BigQuery IO wrapper. (#16786)

[noreply] Update CHANGES.md with Go SDK milestones. (#16787)

[noreply] [BEAM-13193] Allow BeamFnDataOutboundObserver to flush elements.


------------------------------------------
[...truncated 44.85 KB...]
da1552893639: Preparing
2a252c3d0a1d: Preparing
841ed3dbd32a: Preparing
92399d10a8ad: Preparing
8cc024575d60: Preparing
3a45bf8d75fc: Preparing
5331ce361341: Preparing
babf5ef2e3ab: Preparing
613ab28cf833: Preparing
bed676ceab7a: Preparing
6398d5cccd2c: Preparing
0b0f2f2f5279: Preparing
da1552893639: Waiting
babf5ef2e3ab: Waiting
613ab28cf833: Waiting
bed676ceab7a: Waiting
2a252c3d0a1d: Waiting
6398d5cccd2c: Waiting
15ee06e1f7d7: Waiting
0b0f2f2f5279: Waiting
841ed3dbd32a: Waiting
112ca825abfb: Waiting
92399d10a8ad: Waiting
3a45bf8d75fc: Waiting
5331ce361341: Waiting
8cc024575d60: Waiting
23df38581584: Pushed
ee1669083a8c: Pushed
0e85edd95852: Pushed
376adae3d876: Pushed
da1552893639: Pushed
67b4b7ad98b5: Pushed
15ee06e1f7d7: Pushed
841ed3dbd32a: Pushed
8cc024575d60: Pushed
2a252c3d0a1d: Pushed
5331ce361341: Layer already exists
babf5ef2e3ab: Layer already exists
112ca825abfb: Pushed
bed676ceab7a: Layer already exists
613ab28cf833: Layer already exists
6398d5cccd2c: Layer already exists
0b0f2f2f5279: Layer already exists
3a45bf8d75fc: Pushed
92399d10a8ad: Pushed
20220209123414: digest: sha256:d8b4146d1cb4685782b97141ab1f5f215afee4a32cd804296233e60998c23adb size: 4314

> Task :sdks:java:testing:load-tests:run
Feb 09, 2022 12:36:24 PM org.apache.beam.runners.dataflow.options.DataflowPipelineOptions$StagingLocationFactory create
INFO: No stagingLocation provided, falling back to gcpTempLocation
Feb 09, 2022 12:36:24 PM org.apache.beam.runners.dataflow.DataflowRunner fromOptions
INFO: PipelineOptions.filesToStage was not specified. Defaulting to files from the classpath: will stage 204 files. Enable logging at DEBUG level to see which files will be staged.
Feb 09, 2022 12:36:25 PM org.apache.beam.sdk.Pipeline validate
WARNING: The following transforms do not have stable unique names: Window.Into()
Feb 09, 2022 12:36:25 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Executing pipeline on the Dataflow Service, which will have billing implications related to Google Compute Engine usage and other Google Cloud Services.
Feb 09, 2022 12:36:27 PM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Uploading 204 files from PipelineOptions.filesToStage to staging location to prepare for execution.
Feb 09, 2022 12:36:28 PM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Staging files complete: 204 files cached, 0 files newly uploaded in 0 seconds
Feb 09, 2022 12:36:28 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Staging portable pipeline proto to gs://temp-storage-for-perf-tests/loadtests/staging/
Feb 09, 2022 12:36:28 PM org.apache.beam.runners.dataflow.util.PackageUtil tryStagePackage
INFO: Uploading <114077 bytes, hash f3d1f1139e5e054fdba3d9ee7a54ed7fb7be3bb3faf83a7db79251b6d9848e02> to gs://temp-storage-for-perf-tests/loadtests/staging/pipeline-89HxE55eBU_bo9nuelTtf7e-O7P6-Dp9t5JRttmEjgI.pb
Feb 09, 2022 12:36:30 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read input/DataflowRunner.StreamingUnboundedRead.ReadWithIds as step s1
Feb 09, 2022 12:36:30 PM org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource split
INFO: Split into 20 bundles of sizes: [org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3dd31157, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@31c628e7, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3240b2a4, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@58434b19, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7d3fb0ef, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7dbe2ebf, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4adc663e, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@885e7ff, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@8bd86c8, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4fa9ab6, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2d3ef181, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@a2341c6, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6e4c0d8c, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3e3315d9, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@64db4967, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@74e6094b, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7a485a36, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5cf3157b, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@625dfff3, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@26350ea2]
Feb 09, 2022 12:36:30 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read input/StripIds as step s2
Feb 09, 2022 12:36:30 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect start time metrics (input) as step s3
Feb 09, 2022 12:36:30 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Window.Into()/Window.Assign as step s4
Feb 09, 2022 12:36:30 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read co-input/DataflowRunner.StreamingUnboundedRead.ReadWithIds as step s5
Feb 09, 2022 12:36:30 PM org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource split
INFO: Split into 20 bundles of sizes: [org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@806996, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@78b612c6, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@257e0827, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@22752544, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@21ba2445, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@69d23296, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3c3820bb, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@376c7d7d, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4784efd9, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3fba233d, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@427ae189, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@16a9eb2e, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@76332405, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@187e5235, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@d1d8e1a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5434e40c, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3b48e183, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@514de325, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@30c1da48, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@43a65cd8]
Feb 09, 2022 12:36:30 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read co-input/StripIds as step s6
Feb 09, 2022 12:36:30 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect start time metrics (co-input) as step s7
Feb 09, 2022 12:36:30 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Window.Into()2/Window.Assign as step s8
Feb 09, 2022 12:36:30 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/MakeUnionTable0 as step s9
Feb 09, 2022 12:36:30 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/MakeUnionTable1 as step s10
Feb 09, 2022 12:36:30 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/Flatten as step s11
Feb 09, 2022 12:36:30 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/GBK as step s12
Feb 09, 2022 12:36:30 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/ConstructCoGbkResultFn as step s13
Feb 09, 2022 12:36:30 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Ungroup and reiterate as step s14
Feb 09, 2022 12:36:30 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect total bytes as step s15
Feb 09, 2022 12:36:30 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect end time metrics as step s16
Feb 09, 2022 12:36:30 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Dataflow SDK version: 2.37.0-SNAPSHOT
Feb 09, 2022 12:36:31 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To access the Dataflow monitoring console, please navigate to https://console.cloud.google.com/dataflow/jobs/us-central1/2022-02-09_04_36_30-16266957799977566730?project=apache-beam-testing
Feb 09, 2022 12:36:31 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Submitted job: 2022-02-09_04_36_30-16266957799977566730
Feb 09, 2022 12:36:31 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To cancel the job using the 'gcloud' tool, run:
> gcloud dataflow jobs --project=apache-beam-testing cancel --region=us-central1 2022-02-09_04_36_30-16266957799977566730
Feb 09, 2022 12:36:39 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
WARNING: 2022-02-09T12:36:38.855Z: The workflow name is not a valid Cloud Label. Labels applied to Cloud resources (such as GCE Instances) for monitoring will be labeled with this modified job name: load0tests0java170dataflow0v20streaming0cogbk01-jenkins-02-h8r. For the best monitoring experience, please name your job with a valid Cloud Label. For details, see: https://cloud.google.com/compute/docs/labeling-resources#restrictions
Feb 09, 2022 12:36:44 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-09T12:36:43.560Z: Worker configuration: e2-standard-2 in us-central1-b.
Feb 09, 2022 12:36:45 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-09T12:36:44.208Z: Expanding SplittableParDo operations into optimizable parts.
Feb 09, 2022 12:36:45 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-09T12:36:44.236Z: Expanding CollectionToSingleton operations into optimizable parts.
Feb 09, 2022 12:36:45 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-09T12:36:44.314Z: Expanding CoGroupByKey operations into optimizable parts.
Feb 09, 2022 12:36:45 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-09T12:36:44.371Z: Expanding SplittableProcessKeyed operations into optimizable parts.
Feb 09, 2022 12:36:45 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-09T12:36:44.393Z: Expanding GroupByKey operations into streaming Read/Write steps
Feb 09, 2022 12:36:45 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-09T12:36:44.463Z: Lifting ValueCombiningMappingFns into MergeBucketsMappingFns
Feb 09, 2022 12:36:45 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-09T12:36:44.564Z: Fusing adjacent ParDo, Read, Write, and Flatten operations
Feb 09, 2022 12:36:45 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-09T12:36:44.595Z: Unzipping flatten CoGroupByKey-Flatten for input CoGroupByKey-MakeUnionTable0-ParMultiDo-ConstructUnionTable-.output
Feb 09, 2022 12:36:45 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-09T12:36:44.637Z: Fusing unzipped copy of CoGroupByKey/GBK/WriteStream, through flatten CoGroupByKey/Flatten, into producer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable)
Feb 09, 2022 12:36:45 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-09T12:36:44.671Z: Fusing consumer CoGroupByKey/GBK/WriteStream into CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable)
Feb 09, 2022 12:36:45 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-09T12:36:44.703Z: Fusing consumer Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read input/Impulse
Feb 09, 2022 12:36:45 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-09T12:36:44.736Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
Feb 09, 2022 12:36:45 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-09T12:36:44.768Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
Feb 09, 2022 12:36:45 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-09T12:36:44.801Z: Fusing consumer Read input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
Feb 09, 2022 12:36:45 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-09T12:36:44.834Z: Fusing consumer Collect start time metrics (input)/ParMultiDo(TimeMonitor) into Read input/ParDo(StripIds)/ParMultiDo(StripIds)
Feb 09, 2022 12:36:45 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-09T12:36:44.856Z: Fusing consumer Window.Into()/Window.Assign into Collect start time metrics (input)/ParMultiDo(TimeMonitor)
Feb 09, 2022 12:36:45 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-09T12:36:44.877Z: Fusing consumer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable) into Window.Into()/Window.Assign
Feb 09, 2022 12:36:45 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-09T12:36:44.909Z: Fusing consumer Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read co-input/Impulse
Feb 09, 2022 12:36:45 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-09T12:36:44.932Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
Feb 09, 2022 12:36:45 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-09T12:36:44.964Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
Feb 09, 2022 12:36:45 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-09T12:36:44.996Z: Fusing consumer Read co-input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
Feb 09, 2022 12:36:45 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-09T12:36:45.029Z: Fusing consumer Collect start time metrics (co-input)/ParMultiDo(TimeMonitor) into Read co-input/ParDo(StripIds)/ParMultiDo(StripIds)
Feb 09, 2022 12:36:45 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-09T12:36:45.075Z: Fusing consumer Window.Into()2/Window.Assign into Collect start time metrics (co-input)/ParMultiDo(TimeMonitor)
Feb 09, 2022 12:36:45 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-09T12:36:45.101Z: Fusing consumer CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable) into Window.Into()2/Window.Assign
Feb 09, 2022 12:36:45 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-09T12:36:45.125Z: Fusing consumer CoGroupByKey/GBK/MergeBuckets into CoGroupByKey/GBK/ReadStream
Feb 09, 2022 12:36:45 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-09T12:36:45.146Z: Fusing consumer CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult) into CoGroupByKey/GBK/MergeBuckets
Feb 09, 2022 12:36:45 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-09T12:36:45.178Z: Fusing consumer Ungroup and reiterate/ParMultiDo(UngroupAndReiterate) into CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult)
Feb 09, 2022 12:36:45 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-09T12:36:45.235Z: Fusing consumer Collect total bytes/ParMultiDo(ByteMonitor) into Ungroup and reiterate/ParMultiDo(UngroupAndReiterate)
Feb 09, 2022 12:36:45 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-09T12:36:45.274Z: Fusing consumer Collect end time metrics/ParMultiDo(TimeMonitor) into Collect total bytes/ParMultiDo(ByteMonitor)
Feb 09, 2022 12:36:47 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-09T12:36:45.655Z: Starting 5 ****s in us-central1-b...
Feb 09, 2022 12:37:00 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-09T12:36:57.791Z: Your project already contains 100 Dataflow-created metric descriptors, so new user metrics of the form custom.googleapis.com/* will not be created. However, all user metrics are also available in the metric dataflow.googleapis.com/job/user_counter. If you rely on the custom metrics, you can delete old / unused metric descriptors. See https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.list and https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.delete
Feb 09, 2022 12:37:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-09T12:37:26.694Z: Autoscaling: Raised the number of ****s to 5 so that the pipeline can catch up with its backlog and keep up with its input rate.
Feb 09, 2022 12:38:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-09T12:38:26.498Z: Workers have started successfully.
Feb 09, 2022 12:38:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-09T12:38:26.525Z: Workers have started successfully.
Feb 09, 2022 4:00:40 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-09T16:00:39.049Z: Cancel request is committed for workflow job: 2022-02-09_04_36_30-16266957799977566730.
Feb 09, 2022 4:00:40 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-09T16:00:39.132Z: Cleaning up.
Feb 09, 2022 4:00:40 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-09T16:00:39.225Z: Stopping **** pool...
Feb 09, 2022 4:00:40 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-09T16:00:39.293Z: Stopping **** pool...
Feb 09, 2022 4:03:02 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-09T16:03:01.829Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Feb 09, 2022 4:03:02 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-09T16:03:01.888Z: Worker pool stopped.
Feb 09, 2022 4:03:07 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-02-09_04_36_30-16266957799977566730 finished with status CANCELLED.
Load test results for test (ID): d605d13e-dc1f-412e-9b59-2ed956956529 and timestamp: 2022-02-09T12:36:24.977000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                 12097.965
dataflow_v2_java17_total_bytes_count              1.9972801E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220209123414
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:d8b4146d1cb4685782b97141ab1f5f215afee4a32cd804296233e60998c23adb
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220209123414]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:d8b4146d1cb4685782b97141ab1f5f215afee4a32cd804296233e60998c23adb]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220209123414] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:d8b4146d1cb4685782b97141ab1f5f215afee4a32cd804296233e60998c23adb])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:d8b4146d1cb4685782b97141ab1f5f215afee4a32cd804296233e60998c23adb
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:d8b4146d1cb4685782b97141ab1f5f215afee4a32cd804296233e60998c23adb
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:d8b4146d1cb4685782b97141ab1f5f215afee4a32cd804296233e60998c23adb].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 29m 23s
109 actionable tasks: 73 executed, 32 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/3ua6j5qbno2ju

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #51

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/51/display/redirect?page=changes>

Changes:

[Kyle Weaver] [BEAM-12976] Log projection pushdown optimizations.

[benjamin.gonzalez] [BEAM-12572] Change jobs to run as cron jobs

[Ismaël Mejía] [BEAM-13839] Upgrade zstd-jni to version 1.5.2-1

[mmack] [BEAM-13840] Fix usage of legacy rawtypes in AWS modules

[alexander.zhuravlev] [BEAM-13820] Changed color of delete icon in pipeline options dropdown,

[noreply] [BEAM-11971] Revert "Fix timer consistency in direct runner" (#16748)

[noreply] [BEAM-13193] Aggregates fn api outbound data/timers of different

[noreply] [BEAM-13767] Migrate a bundle of grade tasks to use configuration

[noreply] Merge pull request #16653 from [BEAM-12164]: Add integration tests for

[noreply] Merge pull request #16728 from [BEAM-13823] Update docs for SnowflakeIO

[noreply] Merge pull request #16660 from [BEAM-13771][Playground] Send multifile

[noreply] Merge pull request #16646 from [BEAM-13643][Playground] Setup running

[noreply] [BEAM-13015] Add state caching benchmark and move benchmarks to their

[noreply] [BEAM-13419] Check for initialization in dataflow runner (#16765)

[noreply] Merge pull request #16701 from [BEAM-13786] [Playground] [Bugfix] Update

[noreply] Merge pull request #16754 from [BEAM-13838][Playground] Add logs in case

[noreply] [BEAM-13293] consistent naming for expansion service address and flag

[noreply] Merge pull request #16700 from [BEAM-13790][Playground] Change logic of

[noreply] [BEAM-13830] update dependency for debeziumio expansion service (#16743)

[noreply] [BEAM-13761] consistent namings for expansion address in Debezium IO

[noreply] [BEAM-13806] Shutting down SchemaIO expansion services from Go VR

[noreply] [release-2.36.0] Update website/changelog for release 2.36.0 (#16627)

[noreply] [BEAM-13848] Update numpy intersphinx link (#16767)

[noreply] [release-23.6.0] Fix JIRA link for 2.36 blog (#16771)

[noreply] [BEAM-13647] Use role for Go worker binary. (#16729)


------------------------------------------
[...truncated 50.04 KB...]
02b95ed3b258: Preparing
c985abc5c46f: Preparing
c6a22f776f38: Preparing
7cdf99bbf626: Preparing
4178adaf21bb: Preparing
1a1fa842e18f: Preparing
5331ce361341: Preparing
babf5ef2e3ab: Preparing
613ab28cf833: Preparing
bed676ceab7a: Preparing
4c921850ec81: Waiting
6398d5cccd2c: Preparing
02b95ed3b258: Waiting
0b0f2f2f5279: Preparing
4178adaf21bb: Waiting
4a36f4ba74c8: Waiting
7cdf99bbf626: Waiting
1a1fa842e18f: Waiting
c985abc5c46f: Waiting
c6a22f776f38: Waiting
bed676ceab7a: Waiting
5331ce361341: Waiting
0b0f2f2f5279: Waiting
613ab28cf833: Waiting
babf5ef2e3ab: Waiting
6398d5cccd2c: Waiting
7b07699a263f: Pushed
75a152b8416c: Pushed
6f493e8c92b3: Pushed
02b95ed3b258: Pushed
4a36f4ba74c8: Pushed
1df098f15f0d: Pushed
8474b8f39618: Pushed
4c921850ec81: Pushed
c6a22f776f38: Pushed
4178adaf21bb: Pushed
5331ce361341: Layer already exists
babf5ef2e3ab: Layer already exists
c985abc5c46f: Pushed
613ab28cf833: Layer already exists
bed676ceab7a: Layer already exists
6398d5cccd2c: Layer already exists
0b0f2f2f5279: Layer already exists
1a1fa842e18f: Pushed
7cdf99bbf626: Pushed
20220208125141: digest: sha256:61a14815d8656a9e6408b730778f5a9435b9355e2f7e67042b9209ea61f99835 size: 4314

> Task :sdks:java:testing:load-tests:run
Feb 08, 2022 12:54:02 PM org.apache.beam.runners.dataflow.options.DataflowPipelineOptions$StagingLocationFactory create
INFO: No stagingLocation provided, falling back to gcpTempLocation
Feb 08, 2022 12:54:03 PM org.apache.beam.runners.dataflow.DataflowRunner fromOptions
INFO: PipelineOptions.filesToStage was not specified. Defaulting to files from the classpath: will stage 204 files. Enable logging at DEBUG level to see which files will be staged.
Feb 08, 2022 12:54:04 PM org.apache.beam.sdk.Pipeline validate
WARNING: The following transforms do not have stable unique names: Window.Into()
Feb 08, 2022 12:54:04 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Executing pipeline on the Dataflow Service, which will have billing implications related to Google Compute Engine usage and other Google Cloud Services.
Feb 08, 2022 12:54:06 PM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Uploading 204 files from PipelineOptions.filesToStage to staging location to prepare for execution.
Feb 08, 2022 12:54:06 PM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Staging files complete: 204 files cached, 0 files newly uploaded in 0 seconds
Feb 08, 2022 12:54:06 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Staging portable pipeline proto to gs://temp-storage-for-perf-tests/loadtests/staging/
Feb 08, 2022 12:54:06 PM org.apache.beam.runners.dataflow.util.PackageUtil tryStagePackage
INFO: Uploading <114077 bytes, hash 95faaad11d2f9027f09e71327ca72386246a489483dd04ffe4dc4e022b44e899> to gs://temp-storage-for-perf-tests/loadtests/staging/pipeline-lfqq0R0vkCfwnnEyfKcjhiRqSJSD3QT_5NxOAitE6Jk.pb
Feb 08, 2022 12:54:08 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read input/DataflowRunner.StreamingUnboundedRead.ReadWithIds as step s1
Feb 08, 2022 12:54:08 PM org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource split
INFO: Split into 20 bundles of sizes: [org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@31c628e7, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3240b2a4, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@58434b19, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7d3fb0ef, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7dbe2ebf, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4adc663e, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@885e7ff, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@8bd86c8, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4fa9ab6, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2d3ef181, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@a2341c6, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6e4c0d8c, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3e3315d9, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@64db4967, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@74e6094b, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7a485a36, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5cf3157b, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@625dfff3, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@26350ea2, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1e9469b8]
Feb 08, 2022 12:54:08 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read input/StripIds as step s2
Feb 08, 2022 12:54:08 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect start time metrics (input) as step s3
Feb 08, 2022 12:54:08 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Window.Into()/Window.Assign as step s4
Feb 08, 2022 12:54:08 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read co-input/DataflowRunner.StreamingUnboundedRead.ReadWithIds as step s5
Feb 08, 2022 12:54:08 PM org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource split
INFO: Split into 20 bundles of sizes: [org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@78b612c6, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@257e0827, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@22752544, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@21ba2445, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@69d23296, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3c3820bb, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@376c7d7d, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4784efd9, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3fba233d, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@427ae189, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@16a9eb2e, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@76332405, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@187e5235, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@d1d8e1a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5434e40c, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3b48e183, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@514de325, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@30c1da48, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@43a65cd8, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3f1ef9d6]
Feb 08, 2022 12:54:08 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read co-input/StripIds as step s6
Feb 08, 2022 12:54:08 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect start time metrics (co-input) as step s7
Feb 08, 2022 12:54:08 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Window.Into()2/Window.Assign as step s8
Feb 08, 2022 12:54:08 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/MakeUnionTable0 as step s9
Feb 08, 2022 12:54:08 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/MakeUnionTable1 as step s10
Feb 08, 2022 12:54:08 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/Flatten as step s11
Feb 08, 2022 12:54:08 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/GBK as step s12
Feb 08, 2022 12:54:08 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/ConstructCoGbkResultFn as step s13
Feb 08, 2022 12:54:08 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Ungroup and reiterate as step s14
Feb 08, 2022 12:54:08 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect total bytes as step s15
Feb 08, 2022 12:54:08 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect end time metrics as step s16
Feb 08, 2022 12:54:08 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Dataflow SDK version: 2.37.0-SNAPSHOT
Feb 08, 2022 12:54:09 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To access the Dataflow monitoring console, please navigate to https://console.cloud.google.com/dataflow/jobs/us-central1/2022-02-08_04_54_08-9881391196564705185?project=apache-beam-testing
Feb 08, 2022 12:54:09 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Submitted job: 2022-02-08_04_54_08-9881391196564705185
Feb 08, 2022 12:54:09 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To cancel the job using the 'gcloud' tool, run:
> gcloud dataflow jobs --project=apache-beam-testing cancel --region=us-central1 2022-02-08_04_54_08-9881391196564705185
Feb 08, 2022 12:54:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
WARNING: 2022-02-08T12:54:21.155Z: The workflow name is not a valid Cloud Label. Labels applied to Cloud resources (such as GCE Instances) for monitoring will be labeled with this modified job name: load0tests0java170dataflow0v20streaming0cogbk01-jenkins-02-t7m. For the best monitoring experience, please name your job with a valid Cloud Label. For details, see: https://cloud.google.com/compute/docs/labeling-resources#restrictions
Feb 08, 2022 12:54:34 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-08T12:54:30.908Z: Worker configuration: e2-standard-2 in us-central1-b.
Feb 08, 2022 12:54:34 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-08T12:54:34.010Z: Expanding SplittableParDo operations into optimizable parts.
Feb 08, 2022 12:54:34 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-08T12:54:34.048Z: Expanding CollectionToSingleton operations into optimizable parts.
Feb 08, 2022 12:54:34 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-08T12:54:34.113Z: Expanding CoGroupByKey operations into optimizable parts.
Feb 08, 2022 12:54:34 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-08T12:54:34.181Z: Expanding SplittableProcessKeyed operations into optimizable parts.
Feb 08, 2022 12:54:34 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-08T12:54:34.219Z: Expanding GroupByKey operations into streaming Read/Write steps
Feb 08, 2022 12:54:34 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-08T12:54:34.272Z: Lifting ValueCombiningMappingFns into MergeBucketsMappingFns
Feb 08, 2022 12:54:34 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-08T12:54:34.371Z: Fusing adjacent ParDo, Read, Write, and Flatten operations
Feb 08, 2022 12:54:34 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-08T12:54:34.404Z: Unzipping flatten CoGroupByKey-Flatten for input CoGroupByKey-MakeUnionTable0-ParMultiDo-ConstructUnionTable-.output
Feb 08, 2022 12:54:34 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-08T12:54:34.449Z: Fusing unzipped copy of CoGroupByKey/GBK/WriteStream, through flatten CoGroupByKey/Flatten, into producer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable)
Feb 08, 2022 12:54:34 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-08T12:54:34.483Z: Fusing consumer CoGroupByKey/GBK/WriteStream into CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable)
Feb 08, 2022 12:54:34 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-08T12:54:34.536Z: Fusing consumer Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read input/Impulse
Feb 08, 2022 12:54:34 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-08T12:54:34.578Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
Feb 08, 2022 12:54:34 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-08T12:54:34.642Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
Feb 08, 2022 12:54:34 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-08T12:54:34.687Z: Fusing consumer Read input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
Feb 08, 2022 12:54:36 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-08T12:54:34.733Z: Fusing consumer Collect start time metrics (input)/ParMultiDo(TimeMonitor) into Read input/ParDo(StripIds)/ParMultiDo(StripIds)
Feb 08, 2022 12:54:36 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-08T12:54:34.852Z: Fusing consumer Window.Into()/Window.Assign into Collect start time metrics (input)/ParMultiDo(TimeMonitor)
Feb 08, 2022 12:54:36 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-08T12:54:34.927Z: Fusing consumer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable) into Window.Into()/Window.Assign
Feb 08, 2022 12:54:36 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-08T12:54:34.978Z: Fusing consumer Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read co-input/Impulse
Feb 08, 2022 12:54:36 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-08T12:54:35.023Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
Feb 08, 2022 12:54:36 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-08T12:54:35.061Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
Feb 08, 2022 12:54:36 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-08T12:54:35.097Z: Fusing consumer Read co-input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
Feb 08, 2022 12:54:36 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-08T12:54:35.129Z: Fusing consumer Collect start time metrics (co-input)/ParMultiDo(TimeMonitor) into Read co-input/ParDo(StripIds)/ParMultiDo(StripIds)
Feb 08, 2022 12:54:36 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-08T12:54:35.160Z: Fusing consumer Window.Into()2/Window.Assign into Collect start time metrics (co-input)/ParMultiDo(TimeMonitor)
Feb 08, 2022 12:54:36 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-08T12:54:35.183Z: Fusing consumer CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable) into Window.Into()2/Window.Assign
Feb 08, 2022 12:54:36 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-08T12:54:35.215Z: Fusing consumer CoGroupByKey/GBK/MergeBuckets into CoGroupByKey/GBK/ReadStream
Feb 08, 2022 12:54:36 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-08T12:54:35.249Z: Fusing consumer CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult) into CoGroupByKey/GBK/MergeBuckets
Feb 08, 2022 12:54:36 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-08T12:54:35.285Z: Fusing consumer Ungroup and reiterate/ParMultiDo(UngroupAndReiterate) into CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult)
Feb 08, 2022 12:54:36 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-08T12:54:35.312Z: Fusing consumer Collect total bytes/ParMultiDo(ByteMonitor) into Ungroup and reiterate/ParMultiDo(UngroupAndReiterate)
Feb 08, 2022 12:54:36 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-08T12:54:35.342Z: Fusing consumer Collect end time metrics/ParMultiDo(TimeMonitor) into Collect total bytes/ParMultiDo(ByteMonitor)
Feb 08, 2022 12:54:36 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-08T12:54:35.727Z: Starting 5 ****s in us-central1-b...
Feb 08, 2022 12:54:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-08T12:54:42.583Z: Your project already contains 100 Dataflow-created metric descriptors, so new user metrics of the form custom.googleapis.com/* will not be created. However, all user metrics are also available in the metric dataflow.googleapis.com/job/user_counter. If you rely on the custom metrics, you can delete old / unused metric descriptors. See https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.list and https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.delete
Feb 08, 2022 12:55:21 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-08T12:55:19.915Z: Autoscaling: Raised the number of ****s to 5 so that the pipeline can catch up with its backlog and keep up with its input rate.
Feb 08, 2022 12:56:19 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-08T12:56:18.270Z: Workers have started successfully.
Feb 08, 2022 12:56:19 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-08T12:56:18.306Z: Workers have started successfully.
Feb 08, 2022 4:01:03 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-08T16:01:01.742Z: Cancel request is committed for workflow job: 2022-02-08_04_54_08-9881391196564705185.
Feb 08, 2022 4:01:03 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-08T16:01:01.821Z: Cleaning up.
Feb 08, 2022 4:01:03 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-08T16:01:01.891Z: Stopping **** pool...
Feb 08, 2022 4:01:03 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-08T16:01:01.944Z: Stopping **** pool...
Feb 08, 2022 4:03:29 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-08T16:03:29.881Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Feb 08, 2022 4:03:29 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-08T16:03:29.916Z: Worker pool stopped.
Feb 08, 2022 4:03:36 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-02-08_04_54_08-9881391196564705185 finished with status CANCELLED.
Load test results for test (ID): 0dc17070-b2c8-4d8f-b39f-c0d00fa7cfd6 and timestamp: 2022-02-08T12:54:03.654000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                 11031.014
dataflow_v2_java17_total_bytes_count             1.92883302E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220208125141
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:61a14815d8656a9e6408b730778f5a9435b9355e2f7e67042b9209ea61f99835
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220208125141]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:61a14815d8656a9e6408b730778f5a9435b9355e2f7e67042b9209ea61f99835]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220208125141] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:61a14815d8656a9e6408b730778f5a9435b9355e2f7e67042b9209ea61f99835])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:61a14815d8656a9e6408b730778f5a9435b9355e2f7e67042b9209ea61f99835
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:61a14815d8656a9e6408b730778f5a9435b9355e2f7e67042b9209ea61f99835
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:61a14815d8656a9e6408b730778f5a9435b9355e2f7e67042b9209ea61f99835].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 12m 16s
109 actionable tasks: 74 executed, 31 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/xxtva5hy5aa4m

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 - Build # 50 - Aborted!

Posted by Apache Jenkins Server <je...@builds.apache.org>.
beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 - Build # 50 - Aborted:

Check console output at https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/50/ to view the results.

Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #49

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/49/display/redirect?page=changes>

Changes:

[mmack] [BEAM-13663] Remove unused duplicate option for AWS client configuration

[mmack] [BEAM-13203] Deprecate SnsIO.writeAsync for AWS Sdk v2 due to risk of

[noreply] [BEAM-13828] Fix stale bot (#16734)

[noreply] Merge pull request #16364 from [BEAM-13182]  Add diagrams to backend

[noreply] [BEAM-13811] Fix save_main_session arg in tests examples (#16709)

[Kiley Sok] Update beam-master version

[noreply] [BEAM-13015] Calculate exception for closing BeamFnDataInboundObserver2

[noreply] Minor doc tweaks for validating vendoring. (#16747)

[noreply] [BEAM-13686] OOM while logging a large pipeline even when logging level

[noreply] [BEAM-13629] Update URL artifact type for Dataflow Go (#16490)

[noreply] [BEAM-13832] Add automated expansion service start-up to JDBCio (#16739)

[noreply] [BEAM-13831] Add automated expansion service infra into Debezium Read()

[noreply] [BEAM-13821] Add automated expansion service start-up to KafkaIO

[noreply] [BEAM-13799] Created a Dataproc cluster manager for Interactive Beam

[noreply] Merge pull request #16727: [BEAM-11971] remove unsafe Concurrent data


------------------------------------------
[...truncated 49.74 KB...]
fac1035c71d4: Pushed
fea2e1306969: Pushed
f55bef4cec0b: Pushed
385a9d1c164e: Pushed
722cda4f3905: Pushed
745d80168905: Pushed
5331ce361341: Layer already exists
babf5ef2e3ab: Layer already exists
613ab28cf833: Layer already exists
bed676ceab7a: Layer already exists
7289c7a56219: Pushed
0b0f2f2f5279: Layer already exists
6398d5cccd2c: Layer already exists
1a6c1e7784fa: Pushed
c806171f29c7: Pushed
20220205123919: digest: sha256:3b53c0719895fade70b978f2eaf629d36e960b6656ba420c33f79fcdace49160 size: 4314

> Task :sdks:java:testing:load-tests:run
Feb 05, 2022 12:41:14 PM org.apache.beam.runners.dataflow.options.DataflowPipelineOptions$StagingLocationFactory create
INFO: No stagingLocation provided, falling back to gcpTempLocation
Feb 05, 2022 12:41:15 PM org.apache.beam.runners.dataflow.DataflowRunner fromOptions
INFO: PipelineOptions.filesToStage was not specified. Defaulting to files from the classpath: will stage 204 files. Enable logging at DEBUG level to see which files will be staged.
Feb 05, 2022 12:41:16 PM org.apache.beam.sdk.Pipeline validate
WARNING: The following transforms do not have stable unique names: Window.Into()
Feb 05, 2022 12:41:16 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Executing pipeline on the Dataflow Service, which will have billing implications related to Google Compute Engine usage and other Google Cloud Services.
Feb 05, 2022 12:41:18 PM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Uploading 204 files from PipelineOptions.filesToStage to staging location to prepare for execution.
Feb 05, 2022 12:41:18 PM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Staging files complete: 204 files cached, 0 files newly uploaded in 0 seconds
Feb 05, 2022 12:41:18 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Staging portable pipeline proto to gs://temp-storage-for-perf-tests/loadtests/staging/
Feb 05, 2022 12:41:18 PM org.apache.beam.runners.dataflow.util.PackageUtil tryStagePackage
INFO: Uploading <114077 bytes, hash 918a4ac6b597bbf72253273237a0fbf086526d2a911343a78fdfef86c315656b> to gs://temp-storage-for-perf-tests/loadtests/staging/pipeline-kYpKxrWXu_ciUycyN6D78IZSbSqRE0Onj9_vhsMVZWs.pb
Feb 05, 2022 12:41:20 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read input/DataflowRunner.StreamingUnboundedRead.ReadWithIds as step s1
Feb 05, 2022 12:41:20 PM org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource split
INFO: Split into 20 bundles of sizes: [org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@31c628e7, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3240b2a4, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@58434b19, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7d3fb0ef, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7dbe2ebf, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4adc663e, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@885e7ff, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@8bd86c8, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4fa9ab6, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2d3ef181, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@a2341c6, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6e4c0d8c, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3e3315d9, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@64db4967, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@74e6094b, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7a485a36, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5cf3157b, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@625dfff3, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@26350ea2, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1e9469b8]
Feb 05, 2022 12:41:20 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read input/StripIds as step s2
Feb 05, 2022 12:41:20 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect start time metrics (input) as step s3
Feb 05, 2022 12:41:20 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Window.Into()/Window.Assign as step s4
Feb 05, 2022 12:41:20 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read co-input/DataflowRunner.StreamingUnboundedRead.ReadWithIds as step s5
Feb 05, 2022 12:41:20 PM org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource split
INFO: Split into 20 bundles of sizes: [org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@78b612c6, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@257e0827, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@22752544, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@21ba2445, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@69d23296, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3c3820bb, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@376c7d7d, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4784efd9, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3fba233d, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@427ae189, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@16a9eb2e, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@76332405, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@187e5235, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@d1d8e1a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5434e40c, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3b48e183, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@514de325, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@30c1da48, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@43a65cd8, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3f1ef9d6]
Feb 05, 2022 12:41:20 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read co-input/StripIds as step s6
Feb 05, 2022 12:41:20 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect start time metrics (co-input) as step s7
Feb 05, 2022 12:41:20 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Window.Into()2/Window.Assign as step s8
Feb 05, 2022 12:41:20 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/MakeUnionTable0 as step s9
Feb 05, 2022 12:41:20 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/MakeUnionTable1 as step s10
Feb 05, 2022 12:41:20 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/Flatten as step s11
Feb 05, 2022 12:41:20 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/GBK as step s12
Feb 05, 2022 12:41:20 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/ConstructCoGbkResultFn as step s13
Feb 05, 2022 12:41:20 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Ungroup and reiterate as step s14
Feb 05, 2022 12:41:20 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect total bytes as step s15
Feb 05, 2022 12:41:20 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect end time metrics as step s16
Feb 05, 2022 12:41:20 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Dataflow SDK version: 2.37.0-SNAPSHOT
Feb 05, 2022 12:41:27 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To access the Dataflow monitoring console, please navigate to https://console.cloud.google.com/dataflow/jobs/us-central1/2022-02-05_04_41_20-13420762200883215816?project=apache-beam-testing
Feb 05, 2022 12:41:27 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Submitted job: 2022-02-05_04_41_20-13420762200883215816
Feb 05, 2022 12:41:27 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To cancel the job using the 'gcloud' tool, run:
> gcloud dataflow jobs --project=apache-beam-testing cancel --region=us-central1 2022-02-05_04_41_20-13420762200883215816
Feb 05, 2022 12:41:35 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
WARNING: 2022-02-05T12:41:34.610Z: The workflow name is not a valid Cloud Label. Labels applied to Cloud resources (such as GCE Instances) for monitoring will be labeled with this modified job name: load0tests0java170dataflow0v20streaming0cogbk01-jenkins-02-du6k. For the best monitoring experience, please name your job with a valid Cloud Label. For details, see: https://cloud.google.com/compute/docs/labeling-resources#restrictions
Feb 05, 2022 12:41:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-05T12:41:39.714Z: Worker configuration: e2-standard-2 in us-central1-b.
Feb 05, 2022 12:41:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-05T12:41:40.351Z: Expanding SplittableParDo operations into optimizable parts.
Feb 05, 2022 12:41:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-05T12:41:40.384Z: Expanding CollectionToSingleton operations into optimizable parts.
Feb 05, 2022 12:41:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-05T12:41:40.461Z: Expanding CoGroupByKey operations into optimizable parts.
Feb 05, 2022 12:41:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-05T12:41:40.526Z: Expanding SplittableProcessKeyed operations into optimizable parts.
Feb 05, 2022 12:41:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-05T12:41:40.552Z: Expanding GroupByKey operations into streaming Read/Write steps
Feb 05, 2022 12:41:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-05T12:41:40.605Z: Lifting ValueCombiningMappingFns into MergeBucketsMappingFns
Feb 05, 2022 12:41:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-05T12:41:40.701Z: Fusing adjacent ParDo, Read, Write, and Flatten operations
Feb 05, 2022 12:41:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-05T12:41:40.756Z: Unzipping flatten CoGroupByKey-Flatten for input CoGroupByKey-MakeUnionTable0-ParMultiDo-ConstructUnionTable-.output
Feb 05, 2022 12:41:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-05T12:41:40.785Z: Fusing unzipped copy of CoGroupByKey/GBK/WriteStream, through flatten CoGroupByKey/Flatten, into producer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable)
Feb 05, 2022 12:41:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-05T12:41:40.807Z: Fusing consumer CoGroupByKey/GBK/WriteStream into CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable)
Feb 05, 2022 12:41:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-05T12:41:40.842Z: Fusing consumer Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read input/Impulse
Feb 05, 2022 12:41:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-05T12:41:40.877Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
Feb 05, 2022 12:41:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-05T12:41:40.909Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
Feb 05, 2022 12:41:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-05T12:41:40.933Z: Fusing consumer Read input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
Feb 05, 2022 12:41:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-05T12:41:40.959Z: Fusing consumer Collect start time metrics (input)/ParMultiDo(TimeMonitor) into Read input/ParDo(StripIds)/ParMultiDo(StripIds)
Feb 05, 2022 12:41:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-05T12:41:40.980Z: Fusing consumer Window.Into()/Window.Assign into Collect start time metrics (input)/ParMultiDo(TimeMonitor)
Feb 05, 2022 12:41:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-05T12:41:41.015Z: Fusing consumer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable) into Window.Into()/Window.Assign
Feb 05, 2022 12:41:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-05T12:41:41.062Z: Fusing consumer Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read co-input/Impulse
Feb 05, 2022 12:41:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-05T12:41:41.086Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
Feb 05, 2022 12:41:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-05T12:41:41.120Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
Feb 05, 2022 12:41:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-05T12:41:41.155Z: Fusing consumer Read co-input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
Feb 05, 2022 12:41:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-05T12:41:41.187Z: Fusing consumer Collect start time metrics (co-input)/ParMultiDo(TimeMonitor) into Read co-input/ParDo(StripIds)/ParMultiDo(StripIds)
Feb 05, 2022 12:41:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-05T12:41:41.223Z: Fusing consumer Window.Into()2/Window.Assign into Collect start time metrics (co-input)/ParMultiDo(TimeMonitor)
Feb 05, 2022 12:41:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-05T12:41:41.247Z: Fusing consumer CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable) into Window.Into()2/Window.Assign
Feb 05, 2022 12:41:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-05T12:41:41.288Z: Fusing consumer CoGroupByKey/GBK/MergeBuckets into CoGroupByKey/GBK/ReadStream
Feb 05, 2022 12:41:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-05T12:41:41.317Z: Fusing consumer CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult) into CoGroupByKey/GBK/MergeBuckets
Feb 05, 2022 12:41:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-05T12:41:41.351Z: Fusing consumer Ungroup and reiterate/ParMultiDo(UngroupAndReiterate) into CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult)
Feb 05, 2022 12:41:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-05T12:41:41.414Z: Fusing consumer Collect total bytes/ParMultiDo(ByteMonitor) into Ungroup and reiterate/ParMultiDo(UngroupAndReiterate)
Feb 05, 2022 12:41:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-05T12:41:41.449Z: Fusing consumer Collect end time metrics/ParMultiDo(TimeMonitor) into Collect total bytes/ParMultiDo(ByteMonitor)
Feb 05, 2022 12:41:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-05T12:41:41.819Z: Starting 5 ****s in us-central1-b...
Feb 05, 2022 12:41:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-05T12:41:54.010Z: Your project already contains 100 Dataflow-created metric descriptors, so new user metrics of the form custom.googleapis.com/* will not be created. However, all user metrics are also available in the metric dataflow.googleapis.com/job/user_counter. If you rely on the custom metrics, you can delete old / unused metric descriptors. See https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.list and https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.delete
Feb 05, 2022 12:42:29 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-05T12:42:28.125Z: Autoscaling: Raised the number of ****s to 5 so that the pipeline can catch up with its backlog and keep up with its input rate.
Feb 05, 2022 12:43:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-05T12:43:27.581Z: Workers have started successfully.
Feb 05, 2022 12:43:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-05T12:43:27.618Z: Workers have started successfully.
Feb 05, 2022 4:00:47 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-05T16:00:45.282Z: Cancel request is committed for workflow job: 2022-02-05_04_41_20-13420762200883215816.
Feb 05, 2022 4:00:47 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-05T16:00:45.381Z: Cleaning up.
Feb 05, 2022 4:00:47 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-05T16:00:45.457Z: Stopping **** pool...
Feb 05, 2022 4:00:47 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-05T16:00:45.519Z: Stopping **** pool...
Feb 05, 2022 4:03:09 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-05T16:03:07.938Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Feb 05, 2022 4:03:09 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-05T16:03:07.972Z: Worker pool stopped.
Feb 05, 2022 4:03:14 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-02-05_04_41_20-13420762200883215816 finished with status CANCELLED.
Load test results for test (ID): 170b76de-fe2b-4bd1-8876-8eb39a0fc703 and timestamp: 2022-02-05T12:41:15.646000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                 11812.461
dataflow_v2_java17_total_bytes_count             2.57368923E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220205123919
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:3b53c0719895fade70b978f2eaf629d36e960b6656ba420c33f79fcdace49160
Deleted: sha256:917e4adc205e3f9e57d208b0b45a0a1d8f5e629ae8dea1c4aa8d45e56daea084
Deleted: sha256:0f50db4240a4dca07fd1ede54632b156a66e0dda1725885ec4b2360efafc4f0a
Deleted: sha256:9bec2edf1263a10d059a7dbd85aada0ac14bfc4a6c478c4490bca785fc85ef93
Deleted: sha256:b8a0a0354755b5ae7ba52fbc5062a71de62a3b7a1b874f13b35214a9aaac9328
Deleted: sha256:73d96b87b73d43215752a1c6e7a0d7cf42ea09b1fea1617446fd5a2af3b9e8d0
Deleted: sha256:217fd70066dfeb3b1d2af29e5ac9e0236134d7596fd7db7e93ba8441456887fc
Deleted: sha256:9943e42a493beab04d65e862eea238a343bdb929836b926c0b40f79e80ff9cb3
Deleted: sha256:7d02fd5c40cc6561857cbb41f64f7f2db52c257f82e126fd1b99c73c4115fefe
Deleted: sha256:d939f2d1299c5cfb6a1ac6c47c5567fc8b16ae11ce9b20b8080e8627cd53a0f3
Deleted: sha256:47d30df79568fe348dd48491dbe9e5ce1ec81faf6d590394377313ed18839d63
Deleted: sha256:27101bd19fb3fcc17c6b3b7522be89306b8028f23051b57c310f96539f5cc2fd
Deleted: sha256:eaaf7b7ca3cc4f1414e80910263e75e3679bf4865ac93e0654294d8b1d086d18
Deleted: sha256:9767184e21fb8adbe58af1b96dc0b44ff4a3627af2302baa05eb20b6033f28d3
Deleted: sha256:993b984f738e844bda4efdb02a1bd26ee6afd093a70fbbd187ff23f7f26523a6
Deleted: sha256:105f9613e734fff35dda1e06e28260de74e2424def75aefa10a8fbb2eda9ca80
Deleted: sha256:78d779611e3b234ca41fa4554be06c0ace38854c5e7123276ded2789089d405b
Deleted: sha256:705fb4ad2aa76818b2dc48bb7eca35930d3eb84da3add6c890070b9a71ecea95
Deleted: sha256:761cce925f4f749c1f8c266e2113d2e655f4b363b7449a8ccfe183f5e9d644d8
Deleted: sha256:65012e554972590e3f9305538a34b664065fb269f8e3803102846cd8705410e5
Deleted: sha256:a95f28b19e171f91f22746c3c668387b877ab128d662ddd007899dfe3824cc19
Deleted: sha256:f5d85d5ec17139110ae9c3ae7ea335c260592b55d5a418e286fc19376d62bc16
Deleted: sha256:2cad9613969d4af93ec404304bdd92b8955cdc4ed15e2766bd3bf4c0888c6c9a
Deleted: sha256:ec18c6fcad1998c1e37dca1bdf5364961fd13035922012e4dc8696cd89a74732
Deleted: sha256:f7a970d3810f7dd18dcf211ceab768394070f2cbf72b9361442b52481f2d3a81
Deleted: sha256:a653ab3deb1f0671396fadee936314df504e2d11a87c068c2c074069eb00736a
Deleted: sha256:bb5ec5ddbf14057c9baff897571abdd3e001b7fba43400b86228fd1b0a4110f0
Deleted: sha256:a45920a7e38feb86d1be31ede7e953f0afcc33a8610470a1160cee3183e531a7
Deleted: sha256:92aa6c45114be59c92c16e2489a9fec11a40cb64aca1ed6caa0da33afad3e2f4
Deleted: sha256:f1bb9413ce93027f3f02b0d5e396702ec3775689b6f07e821a0a08cc49112a7a
Deleted: sha256:b4168c1cd493324baf3184acda9211772ea6b140fc537f9388a5dd5751986f76
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220205123919]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:3b53c0719895fade70b978f2eaf629d36e960b6656ba420c33f79fcdace49160]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220205123919] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:3b53c0719895fade70b978f2eaf629d36e960b6656ba420c33f79fcdace49160])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:3b53c0719895fade70b978f2eaf629d36e960b6656ba420c33f79fcdace49160
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:3b53c0719895fade70b978f2eaf629d36e960b6656ba420c33f79fcdace49160
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:3b53c0719895fade70b978f2eaf629d36e960b6656ba420c33f79fcdace49160].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 24m 16s
109 actionable tasks: 72 executed, 33 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/hqs4pe3yfu6ti

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #48

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/48/display/redirect?page=changes>

Changes:

[Kiley Sok] Allow Java 17 to be used in SDK

[Kiley Sok] add testing support

[Kiley Sok] Add more testing support for java 17

[Kiley Sok] workaround for jamm

[Kiley Sok] Add Jenkins test for Java 17

[Kiley Sok] Fix jvm hex and skip errorprone

[Kiley Sok] Fix display data for anonymous classes

[Kiley Sok] fix jpms tests

[Kiley Sok] skip zetasql

[aydar.zaynutdinov] [BEAM-13737][Playground]

[Robert Bradshaw] Make num-stages counter into an internal counter.

[Kiley Sok] spotless

[Kiley Sok] spotless

[Kenneth Knowles] [BEAM-13768] Fix NullPointerException in BigQueryStorageSourceBase

[career] [BEAM-13734] Support cache directories that use GCS buckets

[Kiley Sok] Fix trigger

[noreply] Merge pull request #16655 from [BEAM-12164]: Add retry protection to

[noreply] Merge pull request #16586 from [BEAM-13731] FhirIO: Add support for

[Kiley Sok] skip checker framework

[Kiley Sok] fix app name

[noreply] [BEAM-13011] Adds a link to Multi-language Pipelines Tips wiki page

[Kiley Sok] remove duplicate property check

[noreply] [BEAM-12572] Run python examples on multiple runners (#16154)

[noreply] [BEAM-13574] Large Wordcount (#16455)

[Robert Bradshaw] Avoid packaging avro in the java harness jar.

[noreply] [BEAM-13293] Refactor JDBC IO Go Wrapper (#16686)

[noreply] Edit license script for Java, add manual licenses for xz (#16692)

[mmack] [BEAM-13563] Restructure Kinesis Source for Aws 2 internally to prepare

[noreply] [BEAM-4665] Allow joining a running dataflow pipeline without throwing

[noreply] [BEAM-13801] Add standard coder tests for state backed iterable.

[noreply] [BEAM-13430]  Fix provided configuration by removing extendsFrom for

[noreply] [BEAM-12830] Print clearer go version fail message (#16693)

[Jan Lukavský] Add reference to Books to Learning Resources in website

[noreply] Use ThreadLocal for DESERIALIZATION_CONTEXT (#16680)

[noreply] Minor: Add apt update after adding deadsnakes repository in dev env

[noreply] [BEAM-13807] Regenerate container images to get TF 2.8.0 (#16707)

[noreply] [BEAM-13399, BEAM-13683] Eagerly materialize artifacts for automated

[noreply] Merge pull request #16617 from [BEAM-13743] [Playground] Add context

[noreply] Merge pull request #16618 from [BEAM-13744] [Playground] Add context

[noreply] Merge pull request #16698 from [BEAM-13802][Playground] [Bugfix] Clean

[noreply] [BEAM-13293][BEAM-13806] Pipe a SchemaIO flag through Go integration

[noreply] [BEAM-13605] Modify groupby.apply implementation in preparation for

[noreply] Merge pull request #16436 from [BEAM-1330] - DatastoreIO Writes should

[Heejong Lee] [BEAM-13813] Add support for URL artifact to extractStagingToPath

[avilovpavel6] Remove Python SQL Test example from catalog

[relax] Fix timer consistency in direct runner

[noreply] [BEAM-13757] adds pane observation in DoFn (#16629)

[Jan Lukavský] Change links to Books from Amazon to Publisher

[noreply] [BEAM-13605] Add support for pandas 1.4.0 (#16590)

[noreply] [BEAM-13761] adds Debezium IO wrapper for Go SDK (#16642)

[noreply] [BEAM-13024] Unify PipelineOptions behavior (#16719)

[noreply] Update sdks/go/pkg/beam/artifact/materialize_test.go

[noreply] Merge pull request #16605 from [BEAM-13634][Playground] Create a

[noreply] Merge pull request #16593 from [BEAM-13725][Playground] Add graph to the

[noreply] Merge pull request #16699 from [BEAM-13789][Playground] Change logic of

[alexander.chermenin] Fixed CSS for Case study page


------------------------------------------
[...truncated 49.55 KB...]
7e7f8e1aea17: Waiting
f1d9484793b9: Waiting
6ef79a002f74: Waiting
51accf6fb8a3: Waiting
ead9462da41f: Waiting
5230ace0e5a0: Pushed
72f6d30be55b: Pushed
709710a5288c: Pushed
b98da1010db8: Pushed
f1d9484793b9: Pushed
6ef79a002f74: Pushed
23bae4c11d26: Pushed
9464bb870ecf: Pushed
3f61a8147dc0: Pushed
ead9462da41f: Pushed
5331ce361341: Layer already exists
613ab28cf833: Layer already exists
51accf6fb8a3: Pushed
babf5ef2e3ab: Layer already exists
bed676ceab7a: Layer already exists
6398d5cccd2c: Layer already exists
0b0f2f2f5279: Layer already exists
4198ef9d5b37: Pushed
7e7f8e1aea17: Pushed
20220204123818: digest: sha256:0035bdbcee488f123ccaa7ee9d663dfccf8e289224ecbc1989320d0e222f2efa size: 4314

> Task :sdks:java:testing:load-tests:run
Feb 04, 2022 12:40:13 PM org.apache.beam.runners.dataflow.options.DataflowPipelineOptions$StagingLocationFactory create
INFO: No stagingLocation provided, falling back to gcpTempLocation
Feb 04, 2022 12:40:14 PM org.apache.beam.runners.dataflow.DataflowRunner fromOptions
INFO: PipelineOptions.filesToStage was not specified. Defaulting to files from the classpath: will stage 204 files. Enable logging at DEBUG level to see which files will be staged.
Feb 04, 2022 12:40:15 PM org.apache.beam.sdk.Pipeline validate
WARNING: The following transforms do not have stable unique names: Window.Into()
Feb 04, 2022 12:40:15 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Executing pipeline on the Dataflow Service, which will have billing implications related to Google Compute Engine usage and other Google Cloud Services.
Feb 04, 2022 12:40:17 PM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Uploading 204 files from PipelineOptions.filesToStage to staging location to prepare for execution.
Feb 04, 2022 12:40:17 PM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Staging files complete: 204 files cached, 0 files newly uploaded in 0 seconds
Feb 04, 2022 12:40:17 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Staging portable pipeline proto to gs://temp-storage-for-perf-tests/loadtests/staging/
Feb 04, 2022 12:40:17 PM org.apache.beam.runners.dataflow.util.PackageUtil tryStagePackage
INFO: Uploading <114077 bytes, hash 71d872be9d1fccb7c05410323f32683a16416e4c7a32647f462899bdb2cadb46> to gs://temp-storage-for-perf-tests/loadtests/staging/pipeline-cdhyvp0fzLfAVBAyPzJoOhZBbkx6MmR_RiiZvbLK20Y.pb
Feb 04, 2022 12:40:19 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read input/DataflowRunner.StreamingUnboundedRead.ReadWithIds as step s1
Feb 04, 2022 12:40:19 PM org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource split
INFO: Split into 20 bundles of sizes: [org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1d805aa1, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@176f7f3b, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@30ca0779, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@58740366, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@47be0f9b, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2bc426f0, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4bd51d3e, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@33425811, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4b74b35, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@e4e1ef5, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6d11ceef, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4cb2918c, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@72e295cc, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@c2584d3, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6fa0450e, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@37468787, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@51ec2856, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@714f3da4, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1caa9eb6, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1f53481b]
Feb 04, 2022 12:40:19 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read input/StripIds as step s2
Feb 04, 2022 12:40:19 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect start time metrics (input) as step s3
Feb 04, 2022 12:40:19 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Window.Into()/Window.Assign as step s4
Feb 04, 2022 12:40:19 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read co-input/DataflowRunner.StreamingUnboundedRead.ReadWithIds as step s5
Feb 04, 2022 12:40:19 PM org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource split
INFO: Split into 20 bundles of sizes: [org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@56399b9e, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@34b9eb03, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@43fda8d9, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@49d831c2, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@53a7a60c, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@28bdbe88, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6a87026, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@ef60710, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@600f5704, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2503ec73, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@606f81b5, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5e1fc42f, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@44b21f9f, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7ee8130e, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6296474f, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4288d98e, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@169268a7, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@285c6918, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@78a0ff63, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7c601d50]
Feb 04, 2022 12:40:19 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read co-input/StripIds as step s6
Feb 04, 2022 12:40:19 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect start time metrics (co-input) as step s7
Feb 04, 2022 12:40:19 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Window.Into()2/Window.Assign as step s8
Feb 04, 2022 12:40:19 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/MakeUnionTable0 as step s9
Feb 04, 2022 12:40:19 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/MakeUnionTable1 as step s10
Feb 04, 2022 12:40:19 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/Flatten as step s11
Feb 04, 2022 12:40:19 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/GBK as step s12
Feb 04, 2022 12:40:19 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/ConstructCoGbkResultFn as step s13
Feb 04, 2022 12:40:19 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Ungroup and reiterate as step s14
Feb 04, 2022 12:40:19 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect total bytes as step s15
Feb 04, 2022 12:40:19 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect end time metrics as step s16
Feb 04, 2022 12:40:19 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Dataflow SDK version: 2.37.0-SNAPSHOT
Feb 04, 2022 12:40:20 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To access the Dataflow monitoring console, please navigate to https://console.cloud.google.com/dataflow/jobs/us-central1/2022-02-04_04_40_19-12414635713987359991?project=apache-beam-testing
Feb 04, 2022 12:40:20 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Submitted job: 2022-02-04_04_40_19-12414635713987359991
Feb 04, 2022 12:40:20 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To cancel the job using the 'gcloud' tool, run:
> gcloud dataflow jobs --project=apache-beam-testing cancel --region=us-central1 2022-02-04_04_40_19-12414635713987359991
Feb 04, 2022 12:40:28 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
WARNING: 2022-02-04T12:40:27.312Z: The workflow name is not a valid Cloud Label. Labels applied to Cloud resources (such as GCE Instances) for monitoring will be labeled with this modified job name: load0tests0java170dataflow0v20streaming0cogbk01-jenkins-02-rgno. For the best monitoring experience, please name your job with a valid Cloud Label. For details, see: https://cloud.google.com/compute/docs/labeling-resources#restrictions
Feb 04, 2022 12:40:32 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-04T12:40:32.878Z: Worker configuration: e2-standard-2 in us-central1-b.
Feb 04, 2022 12:40:35 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-04T12:40:33.588Z: Expanding SplittableParDo operations into optimizable parts.
Feb 04, 2022 12:40:35 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-04T12:40:33.616Z: Expanding CollectionToSingleton operations into optimizable parts.
Feb 04, 2022 12:40:35 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-04T12:40:33.684Z: Expanding CoGroupByKey operations into optimizable parts.
Feb 04, 2022 12:40:35 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-04T12:40:33.755Z: Expanding SplittableProcessKeyed operations into optimizable parts.
Feb 04, 2022 12:40:35 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-04T12:40:33.782Z: Expanding GroupByKey operations into streaming Read/Write steps
Feb 04, 2022 12:40:35 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-04T12:40:33.858Z: Lifting ValueCombiningMappingFns into MergeBucketsMappingFns
Feb 04, 2022 12:40:35 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-04T12:40:33.943Z: Fusing adjacent ParDo, Read, Write, and Flatten operations
Feb 04, 2022 12:40:35 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-04T12:40:33.972Z: Unzipping flatten CoGroupByKey-Flatten for input CoGroupByKey-MakeUnionTable0-ParMultiDo-ConstructUnionTable-.output
Feb 04, 2022 12:40:35 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-04T12:40:34.004Z: Fusing unzipped copy of CoGroupByKey/GBK/WriteStream, through flatten CoGroupByKey/Flatten, into producer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable)
Feb 04, 2022 12:40:35 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-04T12:40:34.038Z: Fusing consumer CoGroupByKey/GBK/WriteStream into CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable)
Feb 04, 2022 12:40:35 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-04T12:40:34.060Z: Fusing consumer Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read input/Impulse
Feb 04, 2022 12:40:35 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-04T12:40:34.086Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
Feb 04, 2022 12:40:35 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-04T12:40:34.119Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
Feb 04, 2022 12:40:35 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-04T12:40:34.143Z: Fusing consumer Read input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
Feb 04, 2022 12:40:35 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-04T12:40:34.165Z: Fusing consumer Collect start time metrics (input)/ParMultiDo(TimeMonitor) into Read input/ParDo(StripIds)/ParMultiDo(StripIds)
Feb 04, 2022 12:40:35 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-04T12:40:34.195Z: Fusing consumer Window.Into()/Window.Assign into Collect start time metrics (input)/ParMultiDo(TimeMonitor)
Feb 04, 2022 12:40:35 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-04T12:40:34.234Z: Fusing consumer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable) into Window.Into()/Window.Assign
Feb 04, 2022 12:40:35 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-04T12:40:34.271Z: Fusing consumer Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read co-input/Impulse
Feb 04, 2022 12:40:35 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-04T12:40:34.305Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
Feb 04, 2022 12:40:35 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-04T12:40:34.338Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
Feb 04, 2022 12:40:35 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-04T12:40:34.369Z: Fusing consumer Read co-input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
Feb 04, 2022 12:40:35 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-04T12:40:34.404Z: Fusing consumer Collect start time metrics (co-input)/ParMultiDo(TimeMonitor) into Read co-input/ParDo(StripIds)/ParMultiDo(StripIds)
Feb 04, 2022 12:40:35 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-04T12:40:34.441Z: Fusing consumer Window.Into()2/Window.Assign into Collect start time metrics (co-input)/ParMultiDo(TimeMonitor)
Feb 04, 2022 12:40:35 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-04T12:40:34.472Z: Fusing consumer CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable) into Window.Into()2/Window.Assign
Feb 04, 2022 12:40:35 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-04T12:40:34.508Z: Fusing consumer CoGroupByKey/GBK/MergeBuckets into CoGroupByKey/GBK/ReadStream
Feb 04, 2022 12:40:35 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-04T12:40:34.538Z: Fusing consumer CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult) into CoGroupByKey/GBK/MergeBuckets
Feb 04, 2022 12:40:35 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-04T12:40:34.575Z: Fusing consumer Ungroup and reiterate/ParMultiDo(UngroupAndReiterate) into CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult)
Feb 04, 2022 12:40:35 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-04T12:40:34.598Z: Fusing consumer Collect total bytes/ParMultiDo(ByteMonitor) into Ungroup and reiterate/ParMultiDo(UngroupAndReiterate)
Feb 04, 2022 12:40:35 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-04T12:40:34.628Z: Fusing consumer Collect end time metrics/ParMultiDo(TimeMonitor) into Collect total bytes/ParMultiDo(ByteMonitor)
Feb 04, 2022 12:40:35 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-04T12:40:34.963Z: Starting 5 ****s in us-central1-b...
Feb 04, 2022 12:41:01 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-04T12:40:59.239Z: Your project already contains 100 Dataflow-created metric descriptors, so new user metrics of the form custom.googleapis.com/* will not be created. However, all user metrics are also available in the metric dataflow.googleapis.com/job/user_counter. If you rely on the custom metrics, you can delete old / unused metric descriptors. See https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.list and https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.delete
Feb 04, 2022 12:41:20 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-04T12:41:19.161Z: Autoscaling: Raised the number of ****s to 5 so that the pipeline can catch up with its backlog and keep up with its input rate.
Feb 04, 2022 12:42:24 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-04T12:42:22.968Z: Workers have started successfully.
Feb 04, 2022 12:42:24 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-04T12:42:23.004Z: Workers have started successfully.
Feb 04, 2022 4:00:38 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-04T16:00:36.891Z: Cancel request is committed for workflow job: 2022-02-04_04_40_19-12414635713987359991.
Feb 04, 2022 4:00:38 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-04T16:00:36.971Z: Cleaning up.
Feb 04, 2022 4:00:38 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-04T16:00:37.031Z: Stopping **** pool...
Feb 04, 2022 4:00:38 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-04T16:00:37.111Z: Stopping **** pool...
Feb 04, 2022 4:02:57 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-04T16:02:55.683Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Feb 04, 2022 4:02:57 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-02-04T16:02:55.727Z: Worker pool stopped.
Feb 04, 2022 4:03:02 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-02-04_04_40_19-12414635713987359991 finished with status CANCELLED.
Load test results for test (ID): f17bfaec-837f-4299-a2a7-496dad19df6f and timestamp: 2022-02-04T12:40:14.672000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                 11871.421
dataflow_v2_java17_total_bytes_count             2.71294816E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220204123818
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:0035bdbcee488f123ccaa7ee9d663dfccf8e289224ecbc1989320d0e222f2efa
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220204123818]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:0035bdbcee488f123ccaa7ee9d663dfccf8e289224ecbc1989320d0e222f2efa]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220204123818] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:0035bdbcee488f123ccaa7ee9d663dfccf8e289224ecbc1989320d0e222f2efa])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:3ee374fec2e9f88c91a044b0d0ff9cdeae20b4dd8e7b92ec1240d584c61368cc
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:3ee374fec2e9f88c91a044b0d0ff9cdeae20b4dd8e7b92ec1240d584c61368cc
ERROR: (gcloud.container.images.delete) Not found: response: {'docker-distribution-api-version': 'registry/2.0', 'content-type': 'application/json', 'date': 'Fri, 04 Feb 2022 16:03:08 GMT', 'server': 'Docker Registry', 'cache-control': 'private', 'x-xss-protection': '0', 'x-frame-options': 'SAMEORIGIN', 'transfer-encoding': 'chunked', 'status': '404', 'content-length': '168', '-content-encoding': 'gzip'}
Failed to compute blob liveness for manifest: 'sha256:3ee374fec2e9f88c91a044b0d0ff9cdeae20b4dd8e7b92ec1240d584c61368cc': None

> Task :runners:google-cloud-dataflow-java:cleanUpDockerJavaImages FAILED

FAILURE: Build completed with 2 failures.

1: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.
==============================================================================

2: Task failed with an exception.
-----------
* Where:
Build file '<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/runners/google-cloud-dataflow-java/build.gradle'> line: 297

* What went wrong:
Execution failed for task ':runners:google-cloud-dataflow-java:cleanUpDockerJavaImages'.
> Process 'command './scripts/cleanup_untagged_gcr_images.sh'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.
==============================================================================

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 25m 3s
109 actionable tasks: 72 executed, 33 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/mxyemdpqerg2s

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #47

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/47/display/redirect?page=changes>

Changes:

[daria.malkova] Support SCIO SDK via sbt projects

[samuelw] [BEAM-11648] Share thread pool across RetryManager instances.

[Pablo Estrada] Exclude per-key order tests on Twister2 runner

[Heejong Lee] Fix Java SDK container image name for load-tests and nexmark

[daria.malkova] Change executable name fo go tests

[avilovpavel6] Fix java test

[noreply] [BEAM-13769] Skip test_main_session_not_staged_when_using_cloudpickle

[noreply] [BEAM-6744] Support implicitly setting project id in Go Dataflow runner

[noreply] Merge pull request #16493 from [BEAM-13632][Playground] Save catalog

[noreply] Exclude jul-to-slf4j from Spark runner in quickstart POM templates

[noreply] [BEAM-11936] Enable a few errorprone checks that were broken by pinned

[noreply] [BEAM-13780] Add CONTRIBUTING.md pointing to main guide (#16666)

[noreply] [BEAM-13777] Accept cache capacity as input parameter instead of default

[noreply] [BEAM-13051][A] Enable pylint warnings

[noreply] [BEAM-13779] Fix pr labeling (#16665)

[noreply] Merge pull request #16581 from [BEAM-12164]: Add

[noreply] Fix labeler trigger (#16674)

[noreply] [BEAM-13781] Exclude grpc-netty-shaded from gax-grpc's dependency

[noreply] [BEAM-13051] Fixed pylint warnings : raising-non-exception (E0710),

[noreply] [BEAM-13740] Correctly install go before running tests (#16673)

[noreply] [BEAM-12830] Update local Docker env Go version. (#16670)

[noreply] [BEAM-13051][B] Enable pylint warnings

[noreply] [BEAM-13430] Revert Spark libraries in Spark runner to provided (#16675)

[noreply] [BEAM-12240] Add Java 17 support (#16568)

[noreply] [BEAM-13760] Add random component to default python dataflow job name


------------------------------------------
[...truncated 95.11 KB...]
    return self.do_open(http.client.HTTPConnection, req)
  File "/usr/lib/python3.8/urllib/request.py", line 1357, in do_open
    raise URLError(err)
urllib.error.URLError: <urlopen error [Errno 101] Network is unreachable>
ERROR:root:Invalid url for jFormatString-3.0.0: http://www.gnu.org/licenses/lgpl.html. Retrying...
Traceback (most recent call last):
  File "/usr/lib/python3.8/urllib/request.py", line 1354, in do_open
    h.request(req.get_method(), req.selector, req.data, headers,
  File "/usr/lib/python3.8/http/client.py", line 1256, in request
    self._send_request(method, url, body, headers, encode_chunked)
  File "/usr/lib/python3.8/http/client.py", line 1302, in _send_request
    self.endheaders(body, encode_chunked=encode_chunked)
  File "/usr/lib/python3.8/http/client.py", line 1251, in endheaders
    self._send_output(message_body, encode_chunked=encode_chunked)
  File "/usr/lib/python3.8/http/client.py", line 1011, in _send_output
    self.send(msg)
  File "/usr/lib/python3.8/http/client.py", line 951, in send
    self.connect()
  File "/usr/lib/python3.8/http/client.py", line 1418, in connect
    super().connect()
  File "/usr/lib/python3.8/http/client.py", line 922, in connect
    self.sock = self._create_connection(
  File "/usr/lib/python3.8/socket.py", line 808, in create_connection
    raise err
  File "/usr/lib/python3.8/socket.py", line 796, in create_connection
    sock.connect(sa)
OSError: [Errno 101] Network is unreachable

During handling of the above exception, another exception occurred:

Traceback (most recent call last):
  File "<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/license_scripts/pull_licenses_java.py",> line 53, in pull_from_url
    url_read = urlopen(url)
  File "/usr/lib/python3.8/urllib/request.py", line 222, in urlopen
    return opener.open(url, data, timeout)
  File "/usr/lib/python3.8/urllib/request.py", line 525, in open
    response = self._open(req, data)
  File "/usr/lib/python3.8/urllib/request.py", line 542, in _open
    result = self._call_chain(self.handle_open, protocol, protocol +
  File "/usr/lib/python3.8/urllib/request.py", line 502, in _call_chain
    result = func(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 1397, in https_open
    return self.do_open(http.client.HTTPSConnection, req,
  File "/usr/lib/python3.8/urllib/request.py", line 1357, in do_open
    raise URLError(err)
urllib.error.URLError: <urlopen error [Errno 101] Network is unreachable>
ERROR:root:Invalid url for spotbugs-annotations-4.0.6: https://www.gnu.org/licenses/old-licenses/lgpl-2.1.en.html. Retrying...
Traceback (most recent call last):
  File "/usr/lib/python3.8/urllib/request.py", line 1354, in do_open
    h.request(req.get_method(), req.selector, req.data, headers,
  File "/usr/lib/python3.8/http/client.py", line 1256, in request
    self._send_request(method, url, body, headers, encode_chunked)
  File "/usr/lib/python3.8/http/client.py", line 1302, in _send_request
    self.endheaders(body, encode_chunked=encode_chunked)
  File "/usr/lib/python3.8/http/client.py", line 1251, in endheaders
    self._send_output(message_body, encode_chunked=encode_chunked)
  File "/usr/lib/python3.8/http/client.py", line 1011, in _send_output
    self.send(msg)
  File "/usr/lib/python3.8/http/client.py", line 951, in send
    self.connect()
  File "/usr/lib/python3.8/http/client.py", line 922, in connect
    self.sock = self._create_connection(
  File "/usr/lib/python3.8/socket.py", line 808, in create_connection
    raise err
  File "/usr/lib/python3.8/socket.py", line 796, in create_connection
    sock.connect(sa)
OSError: [Errno 101] Network is unreachable

During handling of the above exception, another exception occurred:

Traceback (most recent call last):
  File "<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/license_scripts/pull_licenses_java.py",> line 53, in pull_from_url
    url_read = urlopen(url)
  File "/usr/lib/python3.8/urllib/request.py", line 222, in urlopen
    return opener.open(url, data, timeout)
  File "/usr/lib/python3.8/urllib/request.py", line 525, in open
    response = self._open(req, data)
  File "/usr/lib/python3.8/urllib/request.py", line 542, in _open
    result = self._call_chain(self.handle_open, protocol, protocol +
  File "/usr/lib/python3.8/urllib/request.py", line 502, in _call_chain
    result = func(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 1383, in http_open
    return self.do_open(http.client.HTTPConnection, req)
  File "/usr/lib/python3.8/urllib/request.py", line 1357, in do_open
    raise URLError(err)
urllib.error.URLError: <urlopen error [Errno 101] Network is unreachable>
ERROR:root:Invalid url for checkstyle-8.23: http://www.gnu.org/licenses/old-licenses/lgpl-2.1.txt. Retrying...
Traceback (most recent call last):
  File "/usr/lib/python3.8/urllib/request.py", line 1354, in do_open
    h.request(req.get_method(), req.selector, req.data, headers,
  File "/usr/lib/python3.8/http/client.py", line 1256, in request
    self._send_request(method, url, body, headers, encode_chunked)
  File "/usr/lib/python3.8/http/client.py", line 1302, in _send_request
    self.endheaders(body, encode_chunked=encode_chunked)
  File "/usr/lib/python3.8/http/client.py", line 1251, in endheaders
    self._send_output(message_body, encode_chunked=encode_chunked)
  File "/usr/lib/python3.8/http/client.py", line 1011, in _send_output
    self.send(msg)
  File "/usr/lib/python3.8/http/client.py", line 951, in send
    self.connect()
  File "/usr/lib/python3.8/http/client.py", line 922, in connect
    self.sock = self._create_connection(
  File "/usr/lib/python3.8/socket.py", line 808, in create_connection
    raise err
  File "/usr/lib/python3.8/socket.py", line 796, in create_connection
    sock.connect(sa)
OSError: [Errno 101] Network is unreachable

During handling of the above exception, another exception occurred:

Traceback (most recent call last):
  File "<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/license_scripts/pull_licenses_java.py",> line 53, in pull_from_url
    url_read = urlopen(url)
  File "/usr/lib/python3.8/urllib/request.py", line 222, in urlopen
    return opener.open(url, data, timeout)
  File "/usr/lib/python3.8/urllib/request.py", line 525, in open
    response = self._open(req, data)
  File "/usr/lib/python3.8/urllib/request.py", line 542, in _open
    result = self._call_chain(self.handle_open, protocol, protocol +
  File "/usr/lib/python3.8/urllib/request.py", line 502, in _call_chain
    result = func(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 1383, in http_open
    return self.do_open(http.client.HTTPConnection, req)
  File "/usr/lib/python3.8/urllib/request.py", line 1357, in do_open
    raise URLError(err)
urllib.error.URLError: <urlopen error [Errno 101] Network is unreachable>
ERROR:root:Invalid url for checkstyle-8.23: http://www.gnu.org/licenses/old-licenses/lgpl-2.1.txt after 9 retries.
Traceback (most recent call last):
  File "/usr/lib/python3.8/urllib/request.py", line 1354, in do_open
    h.request(req.get_method(), req.selector, req.data, headers,
  File "/usr/lib/python3.8/http/client.py", line 1256, in request
    self._send_request(method, url, body, headers, encode_chunked)
  File "/usr/lib/python3.8/http/client.py", line 1302, in _send_request
    self.endheaders(body, encode_chunked=encode_chunked)
  File "/usr/lib/python3.8/http/client.py", line 1251, in endheaders
    self._send_output(message_body, encode_chunked=encode_chunked)
  File "/usr/lib/python3.8/http/client.py", line 1011, in _send_output
    self.send(msg)
  File "/usr/lib/python3.8/http/client.py", line 951, in send
    self.connect()
  File "/usr/lib/python3.8/http/client.py", line 922, in connect
    self.sock = self._create_connection(
  File "/usr/lib/python3.8/socket.py", line 808, in create_connection
    raise err
  File "/usr/lib/python3.8/socket.py", line 796, in create_connection
    sock.connect(sa)
OSError: [Errno 101] Network is unreachable

During handling of the above exception, another exception occurred:

Traceback (most recent call last):
  File "<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/license_scripts/pull_licenses_java.py",> line 53, in pull_from_url
    url_read = urlopen(url)
  File "/usr/lib/python3.8/urllib/request.py", line 222, in urlopen
    return opener.open(url, data, timeout)
  File "/usr/lib/python3.8/urllib/request.py", line 525, in open
    response = self._open(req, data)
  File "/usr/lib/python3.8/urllib/request.py", line 542, in _open
    result = self._call_chain(self.handle_open, protocol, protocol +
  File "/usr/lib/python3.8/urllib/request.py", line 502, in _call_chain
    result = func(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 1383, in http_open
    return self.do_open(http.client.HTTPConnection, req)
  File "/usr/lib/python3.8/urllib/request.py", line 1357, in do_open
    raise URLError(err)
urllib.error.URLError: <urlopen error [Errno 101] Network is unreachable>
ERROR:root:Invalid url for jFormatString-3.0.0: http://www.gnu.org/licenses/lgpl.html after 9 retries.
Traceback (most recent call last):
  File "/usr/lib/python3.8/urllib/request.py", line 1354, in do_open
    h.request(req.get_method(), req.selector, req.data, headers,
  File "/usr/lib/python3.8/http/client.py", line 1256, in request
    self._send_request(method, url, body, headers, encode_chunked)
  File "/usr/lib/python3.8/http/client.py", line 1302, in _send_request
    self.endheaders(body, encode_chunked=encode_chunked)
  File "/usr/lib/python3.8/http/client.py", line 1251, in endheaders
    self._send_output(message_body, encode_chunked=encode_chunked)
  File "/usr/lib/python3.8/http/client.py", line 1011, in _send_output
    self.send(msg)
  File "/usr/lib/python3.8/http/client.py", line 951, in send
    self.connect()
  File "/usr/lib/python3.8/http/client.py", line 1418, in connect
    super().connect()
  File "/usr/lib/python3.8/http/client.py", line 922, in connect
    self.sock = self._create_connection(
  File "/usr/lib/python3.8/socket.py", line 808, in create_connection
    raise err
  File "/usr/lib/python3.8/socket.py", line 796, in create_connection
    sock.connect(sa)
OSError: [Errno 101] Network is unreachable

During handling of the above exception, another exception occurred:

Traceback (most recent call last):
  File "<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/license_scripts/pull_licenses_java.py",> line 53, in pull_from_url
    url_read = urlopen(url)
  File "/usr/lib/python3.8/urllib/request.py", line 222, in urlopen
    return opener.open(url, data, timeout)
  File "/usr/lib/python3.8/urllib/request.py", line 525, in open
    response = self._open(req, data)
  File "/usr/lib/python3.8/urllib/request.py", line 542, in _open
    result = self._call_chain(self.handle_open, protocol, protocol +
  File "/usr/lib/python3.8/urllib/request.py", line 502, in _call_chain
    result = func(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 1397, in https_open
    return self.do_open(http.client.HTTPSConnection, req,
  File "/usr/lib/python3.8/urllib/request.py", line 1357, in do_open
    raise URLError(err)
urllib.error.URLError: <urlopen error [Errno 101] Network is unreachable>
ERROR:root:Invalid url for spotbugs-annotations-4.0.6: https://www.gnu.org/licenses/old-licenses/lgpl-2.1.en.html after 9 retries.
ERROR:root:['checkstyle-8.23', 'jFormatString-3.0.0', 'spotbugs-annotations-4.0.6']
ERROR:root:**************************************** Licenses were not able to be pulled automatically for some dependencies. Please search source code of the dependencies on the internet and add "license" and "notice" (if available) field to <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/license_scripts/dep_urls_java.yaml> for each missing license. Dependency List: [checkstyle-8.23,jFormatString-3.0.0,spotbugs-annotations-4.0.6]
INFO:root:pull_licenses_java.py failed. It took 1214.168949 seconds with 16 threads.
Traceback (most recent call last):
  File "<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/license_scripts/pull_licenses_java.py",> line 314, in <module>
    raise RuntimeError('{n} error(s) occurred.'.format(n=len(error_msg)),
RuntimeError: ('1 error(s) occurred.', ['**************************************** Licenses were not able to be pulled automatically for some dependencies. Please search source code of the dependencies on the internet and add "license" and "notice" (if available) field to <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/license_scripts/dep_urls_java.yaml> for each missing license. Dependency List: [checkstyle-8.23,jFormatString-3.0.0,spotbugs-annotations-4.0.6]'])

> Task :sdks:java:container:pullLicenses FAILED

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:container:pullLicenses'.
> Process 'command './license_scripts/license_script.sh'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 20m 52s
103 actionable tasks: 66 executed, 33 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/cddygnbizjyk6

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #46

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/46/display/redirect?page=changes>

Changes:

[mrudary] Generalize S3FileSystem to support multiple URI schemes.


------------------------------------------
Started by timer
Running as SYSTEM
[EnvInject] - Loading node environment variables.
Building remotely on apache-beam-jenkins-6 (beam) in workspace <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/>
The recommended git tool is: NONE
No credentials specified
Wiping out workspace first.
Cloning the remote Git repository
Cloning repository https://github.com/apache/beam.git
 > git init <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src> # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git --version # timeout=10
 > git --version # 'git version 2.25.1'
 > git fetch --tags --force --progress -- https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git fetch --tags --force --progress -- https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* +refs/pull/${ghprbPullId}/*:refs/remotes/origin/pr/${ghprbPullId}/* # timeout=10
 > git rev-parse origin/master^{commit} # timeout=10
Checking out Revision e638c1183407999bca3e8e4987119c0d9158d00d (origin/master)
 > git config core.sparsecheckout # timeout=10
 > git checkout -f e638c1183407999bca3e8e4987119c0d9158d00d # timeout=10
Commit message: "Merge pull request #16607: [BEAM-13245] Generalize S3FileSystem to support multiple URI schemes."
 > git rev-list --no-walk 6c9c208197d3d74b1c3643d22716ad3b00213506 # timeout=10
No emails were triggered.
[EnvInject] - Executing scripts and injecting environment variables after the SCM step.
[EnvInject] - Injecting as environment variables the properties content 
SPARK_LOCAL_IP=127.0.0.1
SETUPTOOLS_USE_DISTUTILS=stdlib

[EnvInject] - Variables injected successfully.
[beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17] $ /bin/bash -xe /tmp/jenkins7654227862286655908.sh
+ echo '*** Load test: CoGBK 2GB 100  byte records - single key ***'
*** Load test: CoGBK 2GB 100  byte records - single key ***
[Gradle] - Launching build.
[src] $ <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/gradlew> -PloadTest.mainClass=org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest -Prunner=:runners:google-cloud-dataflow-java '-PloadTest.args=--project=apache-beam-testing --region=us-central1 --appName=load_tests_Java17_Dataflow_V2_streaming_CoGBK_1 --tempLocation=gs://temp-storage-for-perf-tests/loadtests --influxMeasurement=java_streaming_cogbk_1 --influxTags={"runnerVersion":"v2","jdk":"java17"} --publishToInfluxDB=true --sourceOptions={"numRecords":20000000,"keySizeBytes":10,"valueSizeBytes":90,"numHotKeys":1} --coSourceOptions={"numRecords":2000000,"keySizeBytes":10,"valueSizeBytes":90,"numHotKeys":1000} --iterations=1 --numWorkers=5 --autoscalingAlgorithm=NONE --streaming=true --influxDatabase=beam_test_metrics --influxHost=http://10.128.0.96:8086 --inputWindowDurationSec=1200 --coInputWindowDurationSec=1200 --runner=DataflowRunner' -Prunner.version=V2 -PcompileAndRunTestsWithJava17 -Pjava17Home=/usr/lib/jvm/java-17-openjdk-amd64 --continue --max-****s=12 -Dorg.gradle.jvmargs=-Xms2g -Dorg.gradle.jvmargs=-Xmx4g -Dorg.gradle.vfs.watch=false -Pdocker-pull-licenses :sdks:java:testing:load-tests:run
Starting a Gradle Daemon, 3 busy Daemons could not be reused, use --status for details
Configuration on demand is an incubating feature.
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy FROM-CACHE
> Task :buildSrc:pluginDescriptors
> Task :buildSrc:processResources
> Task :buildSrc:classes
> Task :buildSrc:jar
> Task :buildSrc:assemble
> Task :buildSrc:spotlessGroovy FROM-CACHE
> Task :buildSrc:spotlessGroovyCheck UP-TO-DATE
> Task :buildSrc:spotlessGroovyGradle FROM-CACHE
> Task :buildSrc:spotlessGroovyGradleCheck UP-TO-DATE
> Task :buildSrc:spotlessCheck UP-TO-DATE
> Task :buildSrc:pluginUnderTestMetadata
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validatePlugins FROM-CACHE
> Task :buildSrc:check UP-TO-DATE
> Task :buildSrc:build

> Configure project :sdks:go:test
System Go installation: /snap/bin/go is go version go1.16.13 linux/amd64; Preparing to use /home/jenkins/go/bin/go1.16.12
go1.16.12: already downloaded in /home/jenkins/sdk/go1.16.12
GOCMD=/home/jenkins/go/bin/go1.16.12

FAILURE: Build failed with an exception.

* Where:
Build file '<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/testing/load-tests/build.gradle'> line: 113

* What went wrong:
A problem occurred evaluating project ':sdks:java:testing:load-tests'.
> Could not get unknown property 'dockerImageName' for project ':runners:google-cloud-dataflow-java' of type org.gradle.api.Project.

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

BUILD FAILED in 14s
10 actionable tasks: 4 executed, 4 from cache, 2 up-to-date

Publishing build scan...
https://gradle.com/s/3c2fdlxfuiyru

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #45

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/45/display/redirect>

Changes:


------------------------------------------
Started by timer
Running as SYSTEM
[EnvInject] - Loading node environment variables.
Building remotely on apache-beam-jenkins-6 (beam) in workspace <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/>
The recommended git tool is: NONE
No credentials specified
Wiping out workspace first.
Cloning the remote Git repository
Cloning repository https://github.com/apache/beam.git
 > git init <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src> # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git --version # timeout=10
 > git --version # 'git version 2.25.1'
 > git fetch --tags --force --progress -- https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git fetch --tags --force --progress -- https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* +refs/pull/${ghprbPullId}/*:refs/remotes/origin/pr/${ghprbPullId}/* # timeout=10
 > git rev-parse origin/master^{commit} # timeout=10
Checking out Revision 6c9c208197d3d74b1c3643d22716ad3b00213506 (origin/master)
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 6c9c208197d3d74b1c3643d22716ad3b00213506 # timeout=10
Commit message: "[BEAM-10206] Add Go Vet to Github Actions (#16612)"
 > git rev-list --no-walk 6c9c208197d3d74b1c3643d22716ad3b00213506 # timeout=10
No emails were triggered.
[EnvInject] - Executing scripts and injecting environment variables after the SCM step.
[EnvInject] - Injecting as environment variables the properties content 
SPARK_LOCAL_IP=127.0.0.1
SETUPTOOLS_USE_DISTUTILS=stdlib

[EnvInject] - Variables injected successfully.
[beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17] $ /bin/bash -xe /tmp/jenkins2934564344934422900.sh
+ echo '*** Load test: CoGBK 2GB 100  byte records - single key ***'
*** Load test: CoGBK 2GB 100  byte records - single key ***
[Gradle] - Launching build.
[src] $ <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/gradlew> -PloadTest.mainClass=org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest -Prunner=:runners:google-cloud-dataflow-java '-PloadTest.args=--project=apache-beam-testing --region=us-central1 --appName=load_tests_Java17_Dataflow_V2_streaming_CoGBK_1 --tempLocation=gs://temp-storage-for-perf-tests/loadtests --influxMeasurement=java_streaming_cogbk_1 --influxTags={"runnerVersion":"v2","jdk":"java17"} --publishToInfluxDB=true --sourceOptions={"numRecords":20000000,"keySizeBytes":10,"valueSizeBytes":90,"numHotKeys":1} --coSourceOptions={"numRecords":2000000,"keySizeBytes":10,"valueSizeBytes":90,"numHotKeys":1000} --iterations=1 --numWorkers=5 --autoscalingAlgorithm=NONE --streaming=true --influxDatabase=beam_test_metrics --influxHost=http://10.128.0.96:8086 --inputWindowDurationSec=1200 --coInputWindowDurationSec=1200 --runner=DataflowRunner' -Prunner.version=V2 -PcompileAndRunTestsWithJava17 -Pjava17Home=/usr/lib/jvm/java-17-openjdk-amd64 --continue --max-****s=12 -Dorg.gradle.jvmargs=-Xms2g -Dorg.gradle.jvmargs=-Xmx4g -Dorg.gradle.vfs.watch=false -Pdocker-pull-licenses :sdks:java:testing:load-tests:run
Starting a Gradle Daemon, 3 busy Daemons could not be reused, use --status for details
Configuration on demand is an incubating feature.
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy FROM-CACHE
> Task :buildSrc:pluginDescriptors
> Task :buildSrc:processResources
> Task :buildSrc:classes
> Task :buildSrc:jar
> Task :buildSrc:assemble
> Task :buildSrc:spotlessGroovy FROM-CACHE
> Task :buildSrc:spotlessGroovyCheck UP-TO-DATE
> Task :buildSrc:spotlessGroovyGradle FROM-CACHE
> Task :buildSrc:spotlessGroovyGradleCheck UP-TO-DATE
> Task :buildSrc:spotlessCheck UP-TO-DATE
> Task :buildSrc:pluginUnderTestMetadata
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validatePlugins FROM-CACHE
> Task :buildSrc:check UP-TO-DATE
> Task :buildSrc:build

> Configure project :sdks:go:test
System Go installation: /snap/bin/go is go version go1.16.13 linux/amd64; Preparing to use /home/jenkins/go/bin/go1.16.12
go1.16.12: already downloaded in /home/jenkins/sdk/go1.16.12
GOCMD=/home/jenkins/go/bin/go1.16.12

FAILURE: Build failed with an exception.

* Where:
Build file '<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/testing/load-tests/build.gradle'> line: 113

* What went wrong:
A problem occurred evaluating project ':sdks:java:testing:load-tests'.
> Could not get unknown property 'dockerImageName' for project ':runners:google-cloud-dataflow-java' of type org.gradle.api.Project.

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

BUILD FAILED in 13s
10 actionable tasks: 4 executed, 4 from cache, 2 up-to-date

Publishing build scan...
https://gradle.com/s/d5wzmwqqndzsq

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #44

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/44/display/redirect?page=changes>

Changes:

[Kyle Weaver] [BEAM-13751] Don't block on gcloud when attempting to get default GCP

[Kyle Weaver] [BEAM-13751] Parameterize wait timeout so test doesn't waste 2s.

[Kyle Weaver] [BEAM-13751] Add comment explaining sleep.

[noreply] Update Python SDK beam-master tags (#16630)

[noreply] Merge pull request #16592 from [BEAM-13722][Playground] Add precompiling

[noreply] Merge pull request #16505 from [BEAM-13527] [Playground] Pipeline

[noreply] [BEAM-13293] XLang Jdbc IO for Go SDK (#16111)

[noreply] [BEAM-10206] Add Go Vet to Github Actions (#16612)


------------------------------------------
Started by timer
Running as SYSTEM
[EnvInject] - Loading node environment variables.
Building remotely on apache-beam-jenkins-2 (beam) in workspace <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/>
The recommended git tool is: NONE
No credentials specified
Wiping out workspace first.
Cloning the remote Git repository
Cloning repository https://github.com/apache/beam.git
 > git init <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src> # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git --version # timeout=10
 > git --version # 'git version 2.25.1'
 > git fetch --tags --force --progress -- https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git fetch --tags --force --progress -- https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* +refs/pull/${ghprbPullId}/*:refs/remotes/origin/pr/${ghprbPullId}/* # timeout=10
 > git rev-parse origin/master^{commit} # timeout=10
Checking out Revision 6c9c208197d3d74b1c3643d22716ad3b00213506 (origin/master)
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 6c9c208197d3d74b1c3643d22716ad3b00213506 # timeout=10
Commit message: "[BEAM-10206] Add Go Vet to Github Actions (#16612)"
 > git rev-list --no-walk 178cb7b65401d860b63fc7415f02f1dea2c4582f # timeout=10
No emails were triggered.
[EnvInject] - Executing scripts and injecting environment variables after the SCM step.
[EnvInject] - Injecting as environment variables the properties content 
SPARK_LOCAL_IP=127.0.0.1
SETUPTOOLS_USE_DISTUTILS=stdlib

[EnvInject] - Variables injected successfully.
[beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17] $ /bin/bash -xe /tmp/jenkins5750276109620333099.sh
+ echo '*** Load test: CoGBK 2GB 100  byte records - single key ***'
*** Load test: CoGBK 2GB 100  byte records - single key ***
[Gradle] - Launching build.
[src] $ <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/gradlew> -PloadTest.mainClass=org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest -Prunner=:runners:google-cloud-dataflow-java '-PloadTest.args=--project=apache-beam-testing --region=us-central1 --appName=load_tests_Java17_Dataflow_V2_streaming_CoGBK_1 --tempLocation=gs://temp-storage-for-perf-tests/loadtests --influxMeasurement=java_streaming_cogbk_1 --influxTags={"runnerVersion":"v2","jdk":"java17"} --publishToInfluxDB=true --sourceOptions={"numRecords":20000000,"keySizeBytes":10,"valueSizeBytes":90,"numHotKeys":1} --coSourceOptions={"numRecords":2000000,"keySizeBytes":10,"valueSizeBytes":90,"numHotKeys":1000} --iterations=1 --numWorkers=5 --autoscalingAlgorithm=NONE --streaming=true --influxDatabase=beam_test_metrics --influxHost=http://10.128.0.96:8086 --inputWindowDurationSec=1200 --coInputWindowDurationSec=1200 --runner=DataflowRunner' -Prunner.version=V2 -PcompileAndRunTestsWithJava17 -Pjava17Home=/usr/lib/jvm/java-17-openjdk-amd64 --continue --max-****s=12 -Dorg.gradle.jvmargs=-Xms2g -Dorg.gradle.jvmargs=-Xmx4g -Dorg.gradle.vfs.watch=false -Pdocker-pull-licenses :sdks:java:testing:load-tests:run
Starting a Gradle Daemon, 3 busy Daemons could not be reused, use --status for details
Configuration on demand is an incubating feature.
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy FROM-CACHE
> Task :buildSrc:pluginDescriptors
> Task :buildSrc:processResources
> Task :buildSrc:classes
> Task :buildSrc:jar
> Task :buildSrc:assemble
> Task :buildSrc:spotlessGroovy FROM-CACHE
> Task :buildSrc:spotlessGroovyCheck UP-TO-DATE
> Task :buildSrc:spotlessGroovyGradle FROM-CACHE
> Task :buildSrc:spotlessGroovyGradleCheck UP-TO-DATE
> Task :buildSrc:spotlessCheck UP-TO-DATE
> Task :buildSrc:pluginUnderTestMetadata
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validatePlugins FROM-CACHE
> Task :buildSrc:check UP-TO-DATE
> Task :buildSrc:build

> Configure project :sdks:go:test
System Go installation: /snap/bin/go is go version go1.16.13 linux/amd64; Preparing to use /home/jenkins/go/bin/go1.16.12
go1.16.12: already downloaded in /home/jenkins/sdk/go1.16.12
GOCMD=/home/jenkins/go/bin/go1.16.12

FAILURE: Build failed with an exception.

* Where:
Build file '<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/testing/load-tests/build.gradle'> line: 113

* What went wrong:
A problem occurred evaluating project ':sdks:java:testing:load-tests'.
> Could not get unknown property 'dockerImageName' for project ':runners:google-cloud-dataflow-java' of type org.gradle.api.Project.

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

BUILD FAILED in 14s
10 actionable tasks: 4 executed, 4 from cache, 2 up-to-date

Publishing build scan...
https://gradle.com/s/qwgh4bb2nn6cy

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #43

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/43/display/redirect?page=changes>

Changes:

[marcin.kuthan] Get rid of unnessecary logs for BigQuery streaming writes in

[dhuntsperger] added GitHub example references to Python multilang quickstart

[mmack] [adhoc] Test S3Options and AwsOptions for Sdk v2

[noreply] [BEAM-13537] Fix NPE in kafkatopubsub example (#16625)

[noreply] [BEAM-13740] update java_tests.yml to remove setup-go, which is

[Heejong Lee] Fix google3 import error

[noreply] [BEAM-12976] Implement Java projection pushdown optimizer. (#16513)

[noreply] Merge pull request #16579 from Revert "Revert "Merge pull request #15863

[noreply] Merge pull request #16606 from [BEAM-13247] [Playground] Embedding


------------------------------------------
Started by timer
Running as SYSTEM
[EnvInject] - Loading node environment variables.
Building remotely on apache-beam-jenkins-1 (beam) in workspace <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/>
The recommended git tool is: NONE
No credentials specified
Wiping out workspace first.
Cloning the remote Git repository
Cloning repository https://github.com/apache/beam.git
 > git init <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src> # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git --version # timeout=10
 > git --version # 'git version 2.7.4'
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git fetch --tags --progress https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* +refs/pull/${ghprbPullId}/*:refs/remotes/origin/pr/${ghprbPullId}/* # timeout=10
 > git rev-parse origin/master^{commit} # timeout=10
Checking out Revision 178cb7b65401d860b63fc7415f02f1dea2c4582f (origin/master)
 > git config core.sparsecheckout # timeout=10
 > git checkout -f 178cb7b65401d860b63fc7415f02f1dea2c4582f # timeout=10
Commit message: "Merge pull request #16606 from [BEAM-13247] [Playground] Embedding iframe"
 > git rev-list --no-walk f687ece82d3623062fb1ab0f7b3e1366638ad867 # timeout=10
No emails were triggered.
[EnvInject] - Executing scripts and injecting environment variables after the SCM step.
[EnvInject] - Injecting as environment variables the properties content 
SPARK_LOCAL_IP=127.0.0.1
SETUPTOOLS_USE_DISTUTILS=stdlib

[EnvInject] - Variables injected successfully.
[beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17] $ /bin/bash -xe /tmp/jenkins4139279424303748173.sh
+ echo '*** Load test: CoGBK 2GB 100  byte records - single key ***'
*** Load test: CoGBK 2GB 100  byte records - single key ***
[Gradle] - Launching build.
[src] $ <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/gradlew> -PloadTest.mainClass=org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest -Prunner=:runners:google-cloud-dataflow-java '-PloadTest.args=--project=apache-beam-testing --region=us-central1 --appName=load_tests_Java17_Dataflow_V2_streaming_CoGBK_1 --tempLocation=gs://temp-storage-for-perf-tests/loadtests --influxMeasurement=java_streaming_cogbk_1 --influxTags={"runnerVersion":"v2","jdk":"java17"} --publishToInfluxDB=true --sourceOptions={"numRecords":20000000,"keySizeBytes":10,"valueSizeBytes":90,"numHotKeys":1} --coSourceOptions={"numRecords":2000000,"keySizeBytes":10,"valueSizeBytes":90,"numHotKeys":1000} --iterations=1 --numWorkers=5 --autoscalingAlgorithm=NONE --streaming=true --influxDatabase=beam_test_metrics --influxHost=http://10.128.0.96:8086 --inputWindowDurationSec=1200 --coInputWindowDurationSec=1200 --runner=DataflowRunner' -Prunner.version=V2 -PcompileAndRunTestsWithJava17 -Pjava17Home=/usr/lib/jvm/java-17-openjdk-amd64 --continue --max-****s=12 -Dorg.gradle.jvmargs=-Xms2g -Dorg.gradle.jvmargs=-Xmx4g -Dorg.gradle.vfs.watch=false -Pdocker-pull-licenses :sdks:java:testing:load-tests:run
Starting a Gradle Daemon, 4 busy and 1 incompatible Daemons could not be reused, use --status for details
Configuration on demand is an incubating feature.
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy FROM-CACHE
> Task :buildSrc:pluginDescriptors
> Task :buildSrc:processResources
> Task :buildSrc:classes
> Task :buildSrc:jar
> Task :buildSrc:assemble
> Task :buildSrc:spotlessGroovy FROM-CACHE
> Task :buildSrc:spotlessGroovyCheck UP-TO-DATE
> Task :buildSrc:spotlessGroovyGradle FROM-CACHE
> Task :buildSrc:spotlessGroovyGradleCheck UP-TO-DATE
> Task :buildSrc:spotlessCheck UP-TO-DATE
> Task :buildSrc:pluginUnderTestMetadata
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validatePlugins FROM-CACHE
> Task :buildSrc:check UP-TO-DATE
> Task :buildSrc:build

> Configure project :sdks:go:test
System Go installation: /usr/bin/go is go version go1.16.13 linux/amd64; Preparing to use /home/jenkins/go/bin/go1.16.12
go1.16.12: already downloaded in /home/jenkins/sdk/go1.16.12
GOCMD=/home/jenkins/go/bin/go1.16.12

FAILURE: Build failed with an exception.

* Where:
Build file '<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/testing/load-tests/build.gradle'> line: 113

* What went wrong:
A problem occurred evaluating project ':sdks:java:testing:load-tests'.
> Could not get unknown property 'dockerImageName' for project ':runners:google-cloud-dataflow-java' of type org.gradle.api.Project.

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

BUILD FAILED in 14s
10 actionable tasks: 4 executed, 4 from cache, 2 up-to-date

Publishing build scan...
https://gradle.com/s/wbyxpplltn6vk

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #42

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/42/display/redirect?page=changes>

Changes:

[Heejong Lee] [BEAM-13093] Enable JavaUsingPython CrossLanguageValidateRunner test for

[mmack] [BEAM-13746] Fix deserialization of SSECustomerKey for AWS Sdk v2

[noreply] [BEAM-7928] Allow users to specify worker disk type for Dataflow runner

[noreply] Merge pull request #16534 from [BEAM-13671][Playground] Add backend

[noreply] [BEAM-13271] Bump errorprone to 2.10.0 (#16231)

[noreply] [BEAM-13595] Don't load main session when cloudpickle is used. (#16589)

[Heejong Lee] Update readme for XVR tests


------------------------------------------
Started by timer
Running as SYSTEM
[EnvInject] - Loading node environment variables.
Building remotely on apache-beam-jenkins-7 (beam) in workspace <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/>
The recommended git tool is: NONE
No credentials specified
Wiping out workspace first.
Cloning the remote Git repository
Cloning repository https://github.com/apache/beam.git
 > git init <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src> # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git --version # timeout=10
 > git --version # 'git version 2.25.1'
 > git fetch --tags --force --progress -- https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
 > git config remote.origin.url https://github.com/apache/beam.git # timeout=10
Fetching upstream changes from https://github.com/apache/beam.git
 > git fetch --tags --force --progress -- https://github.com/apache/beam.git +refs/heads/*:refs/remotes/origin/* +refs/pull/${ghprbPullId}/*:refs/remotes/origin/pr/${ghprbPullId}/* # timeout=10
 > git rev-parse origin/master^{commit} # timeout=10
Checking out Revision f687ece82d3623062fb1ab0f7b3e1366638ad867 (origin/master)
 > git config core.sparsecheckout # timeout=10
 > git checkout -f f687ece82d3623062fb1ab0f7b3e1366638ad867 # timeout=10
Commit message: "Merge pull request #16626 from ihji/update_readme"
 > git rev-list --no-walk 41d585f82b10195f758d14e3a54076ea1f05aa75 # timeout=10
No emails were triggered.
[EnvInject] - Executing scripts and injecting environment variables after the SCM step.
[EnvInject] - Injecting as environment variables the properties content 
SPARK_LOCAL_IP=127.0.0.1
SETUPTOOLS_USE_DISTUTILS=stdlib

[EnvInject] - Variables injected successfully.
[beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17] $ /bin/bash -xe /tmp/jenkins4631097003613772399.sh
+ echo '*** Load test: CoGBK 2GB 100  byte records - single key ***'
*** Load test: CoGBK 2GB 100  byte records - single key ***
[Gradle] - Launching build.
[src] $ <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/gradlew> -PloadTest.mainClass=org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest -Prunner=:runners:google-cloud-dataflow-java '-PloadTest.args=--project=apache-beam-testing --region=us-central1 --appName=load_tests_Java17_Dataflow_V2_streaming_CoGBK_1 --tempLocation=gs://temp-storage-for-perf-tests/loadtests --influxMeasurement=java_streaming_cogbk_1 --influxTags={"runnerVersion":"v2","jdk":"java17"} --publishToInfluxDB=true --sourceOptions={"numRecords":20000000,"keySizeBytes":10,"valueSizeBytes":90,"numHotKeys":1} --coSourceOptions={"numRecords":2000000,"keySizeBytes":10,"valueSizeBytes":90,"numHotKeys":1000} --iterations=1 --numWorkers=5 --autoscalingAlgorithm=NONE --streaming=true --influxDatabase=beam_test_metrics --influxHost=http://10.128.0.96:8086 --inputWindowDurationSec=1200 --coInputWindowDurationSec=1200 --runner=DataflowRunner' -Prunner.version=V2 -PcompileAndRunTestsWithJava17 -Pjava17Home=/usr/lib/jvm/java-17-openjdk-amd64 --continue --max-****s=12 -Dorg.gradle.jvmargs=-Xms2g -Dorg.gradle.jvmargs=-Xmx4g -Dorg.gradle.vfs.watch=false -Pdocker-pull-licenses :sdks:java:testing:load-tests:run
Starting a Gradle Daemon, 3 busy Daemons could not be reused, use --status for details
Configuration on demand is an incubating feature.
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy FROM-CACHE
> Task :buildSrc:pluginDescriptors
> Task :buildSrc:processResources
> Task :buildSrc:classes
> Task :buildSrc:jar
> Task :buildSrc:assemble
> Task :buildSrc:spotlessGroovy FROM-CACHE
> Task :buildSrc:spotlessGroovyCheck UP-TO-DATE
> Task :buildSrc:spotlessGroovyGradle FROM-CACHE
> Task :buildSrc:spotlessGroovyGradleCheck UP-TO-DATE
> Task :buildSrc:spotlessCheck UP-TO-DATE
> Task :buildSrc:pluginUnderTestMetadata
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validatePlugins FROM-CACHE
> Task :buildSrc:check UP-TO-DATE
> Task :buildSrc:build

> Configure project :sdks:go:test
System Go installation: /snap/bin/go is go version go1.16.13 linux/amd64; Preparing to use /home/jenkins/go/bin/go1.16.12
go1.16.12: already downloaded in /home/jenkins/sdk/go1.16.12
GOCMD=/home/jenkins/go/bin/go1.16.12

FAILURE: Build failed with an exception.

* Where:
Build file '<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/testing/load-tests/build.gradle'> line: 113

* What went wrong:
A problem occurred evaluating project ':sdks:java:testing:load-tests'.
> Could not get unknown property 'dockerImageName' for project ':runners:google-cloud-dataflow-java' of type org.gradle.api.Project.

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

BUILD FAILED in 25s
10 actionable tasks: 4 executed, 4 from cache, 2 up-to-date

Publishing build scan...
https://gradle.com/s/isiniipku7zg2

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #41

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/41/display/redirect?page=changes>

Changes:

[artur.khanin] Privacy policy update regarding Apache Beam Playground

[Daniel Oliveira] [BEAM-13321] Fix exception with BigQuery StreamWriter TraceID.

[mmack] [BEAM-8807] Add integration test for SnsIO.write (Sdk v1 & v2)

[noreply] [BEAM-13736] Make lifting cache exact. (#16603)

[noreply] Merge pull request #16565 from [BEAM-13692][Playground]  Implement

[noreply] Merge pull request #16502 from [BEAM-13650][Playground] Add link for

[noreply] [BEAM-13310] remove call to get offset consumer config, which was rep…


------------------------------------------
[...truncated 50.06 KB...]
bde539237f9f: Pushed
16a0fdc827f8: Pushed
35b30c715808: Pushed
7d5478d2e029: Pushed
bff9fe6e429c: Layer already exists
7c245b2fe4f1: Layer already exists
1f0fe2e1ec6c: Pushed
f9e18e59a565: Layer already exists
26a504e63be4: Layer already exists
8bf42db0de72: Layer already exists
31892cc314cb: Layer already exists
11936051f93b: Layer already exists
7dbd8ac32018: Pushed
d6fbc5a26bbd: Pushed
7033fee51afd: Pushed
20220126123138: digest: sha256:c07e8f47c4537b9f91d3814dfe4b4fbed465cd40fe06f32257aed502f7fb4bb6 size: 4520

> Task :sdks:java:testing:load-tests:run
Jan 26, 2022 12:33:33 PM org.apache.beam.runners.dataflow.options.DataflowPipelineOptions$StagingLocationFactory create
INFO: No stagingLocation provided, falling back to gcpTempLocation
Jan 26, 2022 12:33:34 PM org.apache.beam.runners.dataflow.DataflowRunner fromOptions
INFO: PipelineOptions.filesToStage was not specified. Defaulting to files from the classpath: will stage 204 files. Enable logging at DEBUG level to see which files will be staged.
Jan 26, 2022 12:33:34 PM org.apache.beam.sdk.Pipeline validate
WARNING: The following transforms do not have stable unique names: Window.Into()
Jan 26, 2022 12:33:34 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Executing pipeline on the Dataflow Service, which will have billing implications related to Google Compute Engine usage and other Google Cloud Services.
Jan 26, 2022 12:33:38 PM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Uploading 204 files from PipelineOptions.filesToStage to staging location to prepare for execution.
Jan 26, 2022 12:33:39 PM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Staging files complete: 204 files cached, 0 files newly uploaded in 1 seconds
Jan 26, 2022 12:33:39 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Staging portable pipeline proto to gs://temp-storage-for-perf-tests/loadtests/staging/
Jan 26, 2022 12:33:39 PM org.apache.beam.runners.dataflow.util.PackageUtil tryStagePackage
INFO: Uploading <114075 bytes, hash ae8293b5d2730a1b786c030a105123ffb161118e1deda2435b7dd4fb7bcbf51e> to gs://temp-storage-for-perf-tests/loadtests/staging/pipeline-roKTtdJzCht4bAMKEFEj_7FhEY4d7aJDW33U-3vL9R4.pb
Jan 26, 2022 12:33:41 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read input/DataflowRunner.StreamingUnboundedRead.ReadWithIds as step s1
Jan 26, 2022 12:33:41 PM org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource split
INFO: Split into 20 bundles of sizes: [org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1947596f, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3078cac, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@f6de586, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5f2bd6d9, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7c847072, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@43d9f1a2, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@23f86d8a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2264ea32, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7d3c09ec, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@71e4b308, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@11900483, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@14a049f9, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@94e51e8, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5de6cf3a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4cc36c19, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5a3a1bf9, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2100d047, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4af45442, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1d805aa1, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@176f7f3b]
Jan 26, 2022 12:33:41 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read input/StripIds as step s2
Jan 26, 2022 12:33:41 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect start time metrics (input) as step s3
Jan 26, 2022 12:33:41 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Window.Into()/Window.Assign as step s4
Jan 26, 2022 12:33:41 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read co-input/DataflowRunner.StreamingUnboundedRead.ReadWithIds as step s5
Jan 26, 2022 12:33:41 PM org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource split
INFO: Split into 20 bundles of sizes: [org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1df1ced0, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5349b246, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@32b0876c, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2aaf152b, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@39bbd9e0, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@27fe9713, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@11c3ff67, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4397a639, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@24841372, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@77114efe, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@79a7bfbc, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@77f68df, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3e4e4c1, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7e7f3cfd, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3ae126d1, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@46a488c2, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6242ae3b, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@65ddee5a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@56399b9e, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@34b9eb03]
Jan 26, 2022 12:33:41 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read co-input/StripIds as step s6
Jan 26, 2022 12:33:41 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect start time metrics (co-input) as step s7
Jan 26, 2022 12:33:41 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Window.Into()2/Window.Assign as step s8
Jan 26, 2022 12:33:41 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/MakeUnionTable0 as step s9
Jan 26, 2022 12:33:41 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/MakeUnionTable1 as step s10
Jan 26, 2022 12:33:41 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/Flatten as step s11
Jan 26, 2022 12:33:41 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/GBK as step s12
Jan 26, 2022 12:33:41 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/ConstructCoGbkResultFn as step s13
Jan 26, 2022 12:33:41 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Ungroup and reiterate as step s14
Jan 26, 2022 12:33:41 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect total bytes as step s15
Jan 26, 2022 12:33:41 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect end time metrics as step s16
Jan 26, 2022 12:33:41 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Dataflow SDK version: 2.37.0-SNAPSHOT
Jan 26, 2022 12:33:43 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To access the Dataflow monitoring console, please navigate to https://console.cloud.google.com/dataflow/jobs/us-central1/2022-01-26_04_33_41-9771457597955160432?project=apache-beam-testing
Jan 26, 2022 12:33:43 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Submitted job: 2022-01-26_04_33_41-9771457597955160432
Jan 26, 2022 12:33:43 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To cancel the job using the 'gcloud' tool, run:
> gcloud dataflow jobs --project=apache-beam-testing cancel --region=us-central1 2022-01-26_04_33_41-9771457597955160432
Jan 26, 2022 12:33:58 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
WARNING: 2022-01-26T12:33:51.087Z: The workflow name is not a valid Cloud Label. Labels applied to Cloud resources (such as GCE Instances) for monitoring will be labeled with this modified job name: load0tests0java170dataflow0v20streaming0cogbk01-jenkins-01-1cr3. For the best monitoring experience, please name your job with a valid Cloud Label. For details, see: https://cloud.google.com/compute/docs/labeling-resources#restrictions
Jan 26, 2022 12:34:00 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-26T12:34:00.210Z: Worker configuration: e2-standard-2 in us-central1-b.
Jan 26, 2022 12:34:02 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-26T12:34:00.987Z: Expanding SplittableParDo operations into optimizable parts.
Jan 26, 2022 12:34:02 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-26T12:34:01.014Z: Expanding CollectionToSingleton operations into optimizable parts.
Jan 26, 2022 12:34:02 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-26T12:34:01.075Z: Expanding CoGroupByKey operations into optimizable parts.
Jan 26, 2022 12:34:02 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-26T12:34:01.144Z: Expanding SplittableProcessKeyed operations into optimizable parts.
Jan 26, 2022 12:34:02 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-26T12:34:01.172Z: Expanding GroupByKey operations into streaming Read/Write steps
Jan 26, 2022 12:34:02 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-26T12:34:01.239Z: Lifting ValueCombiningMappingFns into MergeBucketsMappingFns
Jan 26, 2022 12:34:02 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-26T12:34:01.344Z: Fusing adjacent ParDo, Read, Write, and Flatten operations
Jan 26, 2022 12:34:02 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-26T12:34:01.402Z: Unzipping flatten CoGroupByKey-Flatten for input CoGroupByKey-MakeUnionTable0-ParMultiDo-ConstructUnionTable-.output
Jan 26, 2022 12:34:02 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-26T12:34:01.435Z: Fusing unzipped copy of CoGroupByKey/GBK/WriteStream, through flatten CoGroupByKey/Flatten, into producer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable)
Jan 26, 2022 12:34:02 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-26T12:34:01.457Z: Fusing consumer CoGroupByKey/GBK/WriteStream into CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable)
Jan 26, 2022 12:34:02 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-26T12:34:01.488Z: Fusing consumer Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read input/Impulse
Jan 26, 2022 12:34:02 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-26T12:34:01.519Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
Jan 26, 2022 12:34:02 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-26T12:34:01.553Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
Jan 26, 2022 12:34:02 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-26T12:34:01.576Z: Fusing consumer Read input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
Jan 26, 2022 12:34:02 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-26T12:34:01.602Z: Fusing consumer Collect start time metrics (input)/ParMultiDo(TimeMonitor) into Read input/ParDo(StripIds)/ParMultiDo(StripIds)
Jan 26, 2022 12:34:02 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-26T12:34:01.631Z: Fusing consumer Window.Into()/Window.Assign into Collect start time metrics (input)/ParMultiDo(TimeMonitor)
Jan 26, 2022 12:34:02 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-26T12:34:01.663Z: Fusing consumer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable) into Window.Into()/Window.Assign
Jan 26, 2022 12:34:02 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-26T12:34:01.696Z: Fusing consumer Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read co-input/Impulse
Jan 26, 2022 12:34:02 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-26T12:34:01.734Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
Jan 26, 2022 12:34:02 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-26T12:34:01.764Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
Jan 26, 2022 12:34:02 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-26T12:34:01.798Z: Fusing consumer Read co-input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
Jan 26, 2022 12:34:02 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-26T12:34:01.831Z: Fusing consumer Collect start time metrics (co-input)/ParMultiDo(TimeMonitor) into Read co-input/ParDo(StripIds)/ParMultiDo(StripIds)
Jan 26, 2022 12:34:02 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-26T12:34:01.866Z: Fusing consumer Window.Into()2/Window.Assign into Collect start time metrics (co-input)/ParMultiDo(TimeMonitor)
Jan 26, 2022 12:34:02 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-26T12:34:01.898Z: Fusing consumer CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable) into Window.Into()2/Window.Assign
Jan 26, 2022 12:34:02 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-26T12:34:01.933Z: Fusing consumer CoGroupByKey/GBK/MergeBuckets into CoGroupByKey/GBK/ReadStream
Jan 26, 2022 12:34:02 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-26T12:34:01.965Z: Fusing consumer CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult) into CoGroupByKey/GBK/MergeBuckets
Jan 26, 2022 12:34:02 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-26T12:34:01.986Z: Fusing consumer Ungroup and reiterate/ParMultiDo(UngroupAndReiterate) into CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult)
Jan 26, 2022 12:34:02 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-26T12:34:02.023Z: Fusing consumer Collect total bytes/ParMultiDo(ByteMonitor) into Ungroup and reiterate/ParMultiDo(UngroupAndReiterate)
Jan 26, 2022 12:34:02 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-26T12:34:02.058Z: Fusing consumer Collect end time metrics/ParMultiDo(TimeMonitor) into Collect total bytes/ParMultiDo(ByteMonitor)
Jan 26, 2022 12:34:04 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-26T12:34:02.506Z: Starting 5 ****s in us-central1-b...
Jan 26, 2022 12:34:23 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-26T12:34:22.623Z: Your project already contains 100 Dataflow-created metric descriptors, so new user metrics of the form custom.googleapis.com/* will not be created. However, all user metrics are also available in the metric dataflow.googleapis.com/job/user_counter. If you rely on the custom metrics, you can delete old / unused metric descriptors. See https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.list and https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.delete
Jan 26, 2022 12:34:49 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-26T12:34:48.712Z: Autoscaling: Raised the number of ****s to 5 so that the pipeline can catch up with its backlog and keep up with its input rate.
Jan 26, 2022 12:35:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-26T12:35:42.993Z: Workers have started successfully.
Jan 26, 2022 12:35:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-26T12:35:43.028Z: Workers have started successfully.
Jan 26, 2022 4:00:36 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-26T16:00:34.535Z: Cancel request is committed for workflow job: 2022-01-26_04_33_41-9771457597955160432.
Jan 26, 2022 4:00:36 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-26T16:00:34.597Z: Cleaning up.
Jan 26, 2022 4:00:36 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-26T16:00:34.710Z: Stopping **** pool...
Jan 26, 2022 4:00:36 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-26T16:00:34.753Z: Stopping **** pool...
Jan 26, 2022 4:02:59 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-26T16:02:56.780Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Jan 26, 2022 4:02:59 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-26T16:02:56.828Z: Worker pool stopped.
Jan 26, 2022 4:03:03 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-01-26_04_33_41-9771457597955160432 finished with status CANCELLED.
Load test results for test (ID): 78128151-55df-42be-907c-30cfaae3edc3 and timestamp: 2022-01-26T12:33:34.578000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                 12258.111
dataflow_v2_java17_total_bytes_count             2.30981313E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220126123138
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:c07e8f47c4537b9f91d3814dfe4b4fbed465cd40fe06f32257aed502f7fb4bb6
Deleted: sha256:5cfc0a7543d13c42bd775c9bad3ebf9147411acb2b54bd3bf3c27f1de8e6c6ca
Deleted: sha256:331247c8ce500071d571b7cc9dbe20b52810679a358859d83ab1c9a69fd6b4ac
Deleted: sha256:7c4a743ca38fdd9173b3633f61dc06133fc8c26c957cd3a1d278ebb3daac3aed
Deleted: sha256:8c0ef51f28d90f87fec3999e9f68018adb8e4ba68cff7cd2c9ce25f837f13bdf
Deleted: sha256:0f5464314ded60e7ed37936342ac32b3886dcc26aae9cb00c3094b90f7a22bef
Deleted: sha256:3fe6c87a00efba76518ba7825a517afb03c826def7fb7e1ce2b1cd042ca75a8c
Deleted: sha256:3d2930f2016c14737743b6d23fb4f4145a26f63e64343df198a9d3a8435296dc
Deleted: sha256:c6e6b3a0a561c8e574f5845afd47329fcf3af43cc13f3df21c007f7801e4ada9
Deleted: sha256:9e95ce2f7b0f0faa5d31a671d0f49b12aeee2033165b367465a209a367b74ac8
Deleted: sha256:8a582fa62ed8ac0ff4cdfb7e2c9429d5144679b03bc1060652b4a25eaf3c0f71
Deleted: sha256:995148d8114f803ca82ba725121cd66e8740121757d342625bca7f26125157ee
Deleted: sha256:ecd44d3181d7e45fdc330485678e75166770e2e1ff4ec3592d2cabf3adfdf655
Deleted: sha256:e4473159aa55374602e22861391822143241bb29b702d3c976cf8d698c21cf97
Deleted: sha256:ec8cbc7fa41a9211ef673bba65d4da57d116e27aab759fbfe2fc8888b84b34f1
Deleted: sha256:4fe3faab18494fdf652d3e2801b1fa85e70dd11ba5730ecb7302af69379b5992
Deleted: sha256:09db60ba46f963b4c4b8e0c707d0d4f4973f14fa7714550cf966eb9e728423c4
Deleted: sha256:613d5a029eb0eeb344f5d283c903c132887c979b7d48f9857518aaee123ef68e
Deleted: sha256:d21a7ba8ac54964905b502d417f5fd8d2d07b38f8460b495ceeaa6a76ea66634
Deleted: sha256:2097714accc54a895f7ae3dafe953caf69a02ae90ad1534dafe88651be23ae98
Deleted: sha256:8a3748c1e699aeab4b91bdef191ef4c0acade144a0102ad9385018b57038d85a
Deleted: sha256:2c32a6f317c3ac8939bfa3bf54c249dc07dde639b6429bea762cd2e94b573bd8
Deleted: sha256:a033577986f8f374f36c441c2ef10f7a6e3cbc928dfa582ddaefad7090bb0ec6
Deleted: sha256:f0148302bfc087779f20e93743a670c32f77cbc8476b1e2480ff0fed990ade9f
Deleted: sha256:eab8f730698dea4b4a2ea46d19d773a516f1f4a681d0abf95db21b1c95c54a5a
Deleted: sha256:d7f27f15bf18c541adc8e0121c9c3ec8169ac2ddc40d78660062b486b49a0f3c
Deleted: sha256:0841c7cf1bacdda5f86624a3dd07e5d67cbe4af84fcce53837995e9df37e6bce
Deleted: sha256:23580d7db5cba52d3db3a3c60b470dc9e9bb3af7c44bf5a42cec8f4dea6f0c11
Deleted: sha256:928ef8f91ee8ad4b0b7c9e12788e1afd2b971d435dacad59baea0a57264f535a
Deleted: sha256:2faf2546ad138b28222142d09d224c04b40e3baad24d0515c1de4bf7fb616975
Deleted: sha256:f9404b89111d6873d4db59823328e653abb64ddb77a1351f37cf6e4242c28a46
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220126123138]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:c07e8f47c4537b9f91d3814dfe4b4fbed465cd40fe06f32257aed502f7fb4bb6]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220126123138] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:c07e8f47c4537b9f91d3814dfe4b4fbed465cd40fe06f32257aed502f7fb4bb6])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:c07e8f47c4537b9f91d3814dfe4b4fbed465cd40fe06f32257aed502f7fb4bb6
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:c07e8f47c4537b9f91d3814dfe4b4fbed465cd40fe06f32257aed502f7fb4bb6
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:c07e8f47c4537b9f91d3814dfe4b4fbed465cd40fe06f32257aed502f7fb4bb6].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 31m 43s
109 actionable tasks: 72 executed, 33 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/evlgg25psnzlu

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #40

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/40/display/redirect?page=changes>

Changes:

[Heejong Lee] [BEAM-13716] Clear before creating a new virtual environment in

[mmack] [BEAM-13653] Make SnsIO.write topicArn optional. If provided, validate

[noreply] [BEAM-10897] Update the fastavro lower bound due to an issue on Windows

[noreply] [BEAM-13605] Update pandas_doctests_test denylists in preparation for

[noreply] Merge pull request #16538 from [BEAM-13676][Playground][Bugfix]Build Of

[noreply] Merge pull request #16582 from [BEAM-13711] [Playground] [Bugfix] Add

[noreply] Merge pull request #16515 from [BEAM-13636] [Playground] Checking the

[ningkang0957] [BEAM-13275] Removed the explicit selenium dependency from setup

[noreply] [BEAM-10206] Deprecate unused shallow cloning functions (#16600)

[noreply] Bump Dataflow container versions (#16602)

[noreply] Improved multi-language pipelines section of the programming guide

[mmack] [BEAM-13510] Don't retry on invalid SQS receipt handles.


------------------------------------------
[...truncated 49.95 KB...]
8c070ffa5696: Preparing
51039879420c: Preparing
6072fab9e7ec: Preparing
d4a110d75ecf: Preparing
bff9fe6e429c: Preparing
7c245b2fe4f1: Preparing
f9e18e59a565: Preparing
26a504e63be4: Preparing
8bf42db0de72: Preparing
31892cc314cb: Preparing
11936051f93b: Preparing
8c070ffa5696: Waiting
98abb4913b74: Waiting
136b91d1dc01: Waiting
26a504e63be4: Waiting
f9e18e59a565: Waiting
51039879420c: Waiting
11936051f93b: Waiting
8bf42db0de72: Waiting
6072fab9e7ec: Waiting
86d65ca2733e: Waiting
31892cc314cb: Waiting
bff9fe6e429c: Waiting
d4a110d75ecf: Waiting
7c245b2fe4f1: Waiting
a9b1e125b91a: Pushed
81dd55c067a9: Pushed
55e481102ce5: Pushed
e1864db91616: Pushed
86d65ca2733e: Pushed
98abb4913b74: Pushed
2781d63dcaeb: Pushed
8c070ffa5696: Pushed
d758799e696c: Pushed
bff9fe6e429c: Layer already exists
7c245b2fe4f1: Layer already exists
6072fab9e7ec: Pushed
f9e18e59a565: Layer already exists
26a504e63be4: Layer already exists
8bf42db0de72: Layer already exists
31892cc314cb: Layer already exists
11936051f93b: Layer already exists
136b91d1dc01: Pushed
d4a110d75ecf: Pushed
51039879420c: Pushed
20220125123544: digest: sha256:22e38d724a14e914d89c4169fd06e9b31db0e0e0016a1373810663332fb50e2d size: 4520

> Task :sdks:java:testing:load-tests:run
Jan 25, 2022 12:37:52 PM org.apache.beam.runners.dataflow.options.DataflowPipelineOptions$StagingLocationFactory create
INFO: No stagingLocation provided, falling back to gcpTempLocation
Jan 25, 2022 12:37:52 PM org.apache.beam.runners.dataflow.DataflowRunner fromOptions
INFO: PipelineOptions.filesToStage was not specified. Defaulting to files from the classpath: will stage 204 files. Enable logging at DEBUG level to see which files will be staged.
Jan 25, 2022 12:37:53 PM org.apache.beam.sdk.Pipeline validate
WARNING: The following transforms do not have stable unique names: Window.Into()
Jan 25, 2022 12:37:53 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Executing pipeline on the Dataflow Service, which will have billing implications related to Google Compute Engine usage and other Google Cloud Services.
Jan 25, 2022 12:37:56 PM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Uploading 204 files from PipelineOptions.filesToStage to staging location to prepare for execution.
Jan 25, 2022 12:37:56 PM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Staging files complete: 204 files cached, 0 files newly uploaded in 0 seconds
Jan 25, 2022 12:37:56 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Staging portable pipeline proto to gs://temp-storage-for-perf-tests/loadtests/staging/
Jan 25, 2022 12:37:56 PM org.apache.beam.runners.dataflow.util.PackageUtil tryStagePackage
INFO: Uploading <114075 bytes, hash 52a7d948ff7e538f0196f260f240fa00b6459015d3e76925890edf15e7e5606a> to gs://temp-storage-for-perf-tests/loadtests/staging/pipeline-UqfZSP9-U48BlvJg8kD6ALZFkBXT52kliQ7fFeflYGo.pb
Jan 25, 2022 12:37:58 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read input/DataflowRunner.StreamingUnboundedRead.ReadWithIds as step s1
Jan 25, 2022 12:37:58 PM org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource split
INFO: Split into 20 bundles of sizes: [org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3078cac, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@f6de586, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5f2bd6d9, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7c847072, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@43d9f1a2, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@23f86d8a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2264ea32, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7d3c09ec, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@71e4b308, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@11900483, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@14a049f9, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@94e51e8, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5de6cf3a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4cc36c19, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5a3a1bf9, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2100d047, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4af45442, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1d805aa1, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@176f7f3b, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@30ca0779]
Jan 25, 2022 12:37:58 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read input/StripIds as step s2
Jan 25, 2022 12:37:58 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect start time metrics (input) as step s3
Jan 25, 2022 12:37:58 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Window.Into()/Window.Assign as step s4
Jan 25, 2022 12:37:58 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read co-input/DataflowRunner.StreamingUnboundedRead.ReadWithIds as step s5
Jan 25, 2022 12:37:58 PM org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource split
INFO: Split into 20 bundles of sizes: [org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5349b246, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@32b0876c, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2aaf152b, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@39bbd9e0, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@27fe9713, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@11c3ff67, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4397a639, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@24841372, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@77114efe, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@79a7bfbc, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@77f68df, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3e4e4c1, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7e7f3cfd, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3ae126d1, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@46a488c2, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6242ae3b, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@65ddee5a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@56399b9e, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@34b9eb03, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@43fda8d9]
Jan 25, 2022 12:37:58 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read co-input/StripIds as step s6
Jan 25, 2022 12:37:58 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect start time metrics (co-input) as step s7
Jan 25, 2022 12:37:58 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Window.Into()2/Window.Assign as step s8
Jan 25, 2022 12:37:58 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/MakeUnionTable0 as step s9
Jan 25, 2022 12:37:58 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/MakeUnionTable1 as step s10
Jan 25, 2022 12:37:58 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/Flatten as step s11
Jan 25, 2022 12:37:58 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/GBK as step s12
Jan 25, 2022 12:37:58 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/ConstructCoGbkResultFn as step s13
Jan 25, 2022 12:37:58 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Ungroup and reiterate as step s14
Jan 25, 2022 12:37:58 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect total bytes as step s15
Jan 25, 2022 12:37:58 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect end time metrics as step s16
Jan 25, 2022 12:37:58 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Dataflow SDK version: 2.37.0-SNAPSHOT
Jan 25, 2022 12:37:59 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To access the Dataflow monitoring console, please navigate to https://console.cloud.google.com/dataflow/jobs/us-central1/2022-01-25_04_37_58-8254492904851712497?project=apache-beam-testing
Jan 25, 2022 12:37:59 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Submitted job: 2022-01-25_04_37_58-8254492904851712497
Jan 25, 2022 12:37:59 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To cancel the job using the 'gcloud' tool, run:
> gcloud dataflow jobs --project=apache-beam-testing cancel --region=us-central1 2022-01-25_04_37_58-8254492904851712497
Jan 25, 2022 12:38:08 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
WARNING: 2022-01-25T12:38:06.695Z: The workflow name is not a valid Cloud Label. Labels applied to Cloud resources (such as GCE Instances) for monitoring will be labeled with this modified job name: load0tests0java170dataflow0v20streaming0cogbk01-jenkins-01-8286. For the best monitoring experience, please name your job with a valid Cloud Label. For details, see: https://cloud.google.com/compute/docs/labeling-resources#restrictions
Jan 25, 2022 12:38:11 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-25T12:38:11.034Z: Worker configuration: e2-standard-2 in us-central1-b.
Jan 25, 2022 12:38:13 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-25T12:38:11.847Z: Expanding SplittableParDo operations into optimizable parts.
Jan 25, 2022 12:38:13 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-25T12:38:11.872Z: Expanding CollectionToSingleton operations into optimizable parts.
Jan 25, 2022 12:38:13 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-25T12:38:11.934Z: Expanding CoGroupByKey operations into optimizable parts.
Jan 25, 2022 12:38:13 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-25T12:38:11.995Z: Expanding SplittableProcessKeyed operations into optimizable parts.
Jan 25, 2022 12:38:13 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-25T12:38:12.022Z: Expanding GroupByKey operations into streaming Read/Write steps
Jan 25, 2022 12:38:13 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-25T12:38:12.082Z: Lifting ValueCombiningMappingFns into MergeBucketsMappingFns
Jan 25, 2022 12:38:13 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-25T12:38:12.193Z: Fusing adjacent ParDo, Read, Write, and Flatten operations
Jan 25, 2022 12:38:13 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-25T12:38:12.223Z: Unzipping flatten CoGroupByKey-Flatten for input CoGroupByKey-MakeUnionTable0-ParMultiDo-ConstructUnionTable-.output
Jan 25, 2022 12:38:13 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-25T12:38:12.259Z: Fusing unzipped copy of CoGroupByKey/GBK/WriteStream, through flatten CoGroupByKey/Flatten, into producer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable)
Jan 25, 2022 12:38:13 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-25T12:38:12.293Z: Fusing consumer CoGroupByKey/GBK/WriteStream into CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable)
Jan 25, 2022 12:38:13 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-25T12:38:12.329Z: Fusing consumer Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read input/Impulse
Jan 25, 2022 12:38:13 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-25T12:38:12.354Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
Jan 25, 2022 12:38:13 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-25T12:38:12.387Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
Jan 25, 2022 12:38:13 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-25T12:38:12.438Z: Fusing consumer Read input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
Jan 25, 2022 12:38:13 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-25T12:38:12.467Z: Fusing consumer Collect start time metrics (input)/ParMultiDo(TimeMonitor) into Read input/ParDo(StripIds)/ParMultiDo(StripIds)
Jan 25, 2022 12:38:13 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-25T12:38:12.503Z: Fusing consumer Window.Into()/Window.Assign into Collect start time metrics (input)/ParMultiDo(TimeMonitor)
Jan 25, 2022 12:38:13 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-25T12:38:12.526Z: Fusing consumer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable) into Window.Into()/Window.Assign
Jan 25, 2022 12:38:13 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-25T12:38:12.560Z: Fusing consumer Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read co-input/Impulse
Jan 25, 2022 12:38:13 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-25T12:38:12.588Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
Jan 25, 2022 12:38:13 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-25T12:38:12.619Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
Jan 25, 2022 12:38:13 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-25T12:38:12.646Z: Fusing consumer Read co-input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
Jan 25, 2022 12:38:13 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-25T12:38:12.670Z: Fusing consumer Collect start time metrics (co-input)/ParMultiDo(TimeMonitor) into Read co-input/ParDo(StripIds)/ParMultiDo(StripIds)
Jan 25, 2022 12:38:13 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-25T12:38:12.702Z: Fusing consumer Window.Into()2/Window.Assign into Collect start time metrics (co-input)/ParMultiDo(TimeMonitor)
Jan 25, 2022 12:38:13 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-25T12:38:12.734Z: Fusing consumer CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable) into Window.Into()2/Window.Assign
Jan 25, 2022 12:38:13 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-25T12:38:12.812Z: Fusing consumer CoGroupByKey/GBK/MergeBuckets into CoGroupByKey/GBK/ReadStream
Jan 25, 2022 12:38:13 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-25T12:38:12.876Z: Fusing consumer CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult) into CoGroupByKey/GBK/MergeBuckets
Jan 25, 2022 12:38:13 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-25T12:38:12.930Z: Fusing consumer Ungroup and reiterate/ParMultiDo(UngroupAndReiterate) into CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult)
Jan 25, 2022 12:38:13 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-25T12:38:12.977Z: Fusing consumer Collect total bytes/ParMultiDo(ByteMonitor) into Ungroup and reiterate/ParMultiDo(UngroupAndReiterate)
Jan 25, 2022 12:38:13 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-25T12:38:13.013Z: Fusing consumer Collect end time metrics/ParMultiDo(TimeMonitor) into Collect total bytes/ParMultiDo(ByteMonitor)
Jan 25, 2022 12:38:13 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-25T12:38:13.318Z: Starting 5 ****s in us-central1-b...
Jan 25, 2022 12:38:36 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-25T12:38:35.758Z: Your project already contains 100 Dataflow-created metric descriptors, so new user metrics of the form custom.googleapis.com/* will not be created. However, all user metrics are also available in the metric dataflow.googleapis.com/job/user_counter. If you rely on the custom metrics, you can delete old / unused metric descriptors. See https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.list and https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.delete
Jan 25, 2022 12:38:59 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-25T12:38:57.560Z: Autoscaling: Raised the number of ****s to 5 so that the pipeline can catch up with its backlog and keep up with its input rate.
Jan 25, 2022 12:39:55 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-25T12:39:54.376Z: Workers have started successfully.
Jan 25, 2022 12:39:55 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-25T12:39:54.415Z: Workers have started successfully.
Jan 25, 2022 4:00:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-25T16:00:26.271Z: Cancel request is committed for workflow job: 2022-01-25_04_37_58-8254492904851712497.
Jan 25, 2022 4:00:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-25T16:00:26.454Z: Cleaning up.
Jan 25, 2022 4:00:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-25T16:00:26.544Z: Stopping **** pool...
Jan 25, 2022 4:00:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-25T16:00:26.608Z: Stopping **** pool...
Jan 25, 2022 4:02:53 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-25T16:02:53.146Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Jan 25, 2022 4:02:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-25T16:02:53.191Z: Worker pool stopped.
Jan 25, 2022 4:03:03 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-01-25_04_37_58-8254492904851712497 finished with status CANCELLED.
Load test results for test (ID): 3ef62e1f-4f1a-40ac-8a6e-05408e12f361 and timestamp: 2022-01-25T12:37:53.329000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                 11966.561
dataflow_v2_java17_total_bytes_count             2.60561417E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220125123544
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:22e38d724a14e914d89c4169fd06e9b31db0e0e0016a1373810663332fb50e2d
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220125123544]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:22e38d724a14e914d89c4169fd06e9b31db0e0e0016a1373810663332fb50e2d]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220125123544] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:22e38d724a14e914d89c4169fd06e9b31db0e0e0016a1373810663332fb50e2d])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:22e38d724a14e914d89c4169fd06e9b31db0e0e0016a1373810663332fb50e2d
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:22e38d724a14e914d89c4169fd06e9b31db0e0e0016a1373810663332fb50e2d
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:22e38d724a14e914d89c4169fd06e9b31db0e0e0016a1373810663332fb50e2d].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 27m 37s
109 actionable tasks: 74 executed, 31 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/wfrcvabzp5vxm

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #39

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/39/display/redirect>

Changes:


------------------------------------------
[...truncated 68.54 KB...]
    result = func(*args)
  File "/usr/lib/python3.8/urllib/request.py", line 649, in http_error_default
    raise HTTPError(req.full_url, code, msg, hdrs, fp)
urllib.error.HTTPError: HTTP Error 502: Bad Gateway
ERROR:root:Invalid url for reflection-util-0.0.2: https://opensource.org/licenses/MIT. Retrying...
INFO:root:pull_licenses_java.py succeed. It took 111.159471 seconds with 16 threads.
Copying licenses from /home/jenkins/jenkins-slave/workspace/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/src/sdks/java/container/build/target/java_third_party_licenses to /home/jenkins/jenkins-slave/workspace/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/src/sdks/java/container/build/target/third_party_licenses.
Finished license_scripts.sh

> Task :sdks:java:container:java8:copyJavaThirdPartyLicenses
> Task :sdks:java:container:java8:dockerPrepare
> Task :sdks:java:container:java8:docker

> Task :runners:google-cloud-dataflow-java:buildAndPushDockerContainer
WARNING: `gcloud docker` will not be supported for Docker client versions above 18.03.

As an alternative, use `gcloud auth configure-docker` to configure `docker` to
use `gcloud` as a credential helper, then use `docker` as you would for non-GCR
registries, e.g. `docker pull gcr.io/project-id/my-image`. Add
`--verbosity=error` to silence this warning: `gcloud docker
--verbosity=error -- pull gcr.io/project-id/my-image`.

See: https://cloud.google.com/container-registry/docs/support/deprecation-notices#gcloud-docker

The push refers to repository [us.gcr.io/apache-beam-testing/java-postcommit-it/java]
183c3576fd2c: Preparing
afd4d3756e0c: Preparing
b7b4ee8b4da8: Preparing
06a4f5025888: Preparing
5809b69f2310: Preparing
fc0f1b30b1fc: Preparing
908890c9fd3f: Preparing
4509ebd42147: Preparing
25cc806cdb1e: Preparing
d259172796ae: Preparing
4988ae054215: Preparing
c5b41c87c38c: Preparing
13d2fdb0e4d5: Preparing
bff9fe6e429c: Preparing
7c245b2fe4f1: Preparing
f9e18e59a565: Preparing
26a504e63be4: Preparing
8bf42db0de72: Preparing
31892cc314cb: Preparing
11936051f93b: Preparing
25cc806cdb1e: Waiting
bff9fe6e429c: Waiting
8bf42db0de72: Waiting
7c245b2fe4f1: Waiting
31892cc314cb: Waiting
f9e18e59a565: Waiting
11936051f93b: Waiting
26a504e63be4: Waiting
d259172796ae: Waiting
4988ae054215: Waiting
fc0f1b30b1fc: Waiting
4509ebd42147: Waiting
13d2fdb0e4d5: Waiting
afd4d3756e0c: Pushed
5809b69f2310: Pushed
b7b4ee8b4da8: Pushed
183c3576fd2c: Pushed
fc0f1b30b1fc: Pushed
06a4f5025888: Pushed
4509ebd42147: Pushed
d259172796ae: Pushed
908890c9fd3f: Pushed
c5b41c87c38c: Pushed
bff9fe6e429c: Layer already exists
7c245b2fe4f1: Layer already exists
f9e18e59a565: Layer already exists
25cc806cdb1e: Pushed
8bf42db0de72: Layer already exists
31892cc314cb: Layer already exists
11936051f93b: Layer already exists
26a504e63be4: Layer already exists
13d2fdb0e4d5: Pushed
4988ae054215: Pushed
20220124123033: digest: sha256:56fe4a73910633660dbe846b5faa140abfed44c381e7934b675ab9e666e598f5 size: 4520

> Task :sdks:java:testing:load-tests:run
Jan 24, 2022 12:33:12 PM org.apache.beam.runners.dataflow.options.DataflowPipelineOptions$StagingLocationFactory create
INFO: No stagingLocation provided, falling back to gcpTempLocation
Jan 24, 2022 12:33:13 PM org.apache.beam.runners.dataflow.DataflowRunner fromOptions
INFO: PipelineOptions.filesToStage was not specified. Defaulting to files from the classpath: will stage 204 files. Enable logging at DEBUG level to see which files will be staged.
Jan 24, 2022 12:33:13 PM org.apache.beam.sdk.Pipeline validate
WARNING: The following transforms do not have stable unique names: Window.Into()
Jan 24, 2022 12:33:13 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Executing pipeline on the Dataflow Service, which will have billing implications related to Google Compute Engine usage and other Google Cloud Services.
Jan 24, 2022 12:33:17 PM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Uploading 204 files from PipelineOptions.filesToStage to staging location to prepare for execution.
Jan 24, 2022 12:33:17 PM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Staging files complete: 204 files cached, 0 files newly uploaded in 0 seconds
Jan 24, 2022 12:33:18 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Staging portable pipeline proto to gs://temp-storage-for-perf-tests/loadtests/staging/
Jan 24, 2022 12:33:18 PM org.apache.beam.runners.dataflow.util.PackageUtil tryStagePackage
INFO: Uploading <114075 bytes, hash 7dc98a057de4f9bc9ce7efd73d3df2e758230dcc5cedbef5a486d5470f2e3e04> to gs://temp-storage-for-perf-tests/loadtests/staging/pipeline-fcmKBX3k-byc5-_XPT3y51gjDcxc7b71pIbVRw8uPgQ.pb
Jan 24, 2022 12:33:19 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read input/DataflowRunner.StreamingUnboundedRead.ReadWithIds as step s1
Jan 24, 2022 12:33:19 PM org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource split
INFO: Split into 20 bundles of sizes: [org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@44b29496, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1947596f, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3078cac, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@f6de586, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5f2bd6d9, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7c847072, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@43d9f1a2, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@23f86d8a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2264ea32, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7d3c09ec, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@71e4b308, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@11900483, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@14a049f9, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@94e51e8, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5de6cf3a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4cc36c19, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5a3a1bf9, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2100d047, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4af45442, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1d805aa1]
Jan 24, 2022 12:33:19 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read input/StripIds as step s2
Jan 24, 2022 12:33:19 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect start time metrics (input) as step s3
Jan 24, 2022 12:33:19 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Window.Into()/Window.Assign as step s4
Jan 24, 2022 12:33:19 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read co-input/DataflowRunner.StreamingUnboundedRead.ReadWithIds as step s5
Jan 24, 2022 12:33:19 PM org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource split
INFO: Split into 20 bundles of sizes: [org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@441b8382, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1df1ced0, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5349b246, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@32b0876c, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2aaf152b, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@39bbd9e0, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@27fe9713, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@11c3ff67, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4397a639, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@24841372, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@77114efe, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@79a7bfbc, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@77f68df, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3e4e4c1, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7e7f3cfd, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3ae126d1, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@46a488c2, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6242ae3b, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@65ddee5a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@56399b9e]
Jan 24, 2022 12:33:19 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read co-input/StripIds as step s6
Jan 24, 2022 12:33:19 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect start time metrics (co-input) as step s7
Jan 24, 2022 12:33:19 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Window.Into()2/Window.Assign as step s8
Jan 24, 2022 12:33:19 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/MakeUnionTable0 as step s9
Jan 24, 2022 12:33:19 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/MakeUnionTable1 as step s10
Jan 24, 2022 12:33:19 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/Flatten as step s11
Jan 24, 2022 12:33:19 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/GBK as step s12
Jan 24, 2022 12:33:19 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/ConstructCoGbkResultFn as step s13
Jan 24, 2022 12:33:19 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Ungroup and reiterate as step s14
Jan 24, 2022 12:33:19 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect total bytes as step s15
Jan 24, 2022 12:33:19 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect end time metrics as step s16
Jan 24, 2022 12:33:20 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Dataflow SDK version: 2.37.0-SNAPSHOT
Jan 24, 2022 12:33:20 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To access the Dataflow monitoring console, please navigate to https://console.cloud.google.com/dataflow/jobs/us-central1/2022-01-24_04_33_20-14515115744488168432?project=apache-beam-testing
Jan 24, 2022 12:33:20 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Submitted job: 2022-01-24_04_33_20-14515115744488168432
Jan 24, 2022 12:33:20 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To cancel the job using the 'gcloud' tool, run:
> gcloud dataflow jobs --project=apache-beam-testing cancel --region=us-central1 2022-01-24_04_33_20-14515115744488168432
Jan 24, 2022 12:33:30 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
WARNING: 2022-01-24T12:33:27.261Z: The workflow name is not a valid Cloud Label. Labels applied to Cloud resources (such as GCE Instances) for monitoring will be labeled with this modified job name: load0tests0java170dataflow0v20streaming0cogbk01-jenkins-01-j6h5. For the best monitoring experience, please name your job with a valid Cloud Label. For details, see: https://cloud.google.com/compute/docs/labeling-resources#restrictions
Jan 24, 2022 12:33:33 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-24T12:33:32.929Z: Worker configuration: e2-standard-2 in us-central1-b.
Jan 24, 2022 12:33:34 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-24T12:33:33.686Z: Expanding SplittableParDo operations into optimizable parts.
Jan 24, 2022 12:33:34 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-24T12:33:33.714Z: Expanding CollectionToSingleton operations into optimizable parts.
Jan 24, 2022 12:33:34 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-24T12:33:33.771Z: Expanding CoGroupByKey operations into optimizable parts.
Jan 24, 2022 12:33:34 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-24T12:33:33.828Z: Expanding SplittableProcessKeyed operations into optimizable parts.
Jan 24, 2022 12:33:34 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-24T12:33:33.848Z: Expanding GroupByKey operations into streaming Read/Write steps
Jan 24, 2022 12:33:34 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-24T12:33:33.910Z: Lifting ValueCombiningMappingFns into MergeBucketsMappingFns
Jan 24, 2022 12:33:34 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-24T12:33:34.023Z: Fusing adjacent ParDo, Read, Write, and Flatten operations
Jan 24, 2022 12:33:34 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-24T12:33:34.049Z: Unzipping flatten CoGroupByKey-Flatten for input CoGroupByKey-MakeUnionTable0-ParMultiDo-ConstructUnionTable-.output
Jan 24, 2022 12:33:34 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-24T12:33:34.083Z: Fusing unzipped copy of CoGroupByKey/GBK/WriteStream, through flatten CoGroupByKey/Flatten, into producer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable)
Jan 24, 2022 12:33:34 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-24T12:33:34.105Z: Fusing consumer CoGroupByKey/GBK/WriteStream into CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable)
Jan 24, 2022 12:33:34 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-24T12:33:34.128Z: Fusing consumer Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read input/Impulse
Jan 24, 2022 12:33:34 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-24T12:33:34.150Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
Jan 24, 2022 12:33:34 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-24T12:33:34.184Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
Jan 24, 2022 12:33:34 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-24T12:33:34.218Z: Fusing consumer Read input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
Jan 24, 2022 12:33:34 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-24T12:33:34.249Z: Fusing consumer Collect start time metrics (input)/ParMultiDo(TimeMonitor) into Read input/ParDo(StripIds)/ParMultiDo(StripIds)
Jan 24, 2022 12:33:34 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-24T12:33:34.296Z: Fusing consumer Window.Into()/Window.Assign into Collect start time metrics (input)/ParMultiDo(TimeMonitor)
Jan 24, 2022 12:33:34 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-24T12:33:34.319Z: Fusing consumer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable) into Window.Into()/Window.Assign
Jan 24, 2022 12:33:34 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-24T12:33:34.344Z: Fusing consumer Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read co-input/Impulse
Jan 24, 2022 12:33:34 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-24T12:33:34.378Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
Jan 24, 2022 12:33:34 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-24T12:33:34.402Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
Jan 24, 2022 12:33:34 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-24T12:33:34.438Z: Fusing consumer Read co-input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
Jan 24, 2022 12:33:34 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-24T12:33:34.472Z: Fusing consumer Collect start time metrics (co-input)/ParMultiDo(TimeMonitor) into Read co-input/ParDo(StripIds)/ParMultiDo(StripIds)
Jan 24, 2022 12:33:34 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-24T12:33:34.507Z: Fusing consumer Window.Into()2/Window.Assign into Collect start time metrics (co-input)/ParMultiDo(TimeMonitor)
Jan 24, 2022 12:33:34 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-24T12:33:34.564Z: Fusing consumer CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable) into Window.Into()2/Window.Assign
Jan 24, 2022 12:33:34 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-24T12:33:34.596Z: Fusing consumer CoGroupByKey/GBK/MergeBuckets into CoGroupByKey/GBK/ReadStream
Jan 24, 2022 12:33:34 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-24T12:33:34.643Z: Fusing consumer CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult) into CoGroupByKey/GBK/MergeBuckets
Jan 24, 2022 12:33:34 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-24T12:33:34.668Z: Fusing consumer Ungroup and reiterate/ParMultiDo(UngroupAndReiterate) into CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult)
Jan 24, 2022 12:33:37 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-24T12:33:34.699Z: Fusing consumer Collect total bytes/ParMultiDo(ByteMonitor) into Ungroup and reiterate/ParMultiDo(UngroupAndReiterate)
Jan 24, 2022 12:33:37 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-24T12:33:34.734Z: Fusing consumer Collect end time metrics/ParMultiDo(TimeMonitor) into Collect total bytes/ParMultiDo(ByteMonitor)
Jan 24, 2022 12:33:37 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-24T12:33:35.096Z: Starting 5 ****s in us-central1-b...
Jan 24, 2022 12:33:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-24T12:33:39.965Z: Your project already contains 100 Dataflow-created metric descriptors, so new user metrics of the form custom.googleapis.com/* will not be created. However, all user metrics are also available in the metric dataflow.googleapis.com/job/user_counter. If you rely on the custom metrics, you can delete old / unused metric descriptors. See https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.list and https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.delete
Jan 24, 2022 12:34:25 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-24T12:34:24.820Z: Autoscaling: Raised the number of ****s to 5 so that the pipeline can catch up with its backlog and keep up with its input rate.
Jan 24, 2022 12:35:24 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-24T12:35:21.766Z: Workers have started successfully.
Jan 24, 2022 12:35:24 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-24T12:35:21.795Z: Workers have started successfully.
Build timed out (after 240 minutes). Marking the build as aborted.
FATAL: command execution failed
hudson.remoting.ChannelClosedException: Channel "hudson.remoting.Channel@5051cb59:apache-beam-jenkins-10": Remote call on apache-beam-jenkins-10 failed. The channel is closing down or has closed down
	at hudson.remoting.Channel.call(Channel.java:994)
	at hudson.remoting.RemoteInvocationHandler.invoke(RemoteInvocationHandler.java:286)
	at com.sun.proxy.$Proxy123.isAlive(Unknown Source)
	at hudson.Launcher$RemoteLauncher$ProcImpl.isAlive(Launcher.java:1211)
	at hudson.Launcher$RemoteLauncher$ProcImpl.join(Launcher.java:1203)
	at hudson.Launcher$ProcStarter.join(Launcher.java:523)
	at hudson.plugins.gradle.Gradle.perform(Gradle.java:317)
	at hudson.tasks.BuildStepMonitor$1.perform(BuildStepMonitor.java:20)
	at hudson.model.AbstractBuild$AbstractBuildExecution.perform(AbstractBuild.java:803)
	at hudson.model.Build$BuildExecution.build(Build.java:197)
	at hudson.model.Build$BuildExecution.doRun(Build.java:163)
	at hudson.model.AbstractBuild$AbstractBuildExecution.run(AbstractBuild.java:513)
	at hudson.model.Run.execute(Run.java:1906)
	at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:43)
	at hudson.model.ResourceController.execute(ResourceController.java:97)
	at hudson.model.Executor.run(Executor.java:429)
Caused by: java.io.IOException
	at hudson.remoting.Channel.close(Channel.java:1499)
	at hudson.remoting.Channel.close(Channel.java:1455)
	at hudson.slaves.SlaveComputer.closeChannel(SlaveComputer.java:884)
	at hudson.slaves.SlaveComputer.access$100(SlaveComputer.java:110)
	at hudson.slaves.SlaveComputer$2.run(SlaveComputer.java:765)
	at jenkins.util.ContextResettingExecutorService$1.run(ContextResettingExecutorService.java:28)
	at jenkins.security.ImpersonatingExecutorService$1.run(ImpersonatingExecutorService.java:68)
	at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
	at java.util.concurrent.FutureTask.run(FutureTask.java:266)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
	at java.lang.Thread.run(Thread.java:748)
Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure
ERROR: apache-beam-jenkins-10 is offline; cannot locate jdk_1.8_latest

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #38

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/38/display/redirect>

Changes:


------------------------------------------
[...truncated 50.06 KB...]
868a7c94ad8d: Pushed
dbfeed4855f9: Pushed
9e5c354ff101: Pushed
681c51acc5ca: Pushed
768b97a29a57: Pushed
bff9fe6e429c: Layer already exists
937b44b63b65: Pushed
7c245b2fe4f1: Layer already exists
f9e18e59a565: Layer already exists
26a504e63be4: Layer already exists
8bf42db0de72: Layer already exists
31892cc314cb: Layer already exists
11936051f93b: Layer already exists
a20fa729490e: Pushed
755908e7c341: Pushed
20220123123037: digest: sha256:fc4c7af26502b720c760bcd987ccb2771173a86fb4eab595d257b366040f7f08 size: 4520

> Task :sdks:java:testing:load-tests:run
Jan 23, 2022 12:32:20 PM org.apache.beam.runners.dataflow.options.DataflowPipelineOptions$StagingLocationFactory create
INFO: No stagingLocation provided, falling back to gcpTempLocation
Jan 23, 2022 12:32:21 PM org.apache.beam.runners.dataflow.DataflowRunner fromOptions
INFO: PipelineOptions.filesToStage was not specified. Defaulting to files from the classpath: will stage 204 files. Enable logging at DEBUG level to see which files will be staged.
Jan 23, 2022 12:32:21 PM org.apache.beam.sdk.Pipeline validate
WARNING: The following transforms do not have stable unique names: Window.Into()
Jan 23, 2022 12:32:21 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Executing pipeline on the Dataflow Service, which will have billing implications related to Google Compute Engine usage and other Google Cloud Services.
Jan 23, 2022 12:32:24 PM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Uploading 204 files from PipelineOptions.filesToStage to staging location to prepare for execution.
Jan 23, 2022 12:32:24 PM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Staging files complete: 204 files cached, 0 files newly uploaded in 0 seconds
Jan 23, 2022 12:32:24 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Staging portable pipeline proto to gs://temp-storage-for-perf-tests/loadtests/staging/
Jan 23, 2022 12:32:24 PM org.apache.beam.runners.dataflow.util.PackageUtil tryStagePackage
INFO: Uploading <114075 bytes, hash 89d32f994d5b825996520e0cca8f8e7fa4702aa7d4293876ff0a37a56dbda537> to gs://temp-storage-for-perf-tests/loadtests/staging/pipeline-idMvmU1bglmWUg4Myo-Of6RwKqfUKTh2_wo3pW29pTc.pb
Jan 23, 2022 12:32:26 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read input/DataflowRunner.StreamingUnboundedRead.ReadWithIds as step s1
Jan 23, 2022 12:32:26 PM org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource split
INFO: Split into 20 bundles of sizes: [org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1947596f, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3078cac, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@f6de586, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5f2bd6d9, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7c847072, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@43d9f1a2, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@23f86d8a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2264ea32, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7d3c09ec, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@71e4b308, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@11900483, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@14a049f9, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@94e51e8, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5de6cf3a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4cc36c19, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5a3a1bf9, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2100d047, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4af45442, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1d805aa1, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@176f7f3b]
Jan 23, 2022 12:32:26 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read input/StripIds as step s2
Jan 23, 2022 12:32:26 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect start time metrics (input) as step s3
Jan 23, 2022 12:32:26 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Window.Into()/Window.Assign as step s4
Jan 23, 2022 12:32:26 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read co-input/DataflowRunner.StreamingUnboundedRead.ReadWithIds as step s5
Jan 23, 2022 12:32:26 PM org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource split
INFO: Split into 20 bundles of sizes: [org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1df1ced0, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5349b246, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@32b0876c, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2aaf152b, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@39bbd9e0, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@27fe9713, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@11c3ff67, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4397a639, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@24841372, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@77114efe, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@79a7bfbc, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@77f68df, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3e4e4c1, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7e7f3cfd, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3ae126d1, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@46a488c2, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6242ae3b, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@65ddee5a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@56399b9e, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@34b9eb03]
Jan 23, 2022 12:32:26 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read co-input/StripIds as step s6
Jan 23, 2022 12:32:26 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect start time metrics (co-input) as step s7
Jan 23, 2022 12:32:26 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Window.Into()2/Window.Assign as step s8
Jan 23, 2022 12:32:26 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/MakeUnionTable0 as step s9
Jan 23, 2022 12:32:26 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/MakeUnionTable1 as step s10
Jan 23, 2022 12:32:26 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/Flatten as step s11
Jan 23, 2022 12:32:26 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/GBK as step s12
Jan 23, 2022 12:32:26 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/ConstructCoGbkResultFn as step s13
Jan 23, 2022 12:32:26 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Ungroup and reiterate as step s14
Jan 23, 2022 12:32:26 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect total bytes as step s15
Jan 23, 2022 12:32:26 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect end time metrics as step s16
Jan 23, 2022 12:32:26 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Dataflow SDK version: 2.37.0-SNAPSHOT
Jan 23, 2022 12:32:27 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To access the Dataflow monitoring console, please navigate to https://console.cloud.google.com/dataflow/jobs/us-central1/2022-01-23_04_32_26-9038799913148666900?project=apache-beam-testing
Jan 23, 2022 12:32:27 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Submitted job: 2022-01-23_04_32_26-9038799913148666900
Jan 23, 2022 12:32:27 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To cancel the job using the 'gcloud' tool, run:
> gcloud dataflow jobs --project=apache-beam-testing cancel --region=us-central1 2022-01-23_04_32_26-9038799913148666900
Jan 23, 2022 12:32:35 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
WARNING: 2022-01-23T12:32:34.076Z: The workflow name is not a valid Cloud Label. Labels applied to Cloud resources (such as GCE Instances) for monitoring will be labeled with this modified job name: load0tests0java170dataflow0v20streaming0cogbk01-jenkins-01-g5ap. For the best monitoring experience, please name your job with a valid Cloud Label. For details, see: https://cloud.google.com/compute/docs/labeling-resources#restrictions
Jan 23, 2022 12:32:39 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-23T12:32:39.002Z: Worker configuration: e2-standard-2 in us-central1-b.
Jan 23, 2022 12:32:39 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-23T12:32:39.643Z: Expanding SplittableParDo operations into optimizable parts.
Jan 23, 2022 12:32:39 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-23T12:32:39.681Z: Expanding CollectionToSingleton operations into optimizable parts.
Jan 23, 2022 12:32:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-23T12:32:39.737Z: Expanding CoGroupByKey operations into optimizable parts.
Jan 23, 2022 12:32:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-23T12:32:39.804Z: Expanding SplittableProcessKeyed operations into optimizable parts.
Jan 23, 2022 12:32:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-23T12:32:39.830Z: Expanding GroupByKey operations into streaming Read/Write steps
Jan 23, 2022 12:32:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-23T12:32:39.896Z: Lifting ValueCombiningMappingFns into MergeBucketsMappingFns
Jan 23, 2022 12:32:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-23T12:32:40.033Z: Fusing adjacent ParDo, Read, Write, and Flatten operations
Jan 23, 2022 12:32:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-23T12:32:40.062Z: Unzipping flatten CoGroupByKey-Flatten for input CoGroupByKey-MakeUnionTable0-ParMultiDo-ConstructUnionTable-.output
Jan 23, 2022 12:32:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-23T12:32:40.082Z: Fusing unzipped copy of CoGroupByKey/GBK/WriteStream, through flatten CoGroupByKey/Flatten, into producer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable)
Jan 23, 2022 12:32:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-23T12:32:40.127Z: Fusing consumer CoGroupByKey/GBK/WriteStream into CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable)
Jan 23, 2022 12:32:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-23T12:32:40.161Z: Fusing consumer Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read input/Impulse
Jan 23, 2022 12:32:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-23T12:32:40.191Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
Jan 23, 2022 12:32:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-23T12:32:40.213Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
Jan 23, 2022 12:32:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-23T12:32:40.235Z: Fusing consumer Read input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
Jan 23, 2022 12:32:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-23T12:32:40.264Z: Fusing consumer Collect start time metrics (input)/ParMultiDo(TimeMonitor) into Read input/ParDo(StripIds)/ParMultiDo(StripIds)
Jan 23, 2022 12:32:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-23T12:32:40.287Z: Fusing consumer Window.Into()/Window.Assign into Collect start time metrics (input)/ParMultiDo(TimeMonitor)
Jan 23, 2022 12:32:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-23T12:32:40.317Z: Fusing consumer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable) into Window.Into()/Window.Assign
Jan 23, 2022 12:32:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-23T12:32:40.343Z: Fusing consumer Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read co-input/Impulse
Jan 23, 2022 12:32:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-23T12:32:40.379Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
Jan 23, 2022 12:32:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-23T12:32:40.410Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
Jan 23, 2022 12:32:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-23T12:32:40.431Z: Fusing consumer Read co-input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
Jan 23, 2022 12:32:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-23T12:32:40.494Z: Fusing consumer Collect start time metrics (co-input)/ParMultiDo(TimeMonitor) into Read co-input/ParDo(StripIds)/ParMultiDo(StripIds)
Jan 23, 2022 12:32:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-23T12:32:40.570Z: Fusing consumer Window.Into()2/Window.Assign into Collect start time metrics (co-input)/ParMultiDo(TimeMonitor)
Jan 23, 2022 12:32:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-23T12:32:40.648Z: Fusing consumer CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable) into Window.Into()2/Window.Assign
Jan 23, 2022 12:32:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-23T12:32:40.796Z: Fusing consumer CoGroupByKey/GBK/MergeBuckets into CoGroupByKey/GBK/ReadStream
Jan 23, 2022 12:32:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-23T12:32:40.881Z: Fusing consumer CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult) into CoGroupByKey/GBK/MergeBuckets
Jan 23, 2022 12:32:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-23T12:32:40.916Z: Fusing consumer Ungroup and reiterate/ParMultiDo(UngroupAndReiterate) into CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult)
Jan 23, 2022 12:32:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-23T12:32:40.956Z: Fusing consumer Collect total bytes/ParMultiDo(ByteMonitor) into Ungroup and reiterate/ParMultiDo(UngroupAndReiterate)
Jan 23, 2022 12:32:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-23T12:32:41.004Z: Fusing consumer Collect end time metrics/ParMultiDo(TimeMonitor) into Collect total bytes/ParMultiDo(ByteMonitor)
Jan 23, 2022 12:32:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-23T12:32:41.689Z: Starting 5 ****s in us-central1-b...
Jan 23, 2022 12:33:03 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-23T12:33:02.475Z: Your project already contains 100 Dataflow-created metric descriptors, so new user metrics of the form custom.googleapis.com/* will not be created. However, all user metrics are also available in the metric dataflow.googleapis.com/job/user_counter. If you rely on the custom metrics, you can delete old / unused metric descriptors. See https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.list and https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.delete
Jan 23, 2022 12:33:38 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-23T12:33:38.126Z: Autoscaling: Raised the number of ****s to 5 so that the pipeline can catch up with its backlog and keep up with its input rate.
Jan 23, 2022 12:34:34 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-23T12:34:33.311Z: Workers have started successfully.
Jan 23, 2022 12:34:34 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-23T12:34:33.337Z: Workers have started successfully.
Jan 23, 2022 4:00:35 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-23T16:00:34.233Z: Cancel request is committed for workflow job: 2022-01-23_04_32_26-9038799913148666900.
Jan 23, 2022 4:00:35 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-23T16:00:34.402Z: Cleaning up.
Jan 23, 2022 4:00:35 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-23T16:00:34.487Z: Stopping **** pool...
Jan 23, 2022 4:00:35 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-23T16:00:34.543Z: Stopping **** pool...
Jan 23, 2022 4:02:53 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-23T16:02:52.408Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Jan 23, 2022 4:02:53 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-23T16:02:52.455Z: Worker pool stopped.
Jan 23, 2022 4:02:58 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-01-23_04_32_26-9038799913148666900 finished with status CANCELLED.
Load test results for test (ID): b7101405-6336-4671-a761-0fe3261aac26 and timestamp: 2022-01-23T12:32:21.535000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                  12334.62
dataflow_v2_java17_total_bytes_count             1.98486289E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220123123037
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:fc4c7af26502b720c760bcd987ccb2771173a86fb4eab595d257b366040f7f08
Deleted: sha256:727a71a76dd9518995235b34a284d5812c150342e5a5a0690812a0383ea99ccf
Deleted: sha256:6cbb8d330c7bcd8e51c3a7948199f939a4a130b39888dae5326cb3d9c531e574
Deleted: sha256:4912631cddf6ce96bc0ad5956db77d8c9b6626c7619edba015a626e7a825aede
Deleted: sha256:d87d1d2fe59f78f4a00b1f9dd262ba8a69dbdba7ff216ee5e041fa05893de1c6
Deleted: sha256:9afa7111b6449cdc996225dc650e6ee7d807cdaf65f5e805dcc12009b68a64ed
Deleted: sha256:459d1efecf41277748c6961c1f93bbe39dc67f86b67110e36d48873f6c16fe6d
Deleted: sha256:fe5a14476f7c71a1e129b142836044d122dc79f216f7f6662d497c2263bc0d31
Deleted: sha256:ed6364b1f2b8142105234ff9de0013a2f3ed5899966e842fd48ca56ab3927ed6
Deleted: sha256:98c6f47514df4b2de6487f1dd3fbf3645563ec0576092089eba920c24b46dd11
Deleted: sha256:ca7f642dbed875cd1df64dad505cec72ca655554696f9230737ee87bea3bd662
Deleted: sha256:bd8af7c516855dd49d6bd13948fea9b1c77e60fe55ca79ff42df5de326976cb9
Deleted: sha256:2fbdbf030b9ddaf2d80a80a8f4701a138cb32a055b369ee5e163850bbcd4fd12
Deleted: sha256:159c4fd249d9c6ccfc7c665e1ed479e0cfdbbef1e8f01bd0bc60feba75d623cb
Deleted: sha256:942a21eef5ab8e6a0a2618a30936460187419b77342f39ae7051d93808c693fc
Deleted: sha256:de28582ab02dc282b7d419fd7aa22cd3a1d790d505e3c8a940c7fee2a23c6708
Deleted: sha256:a838be439433b0919c572fe5ce7ccb01d2d630282c349050fdb546457c080a03
Deleted: sha256:71aa4d13f0ab1401f87064b7a73faad25c630c20a4e62f797ed0515064d6221d
Deleted: sha256:023733541687e2d70237f4db197eb4a47cbcbb0392d8face76aa5c09564f56fd
Deleted: sha256:9a533ff28ca6047891efd0dcfa2c6787f9b6d34da5d38e8e4b019d5139b62c5c
Deleted: sha256:02aa2b0a391b2d5d4d4258d27e469951cb374eda005a5f359e96702790801a7d
Deleted: sha256:4dd5d9c74d045bf5006413016edf02e63bf7d09011fd91343b83645ccdf13626
Deleted: sha256:9c394dc6fbfa15ba014712134df3967a0ebe110d0b3c20a931076c3a2e0ed349
Deleted: sha256:84e3ec266131300d9f532aff04197c8bea384433350fa87e8498d888bbae02ee
Deleted: sha256:7aba1bd69f9f47a509ce12712786b65d912b845dc21d3d7eaf62962c85c8a720
Deleted: sha256:49ae19aa4cf0dc5e694d89d785dec102cd46fcd1283485e9a5b44f2342635b7d
Deleted: sha256:1532fbadfe052f9fab7c2d814bb1c9c5f4e2e45856fa84af904c7daea0ee7c41
Deleted: sha256:bcb58fbe4d940f175e1c15bd2c6697f8d2b5fd724a0a532db0f830d927c94992
Deleted: sha256:904a7986d34e6b84715c0e0bd5e08253a25bcd5994f4aed9040b859daafacfcc
Deleted: sha256:c95d702b507f77b2fe860888a3a093e3200e78b324186764e6a8b408edb1fb94
Deleted: sha256:0c6a15ef5450170539547a4abbc371141453929ad3476d9c6f83e447f50f7004
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220123123037]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:fc4c7af26502b720c760bcd987ccb2771173a86fb4eab595d257b366040f7f08]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220123123037] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:fc4c7af26502b720c760bcd987ccb2771173a86fb4eab595d257b366040f7f08])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:fc4c7af26502b720c760bcd987ccb2771173a86fb4eab595d257b366040f7f08
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:fc4c7af26502b720c760bcd987ccb2771173a86fb4eab595d257b366040f7f08
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:fc4c7af26502b720c760bcd987ccb2771173a86fb4eab595d257b366040f7f08].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 32m 39s
109 actionable tasks: 72 executed, 33 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/gf5museauleiq

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #37

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/37/display/redirect?page=changes>

Changes:

[ningkang0957] [BEAM-13687] Improved Spanner IO request count metrics

[noreply] [BEAM-10206] Add key for fields in wrapper (#16583)

[noreply] Merge pull request #16530 from Adding JSON support in SpannerIO and

[noreply] [BEAM-13685] Enable users to specify cache directory under Interactive


------------------------------------------
[...truncated 50.66 KB...]
31892cc314cb: Waiting
11936051f93b: Waiting
d1e988f29d0d: Waiting
26a504e63be4: Waiting
36271400b172: Waiting
f9e18e59a565: Waiting
feba03ab4d2d: Waiting
6da597f90efd: Waiting
62dd6ff2b6cd: Pushed
9e4975d2a80b: Pushed
7cf3badcb976: Pushed
f50a0d06154d: Pushed
e944fa471b21: Pushed
4892d80a24b6: Pushed
36271400b172: Pushed
33b824f2cf83: Pushed
7c5e550725ea: Pushed
bff9fe6e429c: Layer already exists
7c245b2fe4f1: Layer already exists
f9e18e59a565: Layer already exists
0398f72b2f5e: Pushed
feba03ab4d2d: Pushed
26a504e63be4: Layer already exists
31892cc314cb: Layer already exists
8bf42db0de72: Layer already exists
11936051f93b: Layer already exists
d1e988f29d0d: Pushed
6da597f90efd: Pushed
20220122123039: digest: sha256:45adc53f21ee28e2a8d2fb954d5eba8c7995f713678cacb6d7567b68943f8d3a size: 4520

> Task :sdks:java:testing:load-tests:run
Jan 22, 2022 12:32:27 PM org.apache.beam.runners.dataflow.options.DataflowPipelineOptions$StagingLocationFactory create
INFO: No stagingLocation provided, falling back to gcpTempLocation
Jan 22, 2022 12:32:28 PM org.apache.beam.runners.dataflow.DataflowRunner fromOptions
INFO: PipelineOptions.filesToStage was not specified. Defaulting to files from the classpath: will stage 204 files. Enable logging at DEBUG level to see which files will be staged.
Jan 22, 2022 12:32:29 PM org.apache.beam.sdk.Pipeline validate
WARNING: The following transforms do not have stable unique names: Window.Into()
Jan 22, 2022 12:32:29 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Executing pipeline on the Dataflow Service, which will have billing implications related to Google Compute Engine usage and other Google Cloud Services.
Jan 22, 2022 12:32:31 PM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Uploading 204 files from PipelineOptions.filesToStage to staging location to prepare for execution.
Jan 22, 2022 12:32:32 PM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Staging files complete: 204 files cached, 0 files newly uploaded in 0 seconds
Jan 22, 2022 12:32:32 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Staging portable pipeline proto to gs://temp-storage-for-perf-tests/loadtests/staging/
Jan 22, 2022 12:32:32 PM org.apache.beam.runners.dataflow.util.PackageUtil tryStagePackage
INFO: Uploading <114075 bytes, hash 8c34e24446d61ac514531847b2356cd110e8acbb2593bcee3bf6748c809d56f4> to gs://temp-storage-for-perf-tests/loadtests/staging/pipeline-jDTiREbWGsUUUxhHsjVs0RDorLslk7zuO_Z0jICdVvQ.pb
Jan 22, 2022 12:32:34 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read input/DataflowRunner.StreamingUnboundedRead.ReadWithIds as step s1
Jan 22, 2022 12:32:34 PM org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource split
INFO: Split into 20 bundles of sizes: [org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1947596f, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3078cac, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@f6de586, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5f2bd6d9, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7c847072, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@43d9f1a2, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@23f86d8a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2264ea32, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7d3c09ec, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@71e4b308, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@11900483, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@14a049f9, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@94e51e8, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5de6cf3a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4cc36c19, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5a3a1bf9, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2100d047, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4af45442, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1d805aa1, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@176f7f3b]
Jan 22, 2022 12:32:34 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read input/StripIds as step s2
Jan 22, 2022 12:32:34 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect start time metrics (input) as step s3
Jan 22, 2022 12:32:34 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Window.Into()/Window.Assign as step s4
Jan 22, 2022 12:32:34 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read co-input/DataflowRunner.StreamingUnboundedRead.ReadWithIds as step s5
Jan 22, 2022 12:32:34 PM org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource split
INFO: Split into 20 bundles of sizes: [org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1df1ced0, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5349b246, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@32b0876c, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2aaf152b, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@39bbd9e0, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@27fe9713, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@11c3ff67, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4397a639, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@24841372, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@77114efe, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@79a7bfbc, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@77f68df, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3e4e4c1, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7e7f3cfd, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3ae126d1, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@46a488c2, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6242ae3b, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@65ddee5a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@56399b9e, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@34b9eb03]
Jan 22, 2022 12:32:34 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read co-input/StripIds as step s6
Jan 22, 2022 12:32:34 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect start time metrics (co-input) as step s7
Jan 22, 2022 12:32:34 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Window.Into()2/Window.Assign as step s8
Jan 22, 2022 12:32:34 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/MakeUnionTable0 as step s9
Jan 22, 2022 12:32:34 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/MakeUnionTable1 as step s10
Jan 22, 2022 12:32:34 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/Flatten as step s11
Jan 22, 2022 12:32:34 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/GBK as step s12
Jan 22, 2022 12:32:34 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/ConstructCoGbkResultFn as step s13
Jan 22, 2022 12:32:34 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Ungroup and reiterate as step s14
Jan 22, 2022 12:32:34 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect total bytes as step s15
Jan 22, 2022 12:32:34 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect end time metrics as step s16
Jan 22, 2022 12:32:34 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Dataflow SDK version: 2.37.0-SNAPSHOT
Jan 22, 2022 12:32:35 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To access the Dataflow monitoring console, please navigate to https://console.cloud.google.com/dataflow/jobs/us-central1/2022-01-22_04_32_34-9402087514964833050?project=apache-beam-testing
Jan 22, 2022 12:32:35 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Submitted job: 2022-01-22_04_32_34-9402087514964833050
Jan 22, 2022 12:32:35 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To cancel the job using the 'gcloud' tool, run:
> gcloud dataflow jobs --project=apache-beam-testing cancel --region=us-central1 2022-01-22_04_32_34-9402087514964833050
Jan 22, 2022 12:32:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
WARNING: 2022-01-22T12:32:42.605Z: The workflow name is not a valid Cloud Label. Labels applied to Cloud resources (such as GCE Instances) for monitoring will be labeled with this modified job name: load0tests0java170dataflow0v20streaming0cogbk01-jenkins-01-x3sc. For the best monitoring experience, please name your job with a valid Cloud Label. For details, see: https://cloud.google.com/compute/docs/labeling-resources#restrictions
Jan 22, 2022 12:32:47 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-22T12:32:47.629Z: Worker configuration: e2-standard-2 in us-central1-b.
Jan 22, 2022 12:32:50 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-22T12:32:48.322Z: Expanding SplittableParDo operations into optimizable parts.
Jan 22, 2022 12:32:50 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-22T12:32:48.346Z: Expanding CollectionToSingleton operations into optimizable parts.
Jan 22, 2022 12:32:50 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-22T12:32:48.428Z: Expanding CoGroupByKey operations into optimizable parts.
Jan 22, 2022 12:32:50 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-22T12:32:48.496Z: Expanding SplittableProcessKeyed operations into optimizable parts.
Jan 22, 2022 12:32:50 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-22T12:32:48.521Z: Expanding GroupByKey operations into streaming Read/Write steps
Jan 22, 2022 12:32:50 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-22T12:32:48.598Z: Lifting ValueCombiningMappingFns into MergeBucketsMappingFns
Jan 22, 2022 12:32:50 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-22T12:32:48.690Z: Fusing adjacent ParDo, Read, Write, and Flatten operations
Jan 22, 2022 12:32:50 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-22T12:32:48.724Z: Unzipping flatten CoGroupByKey-Flatten for input CoGroupByKey-MakeUnionTable0-ParMultiDo-ConstructUnionTable-.output
Jan 22, 2022 12:32:50 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-22T12:32:48.753Z: Fusing unzipped copy of CoGroupByKey/GBK/WriteStream, through flatten CoGroupByKey/Flatten, into producer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable)
Jan 22, 2022 12:32:50 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-22T12:32:48.801Z: Fusing consumer CoGroupByKey/GBK/WriteStream into CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable)
Jan 22, 2022 12:32:50 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-22T12:32:48.875Z: Fusing consumer Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read input/Impulse
Jan 22, 2022 12:32:50 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-22T12:32:48.916Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
Jan 22, 2022 12:32:50 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-22T12:32:48.953Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
Jan 22, 2022 12:32:50 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-22T12:32:48.984Z: Fusing consumer Read input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
Jan 22, 2022 12:32:50 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-22T12:32:49.019Z: Fusing consumer Collect start time metrics (input)/ParMultiDo(TimeMonitor) into Read input/ParDo(StripIds)/ParMultiDo(StripIds)
Jan 22, 2022 12:32:50 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-22T12:32:49.090Z: Fusing consumer Window.Into()/Window.Assign into Collect start time metrics (input)/ParMultiDo(TimeMonitor)
Jan 22, 2022 12:32:50 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-22T12:32:49.118Z: Fusing consumer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable) into Window.Into()/Window.Assign
Jan 22, 2022 12:32:50 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-22T12:32:49.150Z: Fusing consumer Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read co-input/Impulse
Jan 22, 2022 12:32:50 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-22T12:32:49.182Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
Jan 22, 2022 12:32:50 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-22T12:32:49.212Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
Jan 22, 2022 12:32:50 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-22T12:32:49.242Z: Fusing consumer Read co-input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
Jan 22, 2022 12:32:50 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-22T12:32:49.276Z: Fusing consumer Collect start time metrics (co-input)/ParMultiDo(TimeMonitor) into Read co-input/ParDo(StripIds)/ParMultiDo(StripIds)
Jan 22, 2022 12:32:50 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-22T12:32:49.311Z: Fusing consumer Window.Into()2/Window.Assign into Collect start time metrics (co-input)/ParMultiDo(TimeMonitor)
Jan 22, 2022 12:32:50 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-22T12:32:49.334Z: Fusing consumer CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable) into Window.Into()2/Window.Assign
Jan 22, 2022 12:32:50 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-22T12:32:49.376Z: Fusing consumer CoGroupByKey/GBK/MergeBuckets into CoGroupByKey/GBK/ReadStream
Jan 22, 2022 12:32:50 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-22T12:32:49.398Z: Fusing consumer CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult) into CoGroupByKey/GBK/MergeBuckets
Jan 22, 2022 12:32:50 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-22T12:32:49.423Z: Fusing consumer Ungroup and reiterate/ParMultiDo(UngroupAndReiterate) into CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult)
Jan 22, 2022 12:32:50 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-22T12:32:49.453Z: Fusing consumer Collect total bytes/ParMultiDo(ByteMonitor) into Ungroup and reiterate/ParMultiDo(UngroupAndReiterate)
Jan 22, 2022 12:32:50 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-22T12:32:49.488Z: Fusing consumer Collect end time metrics/ParMultiDo(TimeMonitor) into Collect total bytes/ParMultiDo(ByteMonitor)
Jan 22, 2022 12:32:50 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-22T12:32:49.899Z: Starting 5 ****s in us-central1-b...
Jan 22, 2022 12:33:02 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-22T12:33:02.182Z: Your project already contains 100 Dataflow-created metric descriptors, so new user metrics of the form custom.googleapis.com/* will not be created. However, all user metrics are also available in the metric dataflow.googleapis.com/job/user_counter. If you rely on the custom metrics, you can delete old / unused metric descriptors. See https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.list and https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.delete
Jan 22, 2022 12:33:47 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-22T12:33:47.615Z: Autoscaling: Raised the number of ****s to 5 so that the pipeline can catch up with its backlog and keep up with its input rate.
Jan 22, 2022 12:34:45 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-22T12:34:44.127Z: Workers have started successfully.
Jan 22, 2022 12:34:45 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-22T12:34:44.168Z: Workers have started successfully.
Jan 22, 2022 4:00:40 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-22T16:00:39.139Z: Cancel request is committed for workflow job: 2022-01-22_04_32_34-9402087514964833050.
Jan 22, 2022 4:00:40 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-22T16:00:39.216Z: Cleaning up.
Jan 22, 2022 4:00:40 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-22T16:00:39.284Z: Stopping **** pool...
Jan 22, 2022 4:00:40 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-22T16:00:39.344Z: Stopping **** pool...
Jan 22, 2022 4:03:08 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-22T16:03:07.557Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Jan 22, 2022 4:03:08 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-22T16:03:07.623Z: Worker pool stopped.
Jan 22, 2022 4:03:14 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-01-22_04_32_34-9402087514964833050 finished with status CANCELLED.
Load test results for test (ID): 0e6c4fbf-2db8-4b79-a06b-c092cf66cc82 and timestamp: 2022-01-22T12:32:28.784000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                 12334.413
dataflow_v2_java17_total_bytes_count             2.47191095E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220122123039
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:45adc53f21ee28e2a8d2fb954d5eba8c7995f713678cacb6d7567b68943f8d3a
Deleted: sha256:182784b0d8924aab3fd05113a2fde58bc09f8ee4d378ebf48e80350b75021b13
Deleted: sha256:cff79d5bfba0e6aed0c1441b98836d0faecf8451bd4f2cd0aab5ff4dc7a5008b
Deleted: sha256:6f0db8d3e0999877a306f2f72709af1e87c7c893f7e1495feac2b19414e7adaf
Deleted: sha256:7650a5bb4cf3d1b0d61a8ac347562385f4e531b5d91cc8949143b833e4ad929f
Deleted: sha256:f796dd58965c46cc99210085eda11971528e4b1978a7e9f4842dc8ff9e56eabb
Deleted: sha256:74ef671ae207c930c9281b1e5abfb1dc528e74672d4c1a44178ac5f8ddcf56d5
Deleted: sha256:6bd57acf0744a22591e1b0fbe75e41a84abe21496c9bb7e07db965b9e4580db8
Deleted: sha256:a41794fc8730053019911856bfe5e5d47843c59615ec01c26a0cdafb4df3ace5
Deleted: sha256:6754d5124e70436c40b19d487dee20ee6a969e6f04335ed07e2fd4c4d5114656
Deleted: sha256:7071bfe5de4690098d6b94f59a15c9131bed457ff242135a782380d9d1a95f23
Deleted: sha256:c9a6df0ba9dafa80c58850a43c9133d723f621df1973e8e9e5448eca8a37358d
Deleted: sha256:237c15525062c3c60e02391c4f4f0221f832b31a6434ce0d6391242a87f46d1b
Deleted: sha256:46ea16c71a3dd500f80aaf68bf6a5af4f6d23ec9539ee3382db590905e2d13b9
Deleted: sha256:0bc95ff58146f291a6b60182f22f2c400f108a1b9f8ad1f4d35ed96cdc1e1df3
Deleted: sha256:f482ee6972e99e07185166fd989748dcbe8df1cde874bb82305338b6950689d5
Deleted: sha256:69f529cb3dbaf93fb69be9bf1e1404314c2c0ea62f3730ccc9113e0fe50df414
Deleted: sha256:35b4e406598b62e3196d8fe9df985b9a7692f3fed2ecd6acd56822897225c764
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220122123039]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:45adc53f21ee28e2a8d2fb954d5eba8c7995f713678cacb6d7567b68943f8d3a]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220122123039] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:45adc53f21ee28e2a8d2fb954d5eba8c7995f713678cacb6d7567b68943f8d3a])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:45adc53f21ee28e2a8d2fb954d5eba8c7995f713678cacb6d7567b68943f8d3a
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:45adc53f21ee28e2a8d2fb954d5eba8c7995f713678cacb6d7567b68943f8d3a
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:45adc53f21ee28e2a8d2fb954d5eba8c7995f713678cacb6d7567b68943f8d3a].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 32m 53s
109 actionable tasks: 73 executed, 32 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/gf7os53yj753q

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #36

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/36/display/redirect?page=changes>

Changes:

[thiagotnunes] [BEAM-12164]: Add SDF for reading change stream records

[tuyarer] [BEAM-13577] Beam Select's uniquifyNames function loses nullability of

[sergey.kalinin] Update GH Actions to use proper variables names and proper triggers

[dhuntsperger] edited README and comments in Python multi-lang pipes examples

[Pablo Estrada] Revert "Merge pull request #15863 from [BEAM-13184] Autosharding for

[Pablo Estrada] BEAM-13611 reactivating jdbcio xlang test

[Steve Niemitz] [BEAM-13689] Output token elements when BQ batch writes complete.

[noreply] Merge pull request #16371 from [BEAM-13518][Playground] Beam Playground

[noreply] Update Java FnAPI beam master (#16572)

[noreply] [BEAM-13699] Replace fnv with maphash. (#16573)

[noreply] [BEAM-13693] Bump

[noreply] [BEAM-10206] Remove Fatalf calls in non-test goroutines for

[noreply] [BEAM-13430] Re-add provided configuration (#16552)

[noreply] Merge pull request #16540 from [BEAM-13678][Playground]Update Github

[noreply] Merge pull request #16546 from [BEAM-13661] [BEAM-13704] [Playground]

[noreply] Merge pull request #16369 from [BEAM-13558] [Playground] Hide the Graph


------------------------------------------
[...truncated 49.34 KB...]
84d00813a7f3: Preparing
5c56e14d73e3: Preparing
a036ba1fc8e4: Preparing
bff9fe6e429c: Preparing
7c245b2fe4f1: Preparing
f9e18e59a565: Preparing
26a504e63be4: Preparing
8bf42db0de72: Preparing
31892cc314cb: Preparing
11936051f93b: Preparing
eb127d4ed1e5: Waiting
a70778495779: Waiting
a036ba1fc8e4: Waiting
bff9fe6e429c: Waiting
7c245b2fe4f1: Waiting
26a504e63be4: Waiting
8bf42db0de72: Waiting
f9e18e59a565: Waiting
11936051f93b: Waiting
31892cc314cb: Waiting
5c56e14d73e3: Waiting
1e513f38b2d9: Waiting
923be1818e17: Waiting
b8e7120e7b25: Waiting
84d00813a7f3: Waiting
f93e03466565: Pushed
9b85185bbb7b: Pushed
2e01697d9592: Pushed
b7ad33ee5f20: Pushed
6d2bb8725e98: Pushed
b8e7120e7b25: Pushed
eb127d4ed1e5: Pushed
1e513f38b2d9: Pushed
923be1818e17: Pushed
bff9fe6e429c: Layer already exists
7c245b2fe4f1: Layer already exists
f9e18e59a565: Layer already exists
26a504e63be4: Layer already exists
8bf42db0de72: Layer already exists
31892cc314cb: Layer already exists
11936051f93b: Layer already exists
5c56e14d73e3: Pushed
a036ba1fc8e4: Pushed
a70778495779: Pushed
84d00813a7f3: Pushed
20220121123024: digest: sha256:fadf94325b1285d8bd7f2add7cebcacf154139134e4bea9e1cab266eeeb4b64b size: 4520

> Task :sdks:java:testing:load-tests:run
Jan 21, 2022 12:32:23 PM org.apache.beam.runners.dataflow.options.DataflowPipelineOptions$StagingLocationFactory create
INFO: No stagingLocation provided, falling back to gcpTempLocation
Jan 21, 2022 12:32:23 PM org.apache.beam.runners.dataflow.DataflowRunner fromOptions
INFO: PipelineOptions.filesToStage was not specified. Defaulting to files from the classpath: will stage 204 files. Enable logging at DEBUG level to see which files will be staged.
Jan 21, 2022 12:32:24 PM org.apache.beam.sdk.Pipeline validate
WARNING: The following transforms do not have stable unique names: Window.Into()
Jan 21, 2022 12:32:24 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Executing pipeline on the Dataflow Service, which will have billing implications related to Google Compute Engine usage and other Google Cloud Services.
Jan 21, 2022 12:32:26 PM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Uploading 204 files from PipelineOptions.filesToStage to staging location to prepare for execution.
Jan 21, 2022 12:32:27 PM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Staging files complete: 204 files cached, 0 files newly uploaded in 0 seconds
Jan 21, 2022 12:32:27 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Staging portable pipeline proto to gs://temp-storage-for-perf-tests/loadtests/staging/
Jan 21, 2022 12:32:27 PM org.apache.beam.runners.dataflow.util.PackageUtil tryStagePackage
INFO: Uploading <114075 bytes, hash 3d28eb47e0585c21b53c2096d49ef47cc82460b396693696d44eddf22508953c> to gs://temp-storage-for-perf-tests/loadtests/staging/pipeline-PSjrR-BYXCG1PCCW1J70fMgkYLOWaTaW1E7d8iUIlTw.pb
Jan 21, 2022 12:32:29 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read input/DataflowRunner.StreamingUnboundedRead.ReadWithIds as step s1
Jan 21, 2022 12:32:29 PM org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource split
INFO: Split into 20 bundles of sizes: [org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6d5c2745, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@44b29496, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1947596f, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3078cac, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@f6de586, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5f2bd6d9, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7c847072, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@43d9f1a2, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@23f86d8a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2264ea32, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7d3c09ec, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@71e4b308, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@11900483, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@14a049f9, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@94e51e8, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5de6cf3a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4cc36c19, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5a3a1bf9, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2100d047, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4af45442]
Jan 21, 2022 12:32:29 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read input/StripIds as step s2
Jan 21, 2022 12:32:29 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect start time metrics (input) as step s3
Jan 21, 2022 12:32:29 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Window.Into()/Window.Assign as step s4
Jan 21, 2022 12:32:29 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read co-input/DataflowRunner.StreamingUnboundedRead.ReadWithIds as step s5
Jan 21, 2022 12:32:29 PM org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource split
INFO: Split into 20 bundles of sizes: [org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7da39774, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@441b8382, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1df1ced0, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5349b246, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@32b0876c, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2aaf152b, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@39bbd9e0, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@27fe9713, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@11c3ff67, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4397a639, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@24841372, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@77114efe, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@79a7bfbc, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@77f68df, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3e4e4c1, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7e7f3cfd, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3ae126d1, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@46a488c2, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6242ae3b, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@65ddee5a]
Jan 21, 2022 12:32:29 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read co-input/StripIds as step s6
Jan 21, 2022 12:32:29 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect start time metrics (co-input) as step s7
Jan 21, 2022 12:32:29 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Window.Into()2/Window.Assign as step s8
Jan 21, 2022 12:32:29 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/MakeUnionTable0 as step s9
Jan 21, 2022 12:32:29 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/MakeUnionTable1 as step s10
Jan 21, 2022 12:32:29 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/Flatten as step s11
Jan 21, 2022 12:32:29 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/GBK as step s12
Jan 21, 2022 12:32:29 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/ConstructCoGbkResultFn as step s13
Jan 21, 2022 12:32:29 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Ungroup and reiterate as step s14
Jan 21, 2022 12:32:29 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect total bytes as step s15
Jan 21, 2022 12:32:29 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect end time metrics as step s16
Jan 21, 2022 12:32:29 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Dataflow SDK version: 2.37.0-SNAPSHOT
Jan 21, 2022 12:32:32 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To access the Dataflow monitoring console, please navigate to https://console.cloud.google.com/dataflow/jobs/us-central1/2022-01-21_04_32_29-2908488888772510596?project=apache-beam-testing
Jan 21, 2022 12:32:32 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Submitted job: 2022-01-21_04_32_29-2908488888772510596
Jan 21, 2022 12:32:32 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To cancel the job using the 'gcloud' tool, run:
> gcloud dataflow jobs --project=apache-beam-testing cancel --region=us-central1 2022-01-21_04_32_29-2908488888772510596
Jan 21, 2022 12:32:39 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
WARNING: 2022-01-21T12:32:38.525Z: The workflow name is not a valid Cloud Label. Labels applied to Cloud resources (such as GCE Instances) for monitoring will be labeled with this modified job name: load0tests0java170dataflow0v20streaming0cogbk01-jenkins-01-ftof. For the best monitoring experience, please name your job with a valid Cloud Label. For details, see: https://cloud.google.com/compute/docs/labeling-resources#restrictions
Jan 21, 2022 12:32:42 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-21T12:32:42.790Z: Worker configuration: e2-standard-2 in us-central1-b.
Jan 21, 2022 12:32:44 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-21T12:32:43.530Z: Expanding SplittableParDo operations into optimizable parts.
Jan 21, 2022 12:32:44 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-21T12:32:43.562Z: Expanding CollectionToSingleton operations into optimizable parts.
Jan 21, 2022 12:32:44 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-21T12:32:43.651Z: Expanding CoGroupByKey operations into optimizable parts.
Jan 21, 2022 12:32:44 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-21T12:32:43.735Z: Expanding SplittableProcessKeyed operations into optimizable parts.
Jan 21, 2022 12:32:44 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-21T12:32:43.766Z: Expanding GroupByKey operations into streaming Read/Write steps
Jan 21, 2022 12:32:44 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-21T12:32:43.876Z: Lifting ValueCombiningMappingFns into MergeBucketsMappingFns
Jan 21, 2022 12:32:44 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-21T12:32:44.032Z: Fusing adjacent ParDo, Read, Write, and Flatten operations
Jan 21, 2022 12:32:44 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-21T12:32:44.075Z: Unzipping flatten CoGroupByKey-Flatten for input CoGroupByKey-MakeUnionTable0-ParMultiDo-ConstructUnionTable-.output
Jan 21, 2022 12:32:44 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-21T12:32:44.108Z: Fusing unzipped copy of CoGroupByKey/GBK/WriteStream, through flatten CoGroupByKey/Flatten, into producer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable)
Jan 21, 2022 12:32:44 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-21T12:32:44.152Z: Fusing consumer CoGroupByKey/GBK/WriteStream into CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable)
Jan 21, 2022 12:32:44 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-21T12:32:44.193Z: Fusing consumer Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read input/Impulse
Jan 21, 2022 12:32:44 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-21T12:32:44.231Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
Jan 21, 2022 12:32:44 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-21T12:32:44.267Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
Jan 21, 2022 12:32:44 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-21T12:32:44.299Z: Fusing consumer Read input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
Jan 21, 2022 12:32:44 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-21T12:32:44.332Z: Fusing consumer Collect start time metrics (input)/ParMultiDo(TimeMonitor) into Read input/ParDo(StripIds)/ParMultiDo(StripIds)
Jan 21, 2022 12:32:44 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-21T12:32:44.380Z: Fusing consumer Window.Into()/Window.Assign into Collect start time metrics (input)/ParMultiDo(TimeMonitor)
Jan 21, 2022 12:32:44 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-21T12:32:44.438Z: Fusing consumer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable) into Window.Into()/Window.Assign
Jan 21, 2022 12:32:46 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-21T12:32:44.480Z: Fusing consumer Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read co-input/Impulse
Jan 21, 2022 12:32:46 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-21T12:32:44.513Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
Jan 21, 2022 12:32:46 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-21T12:32:44.563Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
Jan 21, 2022 12:32:46 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-21T12:32:44.596Z: Fusing consumer Read co-input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
Jan 21, 2022 12:32:46 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-21T12:32:44.644Z: Fusing consumer Collect start time metrics (co-input)/ParMultiDo(TimeMonitor) into Read co-input/ParDo(StripIds)/ParMultiDo(StripIds)
Jan 21, 2022 12:32:46 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-21T12:32:44.671Z: Fusing consumer Window.Into()2/Window.Assign into Collect start time metrics (co-input)/ParMultiDo(TimeMonitor)
Jan 21, 2022 12:32:46 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-21T12:32:44.704Z: Fusing consumer CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable) into Window.Into()2/Window.Assign
Jan 21, 2022 12:32:46 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-21T12:32:44.727Z: Fusing consumer CoGroupByKey/GBK/MergeBuckets into CoGroupByKey/GBK/ReadStream
Jan 21, 2022 12:32:46 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-21T12:32:44.770Z: Fusing consumer CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult) into CoGroupByKey/GBK/MergeBuckets
Jan 21, 2022 12:32:46 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-21T12:32:44.812Z: Fusing consumer Ungroup and reiterate/ParMultiDo(UngroupAndReiterate) into CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult)
Jan 21, 2022 12:32:46 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-21T12:32:44.845Z: Fusing consumer Collect total bytes/ParMultiDo(ByteMonitor) into Ungroup and reiterate/ParMultiDo(UngroupAndReiterate)
Jan 21, 2022 12:32:46 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-21T12:32:44.926Z: Fusing consumer Collect end time metrics/ParMultiDo(TimeMonitor) into Collect total bytes/ParMultiDo(ByteMonitor)
Jan 21, 2022 12:32:46 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-21T12:32:45.289Z: Starting 5 ****s in us-central1-b...
Jan 21, 2022 12:33:10 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-21T12:33:08.638Z: Your project already contains 100 Dataflow-created metric descriptors, so new user metrics of the form custom.googleapis.com/* will not be created. However, all user metrics are also available in the metric dataflow.googleapis.com/job/user_counter. If you rely on the custom metrics, you can delete old / unused metric descriptors. See https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.list and https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.delete
Jan 21, 2022 12:33:32 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-21T12:33:31.411Z: Autoscaling: Raised the number of ****s to 5 so that the pipeline can catch up with its backlog and keep up with its input rate.
Jan 21, 2022 12:34:32 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-21T12:34:29.858Z: Workers have started successfully.
Jan 21, 2022 12:34:32 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-21T12:34:29.914Z: Workers have started successfully.
Jan 21, 2022 4:00:40 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-21T16:00:38.984Z: Cancel request is committed for workflow job: 2022-01-21_04_32_29-2908488888772510596.
Jan 21, 2022 4:00:40 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-21T16:00:39.096Z: Cleaning up.
Jan 21, 2022 4:00:40 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-21T16:00:39.180Z: Stopping **** pool...
Jan 21, 2022 4:00:40 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-21T16:00:39.239Z: Stopping **** pool...
Jan 21, 2022 4:03:09 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-21T16:03:07.870Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Jan 21, 2022 4:03:09 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-21T16:03:07.916Z: Worker pool stopped.
Jan 21, 2022 4:03:14 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-01-21_04_32_29-2908488888772510596 finished with status CANCELLED.
Load test results for test (ID): cf863084-5fbd-453c-80d9-f7ab25ec397b and timestamp: 2022-01-21T12:32:24.269000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                 12346.522
dataflow_v2_java17_total_bytes_count             2.58446258E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220121123024
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:fadf94325b1285d8bd7f2add7cebcacf154139134e4bea9e1cab266eeeb4b64b
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220121123024]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:fadf94325b1285d8bd7f2add7cebcacf154139134e4bea9e1cab266eeeb4b64b]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220121123024] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:fadf94325b1285d8bd7f2add7cebcacf154139134e4bea9e1cab266eeeb4b64b])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:fadf94325b1285d8bd7f2add7cebcacf154139134e4bea9e1cab266eeeb4b64b
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:fadf94325b1285d8bd7f2add7cebcacf154139134e4bea9e1cab266eeeb4b64b
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:fadf94325b1285d8bd7f2add7cebcacf154139134e4bea9e1cab266eeeb4b64b].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 32m 57s
109 actionable tasks: 72 executed, 33 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/imcplawdlxkro

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #35

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/35/display/redirect?page=changes>

Changes:

[noreply] [BEAM-13015] Provide caching statistics in the status client. (#16495)

[noreply] [BEAM-13611] Skip test_xlang_jdbc_write (#16554)

[noreply] Merge pull request #16370 from [BEAM-13556] playground - color and

[noreply] Merge pull request #16531 from [BEAM-13567] [playground] Handle run code

[noreply] Merge pull request #16533 from [BEAM-13548] [Playground] Add example

[noreply] Merge pull request #16519 from [BEAM-13639] [Playground] Add

[noreply] Merge pull request #16518 from [BEAM-13619] [Playground] Add loading

[noreply] Merge pull request #16243 from

[noreply] [BEAM-13683] Make cross-language SQL example pipeline (#16567)

[noreply] [BEAM-13688] fixed type in BPG 4.5.3 window section (#16560)

[noreply] Remove obsolete commands from Inventory job. (#16564)

[noreply] Disable logging for memoization test. (#16556)

[noreply] Merge pull request #16472: [BEAM-13697] Add SchemaFieldNumber annotation

[noreply] Merge pull request #16373 from [BEAM-13515] [Playground] Hiding lines in


------------------------------------------
[...truncated 49.31 KB...]
43b1b332c33f: Pushed
062e4c56eee6: Pushed
ce62202dd2ee: Pushed
d5050084df75: Pushed
bff9fe6e429c: Layer already exists
7c245b2fe4f1: Layer already exists
f9e18e59a565: Layer already exists
26a504e63be4: Layer already exists
8bf42db0de72: Layer already exists
31892cc314cb: Layer already exists
11936051f93b: Layer already exists
cc46b19d83be: Pushed
fe28ce742de3: Pushed
8e21f50706c6: Pushed
2d222c32ddf4: Pushed
20220120123816: digest: sha256:3e8877f0e7c61f1ddb0b792277475d50f1918a2d0818ab8a470ca3415cd841a4 size: 4520

> Task :sdks:java:testing:load-tests:run
Jan 20, 2022 12:42:09 PM org.apache.beam.runners.dataflow.options.DataflowPipelineOptions$StagingLocationFactory create
INFO: No stagingLocation provided, falling back to gcpTempLocation
Jan 20, 2022 12:42:10 PM org.apache.beam.runners.dataflow.DataflowRunner fromOptions
INFO: PipelineOptions.filesToStage was not specified. Defaulting to files from the classpath: will stage 203 files. Enable logging at DEBUG level to see which files will be staged.
Jan 20, 2022 12:42:11 PM org.apache.beam.sdk.Pipeline validate
WARNING: The following transforms do not have stable unique names: Window.Into()
Jan 20, 2022 12:42:11 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Executing pipeline on the Dataflow Service, which will have billing implications related to Google Compute Engine usage and other Google Cloud Services.
Jan 20, 2022 12:42:16 PM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Uploading 203 files from PipelineOptions.filesToStage to staging location to prepare for execution.
Jan 20, 2022 12:42:17 PM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Staging files complete: 203 files cached, 0 files newly uploaded in 1 seconds
Jan 20, 2022 12:42:17 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Staging portable pipeline proto to gs://temp-storage-for-perf-tests/loadtests/staging/
Jan 20, 2022 12:42:17 PM org.apache.beam.runners.dataflow.util.PackageUtil tryStagePackage
INFO: Uploading <113730 bytes, hash c539a314f01d7902c35f0b6f3b82dba2cd19c5a5a0f90a556e79a94d33d90895> to gs://temp-storage-for-perf-tests/loadtests/staging/pipeline-xTmjFPAdeQLDXwtvO4Lbos0ZxaWg-QpVbnmpTTPZCJU.pb
Jan 20, 2022 12:42:19 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read input/DataflowRunner.StreamingUnboundedRead.ReadWithIds as step s1
Jan 20, 2022 12:42:19 PM org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource split
INFO: Split into 20 bundles of sizes: [org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7e446d92, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@57f9b467, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6d5c2745, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@44b29496, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1947596f, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3078cac, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@f6de586, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5f2bd6d9, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7c847072, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@43d9f1a2, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@23f86d8a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2264ea32, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7d3c09ec, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@71e4b308, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@11900483, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@14a049f9, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@94e51e8, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5de6cf3a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4cc36c19, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5a3a1bf9]
Jan 20, 2022 12:42:19 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read input/StripIds as step s2
Jan 20, 2022 12:42:19 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect start time metrics (input) as step s3
Jan 20, 2022 12:42:19 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Window.Into()/Window.Assign as step s4
Jan 20, 2022 12:42:19 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read co-input/DataflowRunner.StreamingUnboundedRead.ReadWithIds as step s5
Jan 20, 2022 12:42:19 PM org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource split
INFO: Split into 20 bundles of sizes: [org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6fc6deb7, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@367f0121, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7da39774, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@441b8382, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1df1ced0, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5349b246, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@32b0876c, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2aaf152b, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@39bbd9e0, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@27fe9713, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@11c3ff67, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4397a639, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@24841372, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@77114efe, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@79a7bfbc, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@77f68df, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3e4e4c1, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7e7f3cfd, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3ae126d1, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@46a488c2]
Jan 20, 2022 12:42:19 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read co-input/StripIds as step s6
Jan 20, 2022 12:42:19 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect start time metrics (co-input) as step s7
Jan 20, 2022 12:42:19 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Window.Into()2/Window.Assign as step s8
Jan 20, 2022 12:42:19 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/MakeUnionTable0 as step s9
Jan 20, 2022 12:42:19 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/MakeUnionTable1 as step s10
Jan 20, 2022 12:42:19 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/Flatten as step s11
Jan 20, 2022 12:42:19 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/GBK as step s12
Jan 20, 2022 12:42:19 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/ConstructCoGbkResultFn as step s13
Jan 20, 2022 12:42:19 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Ungroup and reiterate as step s14
Jan 20, 2022 12:42:19 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect total bytes as step s15
Jan 20, 2022 12:42:19 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect end time metrics as step s16
Jan 20, 2022 12:42:19 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Dataflow SDK version: 2.37.0-SNAPSHOT
Jan 20, 2022 12:42:20 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To access the Dataflow monitoring console, please navigate to https://console.cloud.google.com/dataflow/jobs/us-central1/2022-01-20_04_42_19-11141015606132013537?project=apache-beam-testing
Jan 20, 2022 12:42:20 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Submitted job: 2022-01-20_04_42_19-11141015606132013537
Jan 20, 2022 12:42:20 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To cancel the job using the 'gcloud' tool, run:
> gcloud dataflow jobs --project=apache-beam-testing cancel --region=us-central1 2022-01-20_04_42_19-11141015606132013537
Jan 20, 2022 12:42:33 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
WARNING: 2022-01-20T12:42:28.728Z: The workflow name is not a valid Cloud Label. Labels applied to Cloud resources (such as GCE Instances) for monitoring will be labeled with this modified job name: load0tests0java170dataflow0v20streaming0cogbk01-jenkins-01-wbj6. For the best monitoring experience, please name your job with a valid Cloud Label. For details, see: https://cloud.google.com/compute/docs/labeling-resources#restrictions
Jan 20, 2022 12:42:36 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-20T12:42:36.244Z: Worker configuration: e2-standard-2 in us-central1-b.
Jan 20, 2022 12:42:38 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-20T12:42:37.137Z: Expanding SplittableParDo operations into optimizable parts.
Jan 20, 2022 12:42:38 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-20T12:42:37.210Z: Expanding CollectionToSingleton operations into optimizable parts.
Jan 20, 2022 12:42:38 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-20T12:42:37.440Z: Expanding CoGroupByKey operations into optimizable parts.
Jan 20, 2022 12:42:38 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-20T12:42:37.534Z: Expanding SplittableProcessKeyed operations into optimizable parts.
Jan 20, 2022 12:42:38 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-20T12:42:37.578Z: Expanding GroupByKey operations into streaming Read/Write steps
Jan 20, 2022 12:42:38 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-20T12:42:37.660Z: Lifting ValueCombiningMappingFns into MergeBucketsMappingFns
Jan 20, 2022 12:42:38 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-20T12:42:37.800Z: Fusing adjacent ParDo, Read, Write, and Flatten operations
Jan 20, 2022 12:42:38 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-20T12:42:37.842Z: Unzipping flatten CoGroupByKey-Flatten for input CoGroupByKey-MakeUnionTable0-ParMultiDo-ConstructUnionTable-.output
Jan 20, 2022 12:42:38 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-20T12:42:37.874Z: Fusing unzipped copy of CoGroupByKey/GBK/WriteStream, through flatten CoGroupByKey/Flatten, into producer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable)
Jan 20, 2022 12:42:38 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-20T12:42:37.919Z: Fusing consumer CoGroupByKey/GBK/WriteStream into CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable)
Jan 20, 2022 12:42:38 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-20T12:42:37.950Z: Fusing consumer Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read input/Impulse
Jan 20, 2022 12:42:38 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-20T12:42:37.995Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
Jan 20, 2022 12:42:38 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-20T12:42:38.052Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
Jan 20, 2022 12:42:38 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-20T12:42:38.123Z: Fusing consumer Read input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
Jan 20, 2022 12:42:39 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-20T12:42:38.220Z: Fusing consumer Collect start time metrics (input)/ParMultiDo(TimeMonitor) into Read input/ParDo(StripIds)/ParMultiDo(StripIds)
Jan 20, 2022 12:42:39 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-20T12:42:38.325Z: Fusing consumer Window.Into()/Window.Assign into Collect start time metrics (input)/ParMultiDo(TimeMonitor)
Jan 20, 2022 12:42:39 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-20T12:42:38.421Z: Fusing consumer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable) into Window.Into()/Window.Assign
Jan 20, 2022 12:42:39 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-20T12:42:38.491Z: Fusing consumer Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read co-input/Impulse
Jan 20, 2022 12:42:39 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-20T12:42:38.578Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
Jan 20, 2022 12:42:39 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-20T12:42:38.612Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
Jan 20, 2022 12:42:39 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-20T12:42:38.649Z: Fusing consumer Read co-input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
Jan 20, 2022 12:42:39 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-20T12:42:38.682Z: Fusing consumer Collect start time metrics (co-input)/ParMultiDo(TimeMonitor) into Read co-input/ParDo(StripIds)/ParMultiDo(StripIds)
Jan 20, 2022 12:42:39 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-20T12:42:38.749Z: Fusing consumer Window.Into()2/Window.Assign into Collect start time metrics (co-input)/ParMultiDo(TimeMonitor)
Jan 20, 2022 12:42:39 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-20T12:42:38.784Z: Fusing consumer CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable) into Window.Into()2/Window.Assign
Jan 20, 2022 12:42:39 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-20T12:42:38.828Z: Fusing consumer CoGroupByKey/GBK/MergeBuckets into CoGroupByKey/GBK/ReadStream
Jan 20, 2022 12:42:39 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-20T12:42:38.863Z: Fusing consumer CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult) into CoGroupByKey/GBK/MergeBuckets
Jan 20, 2022 12:42:39 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-20T12:42:38.898Z: Fusing consumer Ungroup and reiterate/ParMultiDo(UngroupAndReiterate) into CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult)
Jan 20, 2022 12:42:39 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-20T12:42:38.934Z: Fusing consumer Collect total bytes/ParMultiDo(ByteMonitor) into Ungroup and reiterate/ParMultiDo(UngroupAndReiterate)
Jan 20, 2022 12:42:39 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-20T12:42:38.970Z: Fusing consumer Collect end time metrics/ParMultiDo(TimeMonitor) into Collect total bytes/ParMultiDo(ByteMonitor)
Jan 20, 2022 12:42:42 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-20T12:42:39.772Z: Starting 5 ****s in us-central1-b...
Jan 20, 2022 12:43:10 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-20T12:43:10.085Z: Your project already contains 100 Dataflow-created metric descriptors, so new user metrics of the form custom.googleapis.com/* will not be created. However, all user metrics are also available in the metric dataflow.googleapis.com/job/user_counter. If you rely on the custom metrics, you can delete old / unused metric descriptors. See https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.list and https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.delete
Jan 20, 2022 12:43:33 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-20T12:43:31.941Z: Autoscaling: Raised the number of ****s to 5 so that the pipeline can catch up with its backlog and keep up with its input rate.
Jan 20, 2022 12:44:29 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-20T12:44:29.418Z: Workers have started successfully.
Jan 20, 2022 12:44:29 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-20T12:44:29.452Z: Workers have started successfully.
Jan 20, 2022 4:00:29 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-20T16:00:27.377Z: Cancel request is committed for workflow job: 2022-01-20_04_42_19-11141015606132013537.
Jan 20, 2022 4:00:29 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-20T16:00:27.725Z: Cleaning up.
Jan 20, 2022 4:00:29 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-20T16:00:27.862Z: Stopping **** pool...
Jan 20, 2022 4:00:29 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-20T16:00:27.933Z: Stopping **** pool...
Jan 20, 2022 4:02:52 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-20T16:02:50.223Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Jan 20, 2022 4:02:52 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-20T16:02:50.563Z: Worker pool stopped.
Jan 20, 2022 4:03:01 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-01-20_04_42_19-11141015606132013537 finished with status CANCELLED.
Load test results for test (ID): ce228a0e-ebcc-48a5-97aa-48af9eb6a312 and timestamp: 2022-01-20T12:42:11.098000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                 11657.892
dataflow_v2_java17_total_bytes_count             2.15460128E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220120123816
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:3e8877f0e7c61f1ddb0b792277475d50f1918a2d0818ab8a470ca3415cd841a4
Deleted: sha256:6c1a17878515ab8d7609f5fa8b5a4cc60bb7d77eb999b5e9d64a03c84f38e203
Deleted: sha256:129b0f5d2ea289c5a58b9b47b8bd7ac640bc1df14d905834406588915cb1b6da
Deleted: sha256:52fed35969c8a44792276fdef3a63a1ee3f3bce7875e107b800905386af700c3
Deleted: sha256:40cd6f4eff9f04fd1b65ca99edba8dea306c75551f8faa30f237f051467f996f
Deleted: sha256:0bcf8258af86a20e20ec989361ff66bd7a4db067740f5d42c7dc1ee9eca79dac
Deleted: sha256:f90dffdefe162bb72df2d2809e45a93f9db2ca4e81f1ba31c2ef01ae914cccbc
Deleted: sha256:ceebe3bb2d79a39b8ad2492adcde68a4da2d967958830150ce63d441c659230d
Deleted: sha256:dd105e0557af70e3ee6f399d18fb980ecca26f23c93a02febbf97d47779aa6ba
Deleted: sha256:04bf88108c855d28f7d81b961ad3102f113b36a30e28b0aa706416d35da50082
Deleted: sha256:6ed02c71439d5f652beacc4e6d1afb3d021bf876a683ce47ea9328da319b6449
Deleted: sha256:bcefa1af8e4feb2aeb71ad648fe572238f78c3b47bf42c14607845fdfbf4f55f
Deleted: sha256:f49d31439906cf788bb3331ea3a8b5fe272dad2ae567f3d7e6a9016a78238f77
Deleted: sha256:d59fffa18c58dda59179d8e82e1b7123f1f486a16b188fbabd59cc113a1cd85e
Deleted: sha256:db8122707d8ad18e579e6c88fffe688b9d7faa1c8c7e6be655ca2d72c0f09b74
Deleted: sha256:228c31882b4161e3b976fb01f2e38b54e20c596379c82fd5dc3c911aacfd7aae
Deleted: sha256:839d1dc10d61a72162dc239ace4c14d972ce4c9f3debf32f87e324b9d7facd0a
Deleted: sha256:2ab4741ce5266d13b78766514a142aeea5ad2042aa194897bbf3c780d60b20c4
Deleted: sha256:111c08a3eb2c0a8086d49ecf077e02f34520340a3c99d46fdc1e4d003e1356dd
Deleted: sha256:6bc8ade3426fb3e0dd743b7d7f8d0e8670183e6bc5b76da8db932d51d3b3d66e
Deleted: sha256:79986a283c14b17e1d473f85369f0ca5e075766e76d6aa1439642074eede3f69
Deleted: sha256:4728772a98fe139b94a3d2636953532eb4906c537c2b6d7aee3adff44255d408
Deleted: sha256:7620f51d1b04fd48a2de6818a6a168bd5aa2bd18499dbb8eba1396c699cf16dd
Deleted: sha256:0f82182ea4e50ba06ecba62bb254b571e02509e2d25a302e7bbfb0792fe1bbaa
Deleted: sha256:925db3c60a2eaca0d10b52d890c78e04c86f8bdfa351f607b52e1ac7e604b390
Deleted: sha256:e0b3205245e9084c443f42b6d940e9965945cf003cdb1e0d2a98bde63b8b96fa
Deleted: sha256:fd8d296b8293aab13ee895274a40f85882b93b61f506158aed8ca75fbdd3dad8
Deleted: sha256:4588bb34ff974faa2d5d09a532c5a259ce7ab6866196c4d9e5342a8587b53f7d
Deleted: sha256:2083a7e9970e41c004430ad67d997a72a162ab109239a82c7e2e71d6b5618913
Deleted: sha256:a7d945b3710aec4700733460d9ed277baf1b3e90555021095a80885a78385f5e
Deleted: sha256:0d4e8a3fd47341076e626a2fc0263e1cbeb46ed4e5b5678817e74f65fc084e49
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220120123816]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:3e8877f0e7c61f1ddb0b792277475d50f1918a2d0818ab8a470ca3415cd841a4]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220120123816] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:3e8877f0e7c61f1ddb0b792277475d50f1918a2d0818ab8a470ca3415cd841a4])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:3e8877f0e7c61f1ddb0b792277475d50f1918a2d0818ab8a470ca3415cd841a4
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:3e8877f0e7c61f1ddb0b792277475d50f1918a2d0818ab8a470ca3415cd841a4
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:3e8877f0e7c61f1ddb0b792277475d50f1918a2d0818ab8a470ca3415cd841a4].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 24m 59s
109 actionable tasks: 72 executed, 33 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/l2w3liksobaqg

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #34

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/34/display/redirect?page=changes>

Changes:

[Robert Bradshaw] Bump beam container version.

[alexander.zhuravlev] [BEAM-13680] Fixed code_repository (added pipelineUuid to RunCodeResult

[Robert Bradshaw] Also bump FnAPI container.

[noreply] [BEAM-13616][BEAM-13645] Switch to vendored grpc 1.43.2 (#16543)

[noreply] [BEAM-13616][BEAM-13646] Upgrade vendored calcite to 1.28.0:0.2 (#16544)

[noreply] Merge pull request #16486 from [BEAM-13544][Playground] Add logs to

[noreply] [BEAM-13683] Correct SQL transform schema, fix expansion address

[noreply] Update walkthrough.md (#16512)

[noreply] [BEAM-11808][BEAM-9879] Support aggregate functions with two arguments

[noreply] Merge pull request #16506 from [BEAM-13652][Playground] Send examples'

[noreply] Merge pull request #16322 from [BEAM-13407] [Playground] Preload fonts

[noreply] [BEAM-13665] Make SpannerIO projectID optional again (#16547)

[noreply] [BEAM-13015] Add state caching capability to be used as hint for runners

[noreply] Merge pull request #16309: [BEAM-13503] Set a default value to


------------------------------------------
[...truncated 49.57 KB...]
26a504e63be4: Preparing
8bf42db0de72: Preparing
68cea805aa5d: Waiting
31892cc314cb: Preparing
11936051f93b: Preparing
b00de53e288f: Waiting
224e65a07a09: Waiting
47bec37be8ae: Waiting
08be3ef1725f: Waiting
f858438ba3fd: Waiting
bff9fe6e429c: Waiting
f0f3e6ebcbee: Waiting
6735797c3b6f: Waiting
8bf42db0de72: Waiting
f9e18e59a565: Waiting
31892cc314cb: Waiting
26a504e63be4: Waiting
11936051f93b: Waiting
7c245b2fe4f1: Waiting
bebc0265cc30: Pushed
d0f8665d0aa8: Pushed
ebd4087e8fad: Pushed
f451fb852e57: Pushed
222a5e124fdc: Pushed
b00de53e288f: Pushed
68cea805aa5d: Pushed
f858438ba3fd: Pushed
08be3ef1725f: Pushed
bff9fe6e429c: Layer already exists
7c245b2fe4f1: Layer already exists
224e65a07a09: Pushed
f9e18e59a565: Layer already exists
26a504e63be4: Layer already exists
8bf42db0de72: Layer already exists
47bec37be8ae: Pushed
31892cc314cb: Layer already exists
11936051f93b: Layer already exists
f0f3e6ebcbee: Pushed
6735797c3b6f: Pushed
20220119123240: digest: sha256:b6c0a6d0fc4142073531152159f4cfdaf4cd07d8c439035fc7747863d5085822 size: 4520

> Task :sdks:java:testing:load-tests:run
Jan 19, 2022 12:34:33 PM org.apache.beam.runners.dataflow.options.DataflowPipelineOptions$StagingLocationFactory create
INFO: No stagingLocation provided, falling back to gcpTempLocation
Jan 19, 2022 12:34:33 PM org.apache.beam.runners.dataflow.DataflowRunner fromOptions
INFO: PipelineOptions.filesToStage was not specified. Defaulting to files from the classpath: will stage 203 files. Enable logging at DEBUG level to see which files will be staged.
Jan 19, 2022 12:34:34 PM org.apache.beam.sdk.Pipeline validate
WARNING: The following transforms do not have stable unique names: Window.Into()
Jan 19, 2022 12:34:34 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Executing pipeline on the Dataflow Service, which will have billing implications related to Google Compute Engine usage and other Google Cloud Services.
Jan 19, 2022 12:34:37 PM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Uploading 203 files from PipelineOptions.filesToStage to staging location to prepare for execution.
Jan 19, 2022 12:34:38 PM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Staging files complete: 203 files cached, 0 files newly uploaded in 0 seconds
Jan 19, 2022 12:34:38 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Staging portable pipeline proto to gs://temp-storage-for-perf-tests/loadtests/staging/
Jan 19, 2022 12:34:38 PM org.apache.beam.runners.dataflow.util.PackageUtil tryStagePackage
INFO: Uploading <113730 bytes, hash fc78ef45c81441632efc26ebb4e1b35f421ab390eb20c9fa69de59d185c0a11b> to gs://temp-storage-for-perf-tests/loadtests/staging/pipeline-_HjvRcgUQWMu_CbrtOGzX0Ias5DrIMn6ad5Z0YXAoRs.pb
Jan 19, 2022 12:34:40 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read input/DataflowRunner.StreamingUnboundedRead.ReadWithIds as step s1
Jan 19, 2022 12:34:40 PM org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource split
INFO: Split into 20 bundles of sizes: [org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@44b29496, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1947596f, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3078cac, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@f6de586, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5f2bd6d9, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7c847072, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@43d9f1a2, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@23f86d8a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2264ea32, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7d3c09ec, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@71e4b308, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@11900483, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@14a049f9, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@94e51e8, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5de6cf3a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4cc36c19, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5a3a1bf9, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2100d047, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4af45442, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1d805aa1]
Jan 19, 2022 12:34:40 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read input/StripIds as step s2
Jan 19, 2022 12:34:40 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect start time metrics (input) as step s3
Jan 19, 2022 12:34:40 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Window.Into()/Window.Assign as step s4
Jan 19, 2022 12:34:40 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read co-input/DataflowRunner.StreamingUnboundedRead.ReadWithIds as step s5
Jan 19, 2022 12:34:40 PM org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource split
INFO: Split into 20 bundles of sizes: [org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@441b8382, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1df1ced0, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5349b246, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@32b0876c, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2aaf152b, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@39bbd9e0, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@27fe9713, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@11c3ff67, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4397a639, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@24841372, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@77114efe, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@79a7bfbc, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@77f68df, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3e4e4c1, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7e7f3cfd, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3ae126d1, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@46a488c2, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6242ae3b, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@65ddee5a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@56399b9e]
Jan 19, 2022 12:34:40 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read co-input/StripIds as step s6
Jan 19, 2022 12:34:40 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect start time metrics (co-input) as step s7
Jan 19, 2022 12:34:40 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Window.Into()2/Window.Assign as step s8
Jan 19, 2022 12:34:40 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/MakeUnionTable0 as step s9
Jan 19, 2022 12:34:40 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/MakeUnionTable1 as step s10
Jan 19, 2022 12:34:40 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/Flatten as step s11
Jan 19, 2022 12:34:40 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/GBK as step s12
Jan 19, 2022 12:34:40 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/ConstructCoGbkResultFn as step s13
Jan 19, 2022 12:34:40 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Ungroup and reiterate as step s14
Jan 19, 2022 12:34:40 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect total bytes as step s15
Jan 19, 2022 12:34:40 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect end time metrics as step s16
Jan 19, 2022 12:34:40 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Dataflow SDK version: 2.37.0-SNAPSHOT
Jan 19, 2022 12:34:41 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To access the Dataflow monitoring console, please navigate to https://console.cloud.google.com/dataflow/jobs/us-central1/2022-01-19_04_34_40-232627097302293716?project=apache-beam-testing
Jan 19, 2022 12:34:41 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Submitted job: 2022-01-19_04_34_40-232627097302293716
Jan 19, 2022 12:34:41 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To cancel the job using the 'gcloud' tool, run:
> gcloud dataflow jobs --project=apache-beam-testing cancel --region=us-central1 2022-01-19_04_34_40-232627097302293716
Jan 19, 2022 12:34:50 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
WARNING: 2022-01-19T12:34:48.830Z: The workflow name is not a valid Cloud Label. Labels applied to Cloud resources (such as GCE Instances) for monitoring will be labeled with this modified job name: load0tests0java170dataflow0v20streaming0cogbk01-jenkins-01-udk4. For the best monitoring experience, please name your job with a valid Cloud Label. For details, see: https://cloud.google.com/compute/docs/labeling-resources#restrictions
Jan 19, 2022 12:34:53 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-19T12:34:53.063Z: Worker configuration: e2-standard-2 in us-central1-b.
Jan 19, 2022 12:34:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-19T12:34:53.799Z: Expanding SplittableParDo operations into optimizable parts.
Jan 19, 2022 12:34:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-19T12:34:53.830Z: Expanding CollectionToSingleton operations into optimizable parts.
Jan 19, 2022 12:34:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-19T12:34:53.908Z: Expanding CoGroupByKey operations into optimizable parts.
Jan 19, 2022 12:34:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-19T12:34:53.982Z: Expanding SplittableProcessKeyed operations into optimizable parts.
Jan 19, 2022 12:34:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-19T12:34:54.042Z: Expanding GroupByKey operations into streaming Read/Write steps
Jan 19, 2022 12:34:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-19T12:34:54.149Z: Lifting ValueCombiningMappingFns into MergeBucketsMappingFns
Jan 19, 2022 12:34:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-19T12:34:54.279Z: Fusing adjacent ParDo, Read, Write, and Flatten operations
Jan 19, 2022 12:34:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-19T12:34:54.313Z: Unzipping flatten CoGroupByKey-Flatten for input CoGroupByKey-MakeUnionTable0-ParMultiDo-ConstructUnionTable-.output
Jan 19, 2022 12:34:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-19T12:34:54.346Z: Fusing unzipped copy of CoGroupByKey/GBK/WriteStream, through flatten CoGroupByKey/Flatten, into producer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable)
Jan 19, 2022 12:34:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-19T12:34:54.369Z: Fusing consumer CoGroupByKey/GBK/WriteStream into CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable)
Jan 19, 2022 12:34:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-19T12:34:54.401Z: Fusing consumer Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read input/Impulse
Jan 19, 2022 12:34:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-19T12:34:54.424Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
Jan 19, 2022 12:34:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-19T12:34:54.457Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
Jan 19, 2022 12:34:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-19T12:34:54.490Z: Fusing consumer Read input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
Jan 19, 2022 12:34:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-19T12:34:54.522Z: Fusing consumer Collect start time metrics (input)/ParMultiDo(TimeMonitor) into Read input/ParDo(StripIds)/ParMultiDo(StripIds)
Jan 19, 2022 12:34:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-19T12:34:54.557Z: Fusing consumer Window.Into()/Window.Assign into Collect start time metrics (input)/ParMultiDo(TimeMonitor)
Jan 19, 2022 12:34:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-19T12:34:54.593Z: Fusing consumer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable) into Window.Into()/Window.Assign
Jan 19, 2022 12:34:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-19T12:34:54.614Z: Fusing consumer Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read co-input/Impulse
Jan 19, 2022 12:34:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-19T12:34:54.647Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
Jan 19, 2022 12:34:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-19T12:34:54.691Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
Jan 19, 2022 12:34:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-19T12:34:54.715Z: Fusing consumer Read co-input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
Jan 19, 2022 12:34:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-19T12:34:54.752Z: Fusing consumer Collect start time metrics (co-input)/ParMultiDo(TimeMonitor) into Read co-input/ParDo(StripIds)/ParMultiDo(StripIds)
Jan 19, 2022 12:34:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-19T12:34:54.790Z: Fusing consumer Window.Into()2/Window.Assign into Collect start time metrics (co-input)/ParMultiDo(TimeMonitor)
Jan 19, 2022 12:34:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-19T12:34:54.840Z: Fusing consumer CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable) into Window.Into()2/Window.Assign
Jan 19, 2022 12:34:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-19T12:34:54.873Z: Fusing consumer CoGroupByKey/GBK/MergeBuckets into CoGroupByKey/GBK/ReadStream
Jan 19, 2022 12:34:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-19T12:34:54.900Z: Fusing consumer CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult) into CoGroupByKey/GBK/MergeBuckets
Jan 19, 2022 12:34:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-19T12:34:54.933Z: Fusing consumer Ungroup and reiterate/ParMultiDo(UngroupAndReiterate) into CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult)
Jan 19, 2022 12:34:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-19T12:34:54.965Z: Fusing consumer Collect total bytes/ParMultiDo(ByteMonitor) into Ungroup and reiterate/ParMultiDo(UngroupAndReiterate)
Jan 19, 2022 12:34:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-19T12:34:54.998Z: Fusing consumer Collect end time metrics/ParMultiDo(TimeMonitor) into Collect total bytes/ParMultiDo(ByteMonitor)
Jan 19, 2022 12:34:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-19T12:34:55.436Z: Starting 5 ****s in us-central1-b...
Jan 19, 2022 12:35:16 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-19T12:35:15.594Z: Your project already contains 100 Dataflow-created metric descriptors, so new user metrics of the form custom.googleapis.com/* will not be created. However, all user metrics are also available in the metric dataflow.googleapis.com/job/user_counter. If you rely on the custom metrics, you can delete old / unused metric descriptors. See https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.list and https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.delete
Jan 19, 2022 12:35:39 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-19T12:35:38.583Z: Autoscaling: Raised the number of ****s to 5 so that the pipeline can catch up with its backlog and keep up with its input rate.
Jan 19, 2022 12:36:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-19T12:36:38.384Z: Workers have started successfully.
Jan 19, 2022 12:36:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-19T12:36:38.423Z: Workers have started successfully.
Jan 19, 2022 1:40:29 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-19T13:40:27.255Z: Workers have started successfully.
Jan 19, 2022 1:40:33 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-19T13:40:31.813Z: Worker configuration: e2-standard-2 in us-central1-b.
Jan 19, 2022 1:40:36 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-19T13:40:33.966Z: Your project already contains 100 Dataflow-created metric descriptors, so new user metrics of the form custom.googleapis.com/* will not be created. However, all user metrics are also available in the metric dataflow.googleapis.com/job/user_counter. If you rely on the custom metrics, you can delete old / unused metric descriptors. See https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.list and https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.delete
Jan 19, 2022 4:00:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-19T16:00:25.623Z: Cancel request is committed for workflow job: 2022-01-19_04_34_40-232627097302293716.
Jan 19, 2022 4:00:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-19T16:00:25.677Z: Cleaning up.
Jan 19, 2022 4:00:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-19T16:00:25.770Z: Stopping **** pool...
Jan 19, 2022 4:00:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-19T16:00:25.810Z: Stopping **** pool...
Jan 19, 2022 4:02:58 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-19T16:02:56.413Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Jan 19, 2022 4:02:58 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-19T16:02:56.450Z: Worker pool stopped.
Jan 19, 2022 4:03:02 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-01-19_04_34_40-232627097302293716 finished with status CANCELLED.
Load test results for test (ID): 7e9204af-df16-49da-b0b7-23965bb31fac and timestamp: 2022-01-19T12:34:33.984000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                 12179.635
dataflow_v2_java17_total_bytes_count             3.01978284E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220119123240
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:b6c0a6d0fc4142073531152159f4cfdaf4cd07d8c439035fc7747863d5085822
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220119123240]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:b6c0a6d0fc4142073531152159f4cfdaf4cd07d8c439035fc7747863d5085822]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220119123240] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:b6c0a6d0fc4142073531152159f4cfdaf4cd07d8c439035fc7747863d5085822])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:b6c0a6d0fc4142073531152159f4cfdaf4cd07d8c439035fc7747863d5085822
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:b6c0a6d0fc4142073531152159f4cfdaf4cd07d8c439035fc7747863d5085822
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:b6c0a6d0fc4142073531152159f4cfdaf4cd07d8c439035fc7747863d5085822].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 30m 41s
109 actionable tasks: 72 executed, 33 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/4g25jt2vfotxa

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #33

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/33/display/redirect?page=changes>

Changes:

[mmack] [BEAM-8806] Integration test for SqsIO

[mmack] [BEAM-13631] Add deterministic SQS message coder to fix reading from SQS

[aydar.zaynutdinov] [BEAM-13641][Playground]

[noreply] Remove jcenter repositories from gradle configuration. (#16532)

[noreply] [BEAM-13430] Remove jcenter which will no longer contain any updates.

[noreply] [BEAM-13616] Update com.google.cloud:libraries-bom to 24.2.0 (#16509)


------------------------------------------
[...truncated 50.29 KB...]
895812f78955: Pushed
f2ac77edcc70: Pushed
55b894764515: Pushed
3c6559e4f378: Pushed
bff9fe6e429c: Layer already exists
3c19eb2af73a: Pushed
96a1c4724005: Pushed
7c245b2fe4f1: Layer already exists
f9e18e59a565: Layer already exists
26a504e63be4: Layer already exists
8bf42db0de72: Layer already exists
31892cc314cb: Layer already exists
11936051f93b: Layer already exists
a5967e1cae27: Pushed
49e595d77702: Pushed
20220118124120: digest: sha256:838c3eea10304229b5fd7e1fc67a70751ca0462bcd48e213de3fc3e0d41564e5 size: 4520

> Task :sdks:java:testing:load-tests:run
Jan 18, 2022 12:43:31 PM org.apache.beam.runners.dataflow.options.DataflowPipelineOptions$StagingLocationFactory create
INFO: No stagingLocation provided, falling back to gcpTempLocation
Jan 18, 2022 12:43:32 PM org.apache.beam.runners.dataflow.DataflowRunner fromOptions
INFO: PipelineOptions.filesToStage was not specified. Defaulting to files from the classpath: will stage 203 files. Enable logging at DEBUG level to see which files will be staged.
Jan 18, 2022 12:43:33 PM org.apache.beam.sdk.Pipeline validate
WARNING: The following transforms do not have stable unique names: Window.Into()
Jan 18, 2022 12:43:33 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Executing pipeline on the Dataflow Service, which will have billing implications related to Google Compute Engine usage and other Google Cloud Services.
Jan 18, 2022 12:43:36 PM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Uploading 203 files from PipelineOptions.filesToStage to staging location to prepare for execution.
Jan 18, 2022 12:43:37 PM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Staging files complete: 203 files cached, 0 files newly uploaded in 0 seconds
Jan 18, 2022 12:43:37 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Staging portable pipeline proto to gs://temp-storage-for-perf-tests/loadtests/staging/
Jan 18, 2022 12:43:37 PM org.apache.beam.runners.dataflow.util.PackageUtil tryStagePackage
INFO: Uploading <113698 bytes, hash d9e0cf34b930801b248b69d15ce03385cad1935590df0a6f8c4df5e5a1ea0d28> to gs://temp-storage-for-perf-tests/loadtests/staging/pipeline-2eDPNLkwgBski2nRXOAzhcrRk1WQ3wpvjE315aHqDSg.pb
Jan 18, 2022 12:43:39 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read input/DataflowRunner.StreamingUnboundedRead.ReadWithIds as step s1
Jan 18, 2022 12:43:39 PM org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource split
INFO: Split into 20 bundles of sizes: [org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@71c17a57, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@640ab13c, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1e0a864d, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@440e3ce6, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3e67f5f2, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@fd53053, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4527f70a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@707b1a44, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7132a9dc, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@57435801, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2da66a44, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@527fc8e, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@61bfc9bf, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2c7106d9, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@329bad59, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@862f408, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@178f268a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@33ecbd6c, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5c723f2d, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@432f521f]
Jan 18, 2022 12:43:39 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read input/StripIds as step s2
Jan 18, 2022 12:43:39 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect start time metrics (input) as step s3
Jan 18, 2022 12:43:39 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Window.Into()/Window.Assign as step s4
Jan 18, 2022 12:43:39 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read co-input/DataflowRunner.StreamingUnboundedRead.ReadWithIds as step s5
Jan 18, 2022 12:43:39 PM org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource split
INFO: Split into 20 bundles of sizes: [org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3ac3f6f, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1abebef3, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@18f55704, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@67cefd84, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5fbe155, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6add8e3f, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@58a2b917, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@48904d5a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@12bbfc54, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1491344a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@59b65dce, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1386313f, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5e922647, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@433c6abb, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@288f173f, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@a22c4d8, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@45cd7bc5, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6504a875, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@35e26d05, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@29fa6b65]
Jan 18, 2022 12:43:39 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read co-input/StripIds as step s6
Jan 18, 2022 12:43:39 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect start time metrics (co-input) as step s7
Jan 18, 2022 12:43:39 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Window.Into()2/Window.Assign as step s8
Jan 18, 2022 12:43:39 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/MakeUnionTable0 as step s9
Jan 18, 2022 12:43:39 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/MakeUnionTable1 as step s10
Jan 18, 2022 12:43:39 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/Flatten as step s11
Jan 18, 2022 12:43:39 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/GBK as step s12
Jan 18, 2022 12:43:39 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/ConstructCoGbkResultFn as step s13
Jan 18, 2022 12:43:40 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Ungroup and reiterate as step s14
Jan 18, 2022 12:43:40 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect total bytes as step s15
Jan 18, 2022 12:43:40 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect end time metrics as step s16
Jan 18, 2022 12:43:40 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Dataflow SDK version: 2.37.0-SNAPSHOT
Jan 18, 2022 12:43:41 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To access the Dataflow monitoring console, please navigate to https://console.cloud.google.com/dataflow/jobs/us-central1/2022-01-18_04_43_40-8001858609153742492?project=apache-beam-testing
Jan 18, 2022 12:43:41 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Submitted job: 2022-01-18_04_43_40-8001858609153742492
Jan 18, 2022 12:43:41 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To cancel the job using the 'gcloud' tool, run:
> gcloud dataflow jobs --project=apache-beam-testing cancel --region=us-central1 2022-01-18_04_43_40-8001858609153742492
Jan 18, 2022 12:43:48 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
WARNING: 2022-01-18T12:43:46.460Z: The workflow name is not a valid Cloud Label. Labels applied to Cloud resources (such as GCE Instances) for monitoring will be labeled with this modified job name: load0tests0java170dataflow0v20streaming0cogbk01-jenkins-01-1hlr. For the best monitoring experience, please name your job with a valid Cloud Label. For details, see: https://cloud.google.com/compute/docs/labeling-resources#restrictions
Jan 18, 2022 12:43:50 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-18T12:43:50.063Z: Worker configuration: e2-standard-2 in us-central1-b.
Jan 18, 2022 12:43:52 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-18T12:43:50.675Z: Expanding SplittableParDo operations into optimizable parts.
Jan 18, 2022 12:43:52 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-18T12:43:50.701Z: Expanding CollectionToSingleton operations into optimizable parts.
Jan 18, 2022 12:43:52 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-18T12:43:50.758Z: Expanding CoGroupByKey operations into optimizable parts.
Jan 18, 2022 12:43:52 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-18T12:43:50.836Z: Expanding SplittableProcessKeyed operations into optimizable parts.
Jan 18, 2022 12:43:52 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-18T12:43:50.866Z: Expanding GroupByKey operations into streaming Read/Write steps
Jan 18, 2022 12:43:52 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-18T12:43:50.918Z: Lifting ValueCombiningMappingFns into MergeBucketsMappingFns
Jan 18, 2022 12:43:52 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-18T12:43:51.003Z: Fusing adjacent ParDo, Read, Write, and Flatten operations
Jan 18, 2022 12:43:52 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-18T12:43:51.038Z: Unzipping flatten CoGroupByKey-Flatten for input CoGroupByKey-MakeUnionTable0-ParMultiDo-ConstructUnionTable-.output
Jan 18, 2022 12:43:52 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-18T12:43:51.096Z: Fusing unzipped copy of CoGroupByKey/GBK/WriteStream, through flatten CoGroupByKey/Flatten, into producer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable)
Jan 18, 2022 12:43:52 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-18T12:43:51.126Z: Fusing consumer CoGroupByKey/GBK/WriteStream into CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable)
Jan 18, 2022 12:43:52 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-18T12:43:51.156Z: Fusing consumer Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read input/Impulse
Jan 18, 2022 12:43:52 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-18T12:43:51.188Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
Jan 18, 2022 12:43:52 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-18T12:43:51.208Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
Jan 18, 2022 12:43:52 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-18T12:43:51.230Z: Fusing consumer Read input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
Jan 18, 2022 12:43:52 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-18T12:43:51.251Z: Fusing consumer Collect start time metrics (input)/ParMultiDo(TimeMonitor) into Read input/ParDo(StripIds)/ParMultiDo(StripIds)
Jan 18, 2022 12:43:52 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-18T12:43:51.281Z: Fusing consumer Window.Into()/Window.Assign into Collect start time metrics (input)/ParMultiDo(TimeMonitor)
Jan 18, 2022 12:43:52 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-18T12:43:51.309Z: Fusing consumer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable) into Window.Into()/Window.Assign
Jan 18, 2022 12:43:52 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-18T12:43:51.330Z: Fusing consumer Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read co-input/Impulse
Jan 18, 2022 12:43:52 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-18T12:43:51.359Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
Jan 18, 2022 12:43:52 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-18T12:43:51.377Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
Jan 18, 2022 12:43:52 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-18T12:43:51.402Z: Fusing consumer Read co-input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
Jan 18, 2022 12:43:52 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-18T12:43:51.425Z: Fusing consumer Collect start time metrics (co-input)/ParMultiDo(TimeMonitor) into Read co-input/ParDo(StripIds)/ParMultiDo(StripIds)
Jan 18, 2022 12:43:52 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-18T12:43:51.447Z: Fusing consumer Window.Into()2/Window.Assign into Collect start time metrics (co-input)/ParMultiDo(TimeMonitor)
Jan 18, 2022 12:43:52 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-18T12:43:51.474Z: Fusing consumer CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable) into Window.Into()2/Window.Assign
Jan 18, 2022 12:43:52 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-18T12:43:51.499Z: Fusing consumer CoGroupByKey/GBK/MergeBuckets into CoGroupByKey/GBK/ReadStream
Jan 18, 2022 12:43:52 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-18T12:43:51.524Z: Fusing consumer CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult) into CoGroupByKey/GBK/MergeBuckets
Jan 18, 2022 12:43:52 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-18T12:43:51.544Z: Fusing consumer Ungroup and reiterate/ParMultiDo(UngroupAndReiterate) into CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult)
Jan 18, 2022 12:43:52 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-18T12:43:51.561Z: Fusing consumer Collect total bytes/ParMultiDo(ByteMonitor) into Ungroup and reiterate/ParMultiDo(UngroupAndReiterate)
Jan 18, 2022 12:43:52 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-18T12:43:51.589Z: Fusing consumer Collect end time metrics/ParMultiDo(TimeMonitor) into Collect total bytes/ParMultiDo(ByteMonitor)
Jan 18, 2022 12:43:52 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-18T12:43:51.895Z: Starting 5 ****s in us-central1-b...
Jan 18, 2022 12:44:06 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-18T12:44:04.937Z: Your project already contains 100 Dataflow-created metric descriptors, so new user metrics of the form custom.googleapis.com/* will not be created. However, all user metrics are also available in the metric dataflow.googleapis.com/job/user_counter. If you rely on the custom metrics, you can delete old / unused metric descriptors. See https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.list and https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.delete
Jan 18, 2022 12:44:39 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-18T12:44:37.563Z: Autoscaling: Raised the number of ****s to 5 so that the pipeline can catch up with its backlog and keep up with its input rate.
Jan 18, 2022 12:45:37 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-18T12:45:35.247Z: Workers have started successfully.
Jan 18, 2022 12:45:37 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-18T12:45:35.276Z: Workers have started successfully.
Jan 18, 2022 4:00:42 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-18T16:00:41.305Z: Cancel request is committed for workflow job: 2022-01-18_04_43_40-8001858609153742492.
Jan 18, 2022 4:00:42 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-18T16:00:41.394Z: Cleaning up.
Jan 18, 2022 4:00:42 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-18T16:00:41.473Z: Stopping **** pool...
Jan 18, 2022 4:00:42 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-18T16:00:41.514Z: Stopping **** pool...
Jan 18, 2022 4:03:02 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-18T16:03:00.175Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Jan 18, 2022 4:03:02 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-18T16:03:00.216Z: Worker pool stopped.
Jan 18, 2022 4:03:06 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-01-18_04_43_40-8001858609153742492 finished with status CANCELLED.
Load test results for test (ID): 22318b1f-542d-438a-b6ed-2748d497df5b and timestamp: 2022-01-18T12:43:32.869000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                 11631.675
dataflow_v2_java17_total_bytes_count             2.05118615E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220118124120
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:838c3eea10304229b5fd7e1fc67a70751ca0462bcd48e213de3fc3e0d41564e5
Deleted: sha256:dec3a49df98102f9bf918ac9f4ee78f9f2d0b74793a226a3ecd7855f61f1d936
Deleted: sha256:976913ea87fb88ba30198e65f7f7d585c7d6f96a89d8744ffa2c97aca5fa9fe3
Deleted: sha256:08f50e416ab802c92b9106a5b281e2bafe04a0c7dfd068f9c21b8cab5e8d3586
Deleted: sha256:003b0e19bb34c2c6280cd2a979d431f38f72fcc7e050c84a73fd5dc70b580eda
Deleted: sha256:ad00f9cd90a223ae1a6ef282cbbebd069cc6d38fd4aa8774674ee8f50053d754
Deleted: sha256:04228b7b3f5c97e00782ef4dd809ebe7939f2e220dbfdcbf88979e2b2c04b505
Deleted: sha256:b0d0ccf73cb5f6b5cf830d524243f92be6afccbfb53f86a67d608247af5f7678
Deleted: sha256:e199022d0a222ab7b402d3efe84dd9066cd9e41fe15b83d270906af0c6b5e8a9
Deleted: sha256:2addf720b823a2dc97b99b76ae2067c5fc7885a03e9c82da18c063ce297f3378
Deleted: sha256:291900385a6b97e4d3944f2c9929e6a61ed8141bc6b495aa9b6cc0531b142e4c
Deleted: sha256:ec52bf5fcc07309afac3b4280b0e058efc3e5f2acdc7d237a384ed8bbdc52d32
Deleted: sha256:94b9e7e32da97d2ceaa63f2f9f2f5eacdab86bda0adf9eb10f071f8dd8220dd8
Deleted: sha256:eb23ecf6aa7ea52832f84a84c817c1393c2ae3fb6e1d0413c7db85988ef230c1
Deleted: sha256:3b5a9a7685dc8d08eee7381144cc4e7289edb705854d025d93352bd8416ae126
Deleted: sha256:d0a551c67509b100314438289bd5c7cd4361fc863123093dc3eedba33bebff33
Deleted: sha256:2813931d1b72e05ab06032ec22e462e0162e2a4cb3958409db8221c7309af68f
Deleted: sha256:30b1869278441f3c2c6c1d752fdf9e883ee8f9bb27a184c1805a51e1b724cd3d
Deleted: sha256:a428488a5e657655a179ca095812f39a12d7fff2c1367f30cbb86e4f5a66a4bf
Deleted: sha256:e14e156daebe35ab61bf35ef6e6097f20cac10bf1a97d38764f6e4b931d43985
Deleted: sha256:5c3f96bde105a0f464dac707d291824c964e17899bc82ed3d93b2327d43f0718
Deleted: sha256:06093fd08a0157a6bcb6f2934341ee2bcf059a9e489e3521e4899917b419ecb0
Deleted: sha256:652b0c97d27305773b0cbc5f7f6665108f473939e267375f67cd157b736760c1
Deleted: sha256:9b8bf9f2c5c6c1ce08c72267886582eb683aa4bc3c8c80fdcca4dd03f77262ce
Deleted: sha256:099bea83b5932905e020d56e6caf59b59a04e5e0c68b6c0a2b0bcace6a3546ad
Deleted: sha256:ff635755b9fe7295e318e137140ffc66b656a20c17612f54b54bb8d019093967
Deleted: sha256:c0bef9be011932565c7452f4d2d0ee70cace4c22a5579377a44945e8af8e6d2a
Deleted: sha256:6e3fc80abfd2e671311a9c4b01b7334a28ee4078f57054b87e766d1b8d84ffaa
Deleted: sha256:ecc72db241bdef070d279f7615b1aa0014cc2507896aaa40f0d347c05a3a16e9
Deleted: sha256:f7271e15e60872100e92c721ec27db256a6733eb8695a374236680bec9ce3c64
Deleted: sha256:cacc51581c5d57d3f0ceb89bc5ea41c56f16f6e850f06da5a0105a17c5272ed1
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220118124120]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:838c3eea10304229b5fd7e1fc67a70751ca0462bcd48e213de3fc3e0d41564e5]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220118124120] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:838c3eea10304229b5fd7e1fc67a70751ca0462bcd48e213de3fc3e0d41564e5])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:838c3eea10304229b5fd7e1fc67a70751ca0462bcd48e213de3fc3e0d41564e5
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:838c3eea10304229b5fd7e1fc67a70751ca0462bcd48e213de3fc3e0d41564e5
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:838c3eea10304229b5fd7e1fc67a70751ca0462bcd48e213de3fc3e0d41564e5].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 22m 5s
109 actionable tasks: 72 executed, 33 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/meb6wekdpb4y6

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #32

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/32/display/redirect?page=changes>

Changes:

[mmack] [BEAM-8806] Integration test for SqsIO using Localstack

[noreply] Merge pull request #16507: [BEAM-13137] Fixes ES utest size flakiness


------------------------------------------
[...truncated 49.12 KB...]
6bedb60eb8d5: Waiting
bff9fe6e429c: Waiting
20cbe2c032a1: Waiting
7c245b2fe4f1: Waiting
8bf42db0de72: Waiting
505399a542c8: Waiting
31892cc314cb: Waiting
f9e18e59a565: Waiting
b0a272f3006e: Pushed
3be094d3db6e: Pushed
aad777ec5fb0: Pushed
08bf6506e3dc: Pushed
587f6c5d7058: Pushed
da47ad0be2e9: Pushed
64d3cf9a30cc: Pushed
6bedb60eb8d5: Pushed
8ef85e676efc: Pushed
ddd544841562: Pushed
bff9fe6e429c: Layer already exists
ae5be93e21aa: Pushed
7c245b2fe4f1: Layer already exists
f9e18e59a565: Layer already exists
26a504e63be4: Layer already exists
8bf42db0de72: Layer already exists
31892cc314cb: Layer already exists
11936051f93b: Layer already exists
505399a542c8: Pushed
20cbe2c032a1: Pushed
20220117123034: digest: sha256:3722740481a78d117155532a9b4ea66aecae410a29fa840cc6d212f879d0af98 size: 4520

> Task :sdks:java:testing:load-tests:run
Jan 17, 2022 12:32:42 PM org.apache.beam.runners.dataflow.options.DataflowPipelineOptions$StagingLocationFactory create
INFO: No stagingLocation provided, falling back to gcpTempLocation
Jan 17, 2022 12:32:42 PM org.apache.beam.runners.dataflow.DataflowRunner fromOptions
INFO: PipelineOptions.filesToStage was not specified. Defaulting to files from the classpath: will stage 203 files. Enable logging at DEBUG level to see which files will be staged.
Jan 17, 2022 12:32:43 PM org.apache.beam.sdk.Pipeline validate
WARNING: The following transforms do not have stable unique names: Window.Into()
Jan 17, 2022 12:32:43 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Executing pipeline on the Dataflow Service, which will have billing implications related to Google Compute Engine usage and other Google Cloud Services.
Jan 17, 2022 12:32:45 PM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Uploading 203 files from PipelineOptions.filesToStage to staging location to prepare for execution.
Jan 17, 2022 12:32:46 PM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Staging files complete: 203 files cached, 0 files newly uploaded in 0 seconds
Jan 17, 2022 12:32:46 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Staging portable pipeline proto to gs://temp-storage-for-perf-tests/loadtests/staging/
Jan 17, 2022 12:32:46 PM org.apache.beam.runners.dataflow.util.PackageUtil tryStagePackage
INFO: Uploading <113691 bytes, hash 1fb8e4c0f6a2f8edbe7131f93af5fab06d5cc1837fb66036c0a776753f6856a1> to gs://temp-storage-for-perf-tests/loadtests/staging/pipeline-H7jkwPai-O2-cTH5OvX6sG1cwYN_tmA2wKd2dT9oVqE.pb
Jan 17, 2022 12:32:48 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read input/DataflowRunner.StreamingUnboundedRead.ReadWithIds as step s1
Jan 17, 2022 12:32:48 PM org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource split
INFO: Split into 20 bundles of sizes: [org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@60b34931, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4aa21f9d, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@71c17a57, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@640ab13c, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1e0a864d, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@440e3ce6, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3e67f5f2, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@fd53053, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4527f70a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@707b1a44, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7132a9dc, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@57435801, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2da66a44, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@527fc8e, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@61bfc9bf, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2c7106d9, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@329bad59, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@862f408, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@178f268a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@33ecbd6c]
Jan 17, 2022 12:32:48 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read input/StripIds as step s2
Jan 17, 2022 12:32:48 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect start time metrics (input) as step s3
Jan 17, 2022 12:32:48 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Window.Into()/Window.Assign as step s4
Jan 17, 2022 12:32:48 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read co-input/DataflowRunner.StreamingUnboundedRead.ReadWithIds as step s5
Jan 17, 2022 12:32:48 PM org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource split
INFO: Split into 20 bundles of sizes: [org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@b112b13, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@24eb65e3, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3ac3f6f, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1abebef3, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@18f55704, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@67cefd84, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5fbe155, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6add8e3f, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@58a2b917, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@48904d5a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@12bbfc54, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1491344a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@59b65dce, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1386313f, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5e922647, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@433c6abb, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@288f173f, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@a22c4d8, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@45cd7bc5, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6504a875]
Jan 17, 2022 12:32:48 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read co-input/StripIds as step s6
Jan 17, 2022 12:32:48 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect start time metrics (co-input) as step s7
Jan 17, 2022 12:32:48 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Window.Into()2/Window.Assign as step s8
Jan 17, 2022 12:32:48 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/MakeUnionTable0 as step s9
Jan 17, 2022 12:32:48 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/MakeUnionTable1 as step s10
Jan 17, 2022 12:32:48 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/Flatten as step s11
Jan 17, 2022 12:32:48 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/GBK as step s12
Jan 17, 2022 12:32:48 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/ConstructCoGbkResultFn as step s13
Jan 17, 2022 12:32:48 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Ungroup and reiterate as step s14
Jan 17, 2022 12:32:48 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect total bytes as step s15
Jan 17, 2022 12:32:48 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect end time metrics as step s16
Jan 17, 2022 12:32:48 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Dataflow SDK version: 2.37.0-SNAPSHOT
Jan 17, 2022 12:32:49 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To access the Dataflow monitoring console, please navigate to https://console.cloud.google.com/dataflow/jobs/us-central1/2022-01-17_04_32_48-11262187469055311349?project=apache-beam-testing
Jan 17, 2022 12:32:49 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Submitted job: 2022-01-17_04_32_48-11262187469055311349
Jan 17, 2022 12:32:49 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To cancel the job using the 'gcloud' tool, run:
> gcloud dataflow jobs --project=apache-beam-testing cancel --region=us-central1 2022-01-17_04_32_48-11262187469055311349
Jan 17, 2022 12:33:02 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
WARNING: 2022-01-17T12:33:01.318Z: The workflow name is not a valid Cloud Label. Labels applied to Cloud resources (such as GCE Instances) for monitoring will be labeled with this modified job name: load0tests0java170dataflow0v20streaming0cogbk01-jenkins-01-4ad6. For the best monitoring experience, please name your job with a valid Cloud Label. For details, see: https://cloud.google.com/compute/docs/labeling-resources#restrictions
Jan 17, 2022 12:33:07 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-17T12:33:06.411Z: Worker configuration: e2-standard-2 in us-central1-b.
Jan 17, 2022 12:33:07 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-17T12:33:06.999Z: Expanding SplittableParDo operations into optimizable parts.
Jan 17, 2022 12:33:07 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-17T12:33:07.037Z: Expanding CollectionToSingleton operations into optimizable parts.
Jan 17, 2022 12:33:10 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-17T12:33:07.122Z: Expanding CoGroupByKey operations into optimizable parts.
Jan 17, 2022 12:33:10 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-17T12:33:07.194Z: Expanding SplittableProcessKeyed operations into optimizable parts.
Jan 17, 2022 12:33:10 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-17T12:33:07.228Z: Expanding GroupByKey operations into streaming Read/Write steps
Jan 17, 2022 12:33:10 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-17T12:33:07.284Z: Lifting ValueCombiningMappingFns into MergeBucketsMappingFns
Jan 17, 2022 12:33:10 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-17T12:33:07.427Z: Fusing adjacent ParDo, Read, Write, and Flatten operations
Jan 17, 2022 12:33:10 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-17T12:33:07.464Z: Unzipping flatten CoGroupByKey-Flatten for input CoGroupByKey-MakeUnionTable0-ParMultiDo-ConstructUnionTable-.output
Jan 17, 2022 12:33:10 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-17T12:33:07.488Z: Fusing unzipped copy of CoGroupByKey/GBK/WriteStream, through flatten CoGroupByKey/Flatten, into producer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable)
Jan 17, 2022 12:33:10 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-17T12:33:07.511Z: Fusing consumer CoGroupByKey/GBK/WriteStream into CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable)
Jan 17, 2022 12:33:10 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-17T12:33:07.536Z: Fusing consumer Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read input/Impulse
Jan 17, 2022 12:33:10 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-17T12:33:07.596Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
Jan 17, 2022 12:33:10 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-17T12:33:07.620Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
Jan 17, 2022 12:33:10 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-17T12:33:07.653Z: Fusing consumer Read input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
Jan 17, 2022 12:33:10 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-17T12:33:07.688Z: Fusing consumer Collect start time metrics (input)/ParMultiDo(TimeMonitor) into Read input/ParDo(StripIds)/ParMultiDo(StripIds)
Jan 17, 2022 12:33:10 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-17T12:33:07.720Z: Fusing consumer Window.Into()/Window.Assign into Collect start time metrics (input)/ParMultiDo(TimeMonitor)
Jan 17, 2022 12:33:10 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-17T12:33:07.743Z: Fusing consumer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable) into Window.Into()/Window.Assign
Jan 17, 2022 12:33:10 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-17T12:33:07.774Z: Fusing consumer Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read co-input/Impulse
Jan 17, 2022 12:33:10 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-17T12:33:07.810Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
Jan 17, 2022 12:33:10 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-17T12:33:07.834Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
Jan 17, 2022 12:33:10 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-17T12:33:07.871Z: Fusing consumer Read co-input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
Jan 17, 2022 12:33:10 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-17T12:33:07.906Z: Fusing consumer Collect start time metrics (co-input)/ParMultiDo(TimeMonitor) into Read co-input/ParDo(StripIds)/ParMultiDo(StripIds)
Jan 17, 2022 12:33:10 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-17T12:33:07.942Z: Fusing consumer Window.Into()2/Window.Assign into Collect start time metrics (co-input)/ParMultiDo(TimeMonitor)
Jan 17, 2022 12:33:10 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-17T12:33:07.974Z: Fusing consumer CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable) into Window.Into()2/Window.Assign
Jan 17, 2022 12:33:10 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-17T12:33:08.010Z: Fusing consumer CoGroupByKey/GBK/MergeBuckets into CoGroupByKey/GBK/ReadStream
Jan 17, 2022 12:33:10 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-17T12:33:08.044Z: Fusing consumer CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult) into CoGroupByKey/GBK/MergeBuckets
Jan 17, 2022 12:33:10 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-17T12:33:08.088Z: Fusing consumer Ungroup and reiterate/ParMultiDo(UngroupAndReiterate) into CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult)
Jan 17, 2022 12:33:10 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-17T12:33:08.120Z: Fusing consumer Collect total bytes/ParMultiDo(ByteMonitor) into Ungroup and reiterate/ParMultiDo(UngroupAndReiterate)
Jan 17, 2022 12:33:10 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-17T12:33:08.151Z: Fusing consumer Collect end time metrics/ParMultiDo(TimeMonitor) into Collect total bytes/ParMultiDo(ByteMonitor)
Jan 17, 2022 12:33:10 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-17T12:33:08.559Z: Starting 5 ****s in us-central1-b...
Jan 17, 2022 12:33:36 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-17T12:33:35.803Z: Your project already contains 100 Dataflow-created metric descriptors, so new user metrics of the form custom.googleapis.com/* will not be created. However, all user metrics are also available in the metric dataflow.googleapis.com/job/user_counter. If you rely on the custom metrics, you can delete old / unused metric descriptors. See https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.list and https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.delete
Jan 17, 2022 12:34:00 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-17T12:33:59.283Z: Autoscaling: Raised the number of ****s to 5 so that the pipeline can catch up with its backlog and keep up with its input rate.
Jan 17, 2022 12:34:59 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-17T12:34:57.703Z: Workers have started successfully.
Jan 17, 2022 12:34:59 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-17T12:34:57.731Z: Workers have started successfully.
Jan 17, 2022 4:00:34 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-17T16:00:33.126Z: Cancel request is committed for workflow job: 2022-01-17_04_32_48-11262187469055311349.
Jan 17, 2022 4:00:34 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-17T16:00:33.281Z: Cleaning up.
Jan 17, 2022 4:00:34 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-17T16:00:33.399Z: Stopping **** pool...
Jan 17, 2022 4:00:34 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-17T16:00:33.443Z: Stopping **** pool...
Jan 17, 2022 4:03:00 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-17T16:02:59.497Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Jan 17, 2022 4:03:00 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-17T16:02:59.553Z: Worker pool stopped.
Jan 17, 2022 4:03:05 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-01-17_04_32_48-11262187469055311349 finished with status CANCELLED.
Load test results for test (ID): 7e7194da-3aa8-4942-a86e-003acb796371 and timestamp: 2022-01-17T12:32:43.312000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                 12248.312
dataflow_v2_java17_total_bytes_count             1.69792821E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220117123034
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:3722740481a78d117155532a9b4ea66aecae410a29fa840cc6d212f879d0af98
Deleted: sha256:810d0246cf094bf8af97d2eeb39e9e51302cd62f5af4969180eabe4a4da7a549
Deleted: sha256:7f5287f1eee43f81df1b7c48b8f7320428320a55a8f8f0c4f2e2bacb5bc2afee
Deleted: sha256:44d07045b9a3446d794945de7d845247722a5466e50d2109d18de16348fec4e7
Deleted: sha256:7482267b5d56fe46f0fec022a83656b2490583e6a66579a35e19bc36eaa571cd
Deleted: sha256:e3a6a88597ca0074c80373094ade3c4ddf080aed6219f4b3ace1adcb849b7413
Deleted: sha256:78047f8e6ae8825e0a9f7c9f102da5905c037d1069863baef72e31ea9f2763b0
Deleted: sha256:d14820011712c49c5af914efa72025e40574955580c7dc40a0390f18c574836e
Deleted: sha256:dfb349379f26ce08e6fdb10a6fceef404cf9181e7747bf1063ef0f73c027a3f5
Deleted: sha256:63ddb793ccf743b7303b09f247b783a2851bcc78925e5dbf2d451a9c7e56c664
Deleted: sha256:64e1f860d961befe4b6258c75477b99b10a8f12c9fa6c812ef1f2edc29c8e55b
Deleted: sha256:e27395e5392a18dc19268d6b0fb72a32d61057f011e57362395d62bb13f6b6ca
Deleted: sha256:14d24537863060434f892c1c7a1a043d56d8b7557f3a2defdab594a17eed977e
Deleted: sha256:11be59ddaccaa73d54f10823bf00834bd7adbffc3a9d94f7c25067700132f241
Deleted: sha256:179dd84a6dcd1d87b5d9c6fd4eab3e05e943cac1bdb875579de8c2e31e449fa0
Deleted: sha256:177d3cb7df7d94fe8b7a6d369fec89fa750ebc405505473e96740292ca7cf440
Deleted: sha256:aa1cd4e2c3915520001c69c08fc2c226f3f094094d35a6027a8bc493592eb611
Deleted: sha256:192a0efac62bfb344dcc4099bbbd3d6ed9f250cff1035d8b43c73777e14f450d
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220117123034]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:3722740481a78d117155532a9b4ea66aecae410a29fa840cc6d212f879d0af98]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220117123034] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:3722740481a78d117155532a9b4ea66aecae410a29fa840cc6d212f879d0af98])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:3722740481a78d117155532a9b4ea66aecae410a29fa840cc6d212f879d0af98
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:3722740481a78d117155532a9b4ea66aecae410a29fa840cc6d212f879d0af98
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:3722740481a78d117155532a9b4ea66aecae410a29fa840cc6d212f879d0af98].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 32m 48s
109 actionable tasks: 72 executed, 33 from cache, 4 up-to-date

Publishing build scan...
https://scans.gradle.com/s/2l4xlyx5u77am

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #31

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/31/display/redirect>

Changes:


------------------------------------------
[...truncated 49.69 KB...]
9ef7d3a9807c: Preparing
9ad2e23d7801: Preparing
9d249883d643: Preparing
f3ab44eef6ce: Preparing
bceeb55ed9c0: Waiting
2e7e34f9ecfa: Preparing
7cec769fce73: Waiting
bff9fe6e429c: Preparing
bb49150f06cf: Waiting
7c245b2fe4f1: Preparing
f9e18e59a565: Preparing
9ef7d3a9807c: Waiting
26a504e63be4: Preparing
9ad2e23d7801: Waiting
8bf42db0de72: Preparing
31892cc314cb: Preparing
11936051f93b: Preparing
9d249883d643: Waiting
bff9fe6e429c: Waiting
7c245b2fe4f1: Waiting
f3ab44eef6ce: Waiting
f9e18e59a565: Waiting
8bf42db0de72: Waiting
11936051f93b: Waiting
26a504e63be4: Waiting
5057c979272e: Pushed
f81272eabb31: Pushed
0c43e11764ae: Pushed
2a0680d7005c: Pushed
bceeb55ed9c0: Pushed
bb49150f06cf: Pushed
f3be89cb631b: Pushed
9ad2e23d7801: Pushed
f3ab44eef6ce: Pushed
9ef7d3a9807c: Pushed
7cec769fce73: Pushed
bff9fe6e429c: Layer already exists
7c245b2fe4f1: Layer already exists
f9e18e59a565: Layer already exists
26a504e63be4: Layer already exists
8bf42db0de72: Layer already exists
31892cc314cb: Layer already exists
11936051f93b: Layer already exists
2e7e34f9ecfa: Pushed
9d249883d643: Pushed
20220116123033: digest: sha256:879140a2f0503f7fdd9814fb9526c1ed10806ccda9668bb2898c185419894c2d size: 4520

> Task :sdks:java:testing:load-tests:run
Jan 16, 2022 12:32:16 PM org.apache.beam.runners.dataflow.options.DataflowPipelineOptions$StagingLocationFactory create
INFO: No stagingLocation provided, falling back to gcpTempLocation
Jan 16, 2022 12:32:16 PM org.apache.beam.runners.dataflow.DataflowRunner fromOptions
INFO: PipelineOptions.filesToStage was not specified. Defaulting to files from the classpath: will stage 203 files. Enable logging at DEBUG level to see which files will be staged.
Jan 16, 2022 12:32:17 PM org.apache.beam.sdk.Pipeline validate
WARNING: The following transforms do not have stable unique names: Window.Into()
Jan 16, 2022 12:32:17 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Executing pipeline on the Dataflow Service, which will have billing implications related to Google Compute Engine usage and other Google Cloud Services.
Jan 16, 2022 12:32:20 PM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Uploading 203 files from PipelineOptions.filesToStage to staging location to prepare for execution.
Jan 16, 2022 12:32:20 PM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Staging files complete: 203 files cached, 0 files newly uploaded in 0 seconds
Jan 16, 2022 12:32:20 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Staging portable pipeline proto to gs://temp-storage-for-perf-tests/loadtests/staging/
Jan 16, 2022 12:32:20 PM org.apache.beam.runners.dataflow.util.PackageUtil tryStagePackage
INFO: Uploading <113691 bytes, hash ff1cb94c0bc67b9294ac703c031aac0abe3f57dc3490d18c384d585e8f2ca390> to gs://temp-storage-for-perf-tests/loadtests/staging/pipeline-_xy5TAvGe5KUrHA8AxqsCr4_V9w0kNGMOE1YXo8so5A.pb
Jan 16, 2022 12:32:22 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read input/DataflowRunner.StreamingUnboundedRead.ReadWithIds as step s1
Jan 16, 2022 12:32:22 PM org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource split
INFO: Split into 20 bundles of sizes: [org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@60b34931, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4aa21f9d, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@71c17a57, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@640ab13c, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1e0a864d, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@440e3ce6, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3e67f5f2, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@fd53053, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4527f70a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@707b1a44, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7132a9dc, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@57435801, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2da66a44, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@527fc8e, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@61bfc9bf, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2c7106d9, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@329bad59, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@862f408, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@178f268a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@33ecbd6c]
Jan 16, 2022 12:32:22 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read input/StripIds as step s2
Jan 16, 2022 12:32:22 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect start time metrics (input) as step s3
Jan 16, 2022 12:32:22 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Window.Into()/Window.Assign as step s4
Jan 16, 2022 12:32:22 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read co-input/DataflowRunner.StreamingUnboundedRead.ReadWithIds as step s5
Jan 16, 2022 12:32:22 PM org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource split
INFO: Split into 20 bundles of sizes: [org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@b112b13, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@24eb65e3, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3ac3f6f, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1abebef3, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@18f55704, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@67cefd84, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5fbe155, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6add8e3f, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@58a2b917, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@48904d5a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@12bbfc54, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1491344a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@59b65dce, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1386313f, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5e922647, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@433c6abb, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@288f173f, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@a22c4d8, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@45cd7bc5, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6504a875]
Jan 16, 2022 12:32:22 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read co-input/StripIds as step s6
Jan 16, 2022 12:32:22 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect start time metrics (co-input) as step s7
Jan 16, 2022 12:32:22 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Window.Into()2/Window.Assign as step s8
Jan 16, 2022 12:32:22 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/MakeUnionTable0 as step s9
Jan 16, 2022 12:32:22 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/MakeUnionTable1 as step s10
Jan 16, 2022 12:32:22 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/Flatten as step s11
Jan 16, 2022 12:32:22 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/GBK as step s12
Jan 16, 2022 12:32:22 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/ConstructCoGbkResultFn as step s13
Jan 16, 2022 12:32:22 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Ungroup and reiterate as step s14
Jan 16, 2022 12:32:22 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect total bytes as step s15
Jan 16, 2022 12:32:22 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect end time metrics as step s16
Jan 16, 2022 12:32:22 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Dataflow SDK version: 2.37.0-SNAPSHOT
Jan 16, 2022 12:32:23 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To access the Dataflow monitoring console, please navigate to https://console.cloud.google.com/dataflow/jobs/us-central1/2022-01-16_04_32_22-7158003769059064304?project=apache-beam-testing
Jan 16, 2022 12:32:23 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Submitted job: 2022-01-16_04_32_22-7158003769059064304
Jan 16, 2022 12:32:23 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To cancel the job using the 'gcloud' tool, run:
> gcloud dataflow jobs --project=apache-beam-testing cancel --region=us-central1 2022-01-16_04_32_22-7158003769059064304
Jan 16, 2022 12:32:33 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
WARNING: 2022-01-16T12:32:30.258Z: The workflow name is not a valid Cloud Label. Labels applied to Cloud resources (such as GCE Instances) for monitoring will be labeled with this modified job name: load0tests0java170dataflow0v20streaming0cogbk01-jenkins-01-hova. For the best monitoring experience, please name your job with a valid Cloud Label. For details, see: https://cloud.google.com/compute/docs/labeling-resources#restrictions
Jan 16, 2022 12:32:37 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-16T12:32:34.359Z: Worker configuration: e2-standard-2 in us-central1-b.
Jan 16, 2022 12:32:37 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-16T12:32:35.002Z: Expanding SplittableParDo operations into optimizable parts.
Jan 16, 2022 12:32:37 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-16T12:32:35.037Z: Expanding CollectionToSingleton operations into optimizable parts.
Jan 16, 2022 12:32:37 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-16T12:32:35.157Z: Expanding CoGroupByKey operations into optimizable parts.
Jan 16, 2022 12:32:37 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-16T12:32:35.218Z: Expanding SplittableProcessKeyed operations into optimizable parts.
Jan 16, 2022 12:32:37 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-16T12:32:35.257Z: Expanding GroupByKey operations into streaming Read/Write steps
Jan 16, 2022 12:32:37 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-16T12:32:35.303Z: Lifting ValueCombiningMappingFns into MergeBucketsMappingFns
Jan 16, 2022 12:32:37 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-16T12:32:35.392Z: Fusing adjacent ParDo, Read, Write, and Flatten operations
Jan 16, 2022 12:32:37 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-16T12:32:35.424Z: Unzipping flatten CoGroupByKey-Flatten for input CoGroupByKey-MakeUnionTable0-ParMultiDo-ConstructUnionTable-.output
Jan 16, 2022 12:32:37 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-16T12:32:35.454Z: Fusing unzipped copy of CoGroupByKey/GBK/WriteStream, through flatten CoGroupByKey/Flatten, into producer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable)
Jan 16, 2022 12:32:37 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-16T12:32:35.487Z: Fusing consumer CoGroupByKey/GBK/WriteStream into CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable)
Jan 16, 2022 12:32:37 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-16T12:32:35.523Z: Fusing consumer Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read input/Impulse
Jan 16, 2022 12:32:37 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-16T12:32:35.547Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
Jan 16, 2022 12:32:37 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-16T12:32:35.579Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
Jan 16, 2022 12:32:37 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-16T12:32:35.614Z: Fusing consumer Read input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
Jan 16, 2022 12:32:37 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-16T12:32:35.646Z: Fusing consumer Collect start time metrics (input)/ParMultiDo(TimeMonitor) into Read input/ParDo(StripIds)/ParMultiDo(StripIds)
Jan 16, 2022 12:32:37 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-16T12:32:35.673Z: Fusing consumer Window.Into()/Window.Assign into Collect start time metrics (input)/ParMultiDo(TimeMonitor)
Jan 16, 2022 12:32:37 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-16T12:32:35.695Z: Fusing consumer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable) into Window.Into()/Window.Assign
Jan 16, 2022 12:32:37 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-16T12:32:35.722Z: Fusing consumer Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read co-input/Impulse
Jan 16, 2022 12:32:37 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-16T12:32:35.766Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
Jan 16, 2022 12:32:37 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-16T12:32:35.794Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
Jan 16, 2022 12:32:37 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-16T12:32:35.823Z: Fusing consumer Read co-input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
Jan 16, 2022 12:32:37 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-16T12:32:35.859Z: Fusing consumer Collect start time metrics (co-input)/ParMultiDo(TimeMonitor) into Read co-input/ParDo(StripIds)/ParMultiDo(StripIds)
Jan 16, 2022 12:32:37 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-16T12:32:35.891Z: Fusing consumer Window.Into()2/Window.Assign into Collect start time metrics (co-input)/ParMultiDo(TimeMonitor)
Jan 16, 2022 12:32:37 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-16T12:32:35.918Z: Fusing consumer CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable) into Window.Into()2/Window.Assign
Jan 16, 2022 12:32:37 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-16T12:32:35.949Z: Fusing consumer CoGroupByKey/GBK/MergeBuckets into CoGroupByKey/GBK/ReadStream
Jan 16, 2022 12:32:37 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-16T12:32:35.983Z: Fusing consumer CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult) into CoGroupByKey/GBK/MergeBuckets
Jan 16, 2022 12:32:37 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-16T12:32:36.013Z: Fusing consumer Ungroup and reiterate/ParMultiDo(UngroupAndReiterate) into CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult)
Jan 16, 2022 12:32:37 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-16T12:32:36.046Z: Fusing consumer Collect total bytes/ParMultiDo(ByteMonitor) into Ungroup and reiterate/ParMultiDo(UngroupAndReiterate)
Jan 16, 2022 12:32:37 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-16T12:32:36.081Z: Fusing consumer Collect end time metrics/ParMultiDo(TimeMonitor) into Collect total bytes/ParMultiDo(ByteMonitor)
Jan 16, 2022 12:32:37 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-16T12:32:36.418Z: Starting 5 ****s in us-central1-b...
Jan 16, 2022 12:32:53 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-16T12:32:53.433Z: Your project already contains 100 Dataflow-created metric descriptors, so new user metrics of the form custom.googleapis.com/* will not be created. However, all user metrics are also available in the metric dataflow.googleapis.com/job/user_counter. If you rely on the custom metrics, you can delete old / unused metric descriptors. See https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.list and https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.delete
Jan 16, 2022 12:33:28 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-16T12:33:27.616Z: Autoscaling: Raised the number of ****s to 5 so that the pipeline can catch up with its backlog and keep up with its input rate.
Jan 16, 2022 12:34:25 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-16T12:34:24.497Z: Workers have started successfully.
Jan 16, 2022 12:34:25 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-16T12:34:24.531Z: Workers have started successfully.
Jan 16, 2022 4:00:39 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-16T16:00:36.574Z: Cancel request is committed for workflow job: 2022-01-16_04_32_22-7158003769059064304.
Jan 16, 2022 4:00:39 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-16T16:00:36.638Z: Cleaning up.
Jan 16, 2022 4:00:39 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-16T16:00:36.707Z: Stopping **** pool...
Jan 16, 2022 4:00:39 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-16T16:00:36.749Z: Stopping **** pool...
Jan 16, 2022 4:03:03 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-16T16:03:03.543Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Jan 16, 2022 4:03:03 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-16T16:03:03.588Z: Worker pool stopped.
Jan 16, 2022 4:03:09 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-01-16_04_32_22-7158003769059064304 finished with status CANCELLED.
Load test results for test (ID): 8e7a66a4-9fb6-4858-9915-d2903ab657f4 and timestamp: 2022-01-16T12:32:17.193000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                 12347.644
dataflow_v2_java17_total_bytes_count             2.27034289E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220116123033
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:879140a2f0503f7fdd9814fb9526c1ed10806ccda9668bb2898c185419894c2d
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220116123033]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:879140a2f0503f7fdd9814fb9526c1ed10806ccda9668bb2898c185419894c2d]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220116123033] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:879140a2f0503f7fdd9814fb9526c1ed10806ccda9668bb2898c185419894c2d])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:879140a2f0503f7fdd9814fb9526c1ed10806ccda9668bb2898c185419894c2d
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:879140a2f0503f7fdd9814fb9526c1ed10806ccda9668bb2898c185419894c2d
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:879140a2f0503f7fdd9814fb9526c1ed10806ccda9668bb2898c185419894c2d].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 32m 51s
109 actionable tasks: 72 executed, 33 from cache, 4 up-to-date

Publishing build scan...
https://scans.gradle.com/s/i5acpsvfzfepy

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #30

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/30/display/redirect?page=changes>

Changes:

[daria.malkova] Split builder into several builder for each step of pipeline execution

[Valentyn Tymofieiev] Provide API to check whether a hint is known.

[Valentyn Tymofieiev] [BEAM-12558] Fix doc typo.

[arietis27] [BEAM-13400] JDBC IO does not support UUID and JSONB PostgreSQL types

[jrmccluskey] [BEAM-10206] Resolve go vet errors in protox package

[noreply] Merge pull request #16482 from [BEAM-13429][Playground] Add builder for

[noreply] [BEAM-13590] Fix  abc imports from collections (#15850)

[jrmccluskey] Fix staticcheck errors in transforms directory

[jrmccluskey] Remove unnecessary fmt.Sprintf() in partition.go

[jrmccluskey] Replace bytes.Compare() with bytes.Equal() in test cases

[jrmccluskey] Replace string(buf.Bytes()) with buf.String() in coder_test.go

[jrmccluskey] Remove unnecessary blank identifier assignment in harness.go

[jrmccluskey] fix capitalized error strings in expansionx

[jrmccluskey] Clean up string cast of bytes in vet.go and corresponding tests

[jrmccluskey] Remove unnecessary fmt call in universal.go

[Robert Bradshaw] Remove tab from source.

[noreply] Redirecting cross-language transforms content (#16504)

[noreply] doc tweaks (#16498)

[noreply] [BEAM-12621] - Update Jenkins VMs to modern Ubuntu version (#16457)

[noreply] [BEAM-13664] Fix Primitives hashing benchmark (#16523)


------------------------------------------
[...truncated 49.54 KB...]
160b539ba9c2: Preparing
262398613c23: Preparing
bedc7db8444f: Preparing
2505906f4755: Preparing
5bc7edccbdb7: Preparing
a9ba48bdfb7b: Preparing
bff9fe6e429c: Preparing
7c245b2fe4f1: Preparing
f9e18e59a565: Preparing
26a504e63be4: Preparing
8bf42db0de72: Preparing
31892cc314cb: Preparing
11936051f93b: Preparing
26a504e63be4: Waiting
7c245b2fe4f1: Waiting
8bf42db0de72: Waiting
f9e18e59a565: Waiting
160b539ba9c2: Waiting
262398613c23: Waiting
bde2377ec88f: Waiting
31892cc314cb: Waiting
a9ba48bdfb7b: Waiting
5bc7edccbdb7: Waiting
bff9fe6e429c: Waiting
11936051f93b: Waiting
ebc60f590e19: Pushed
fe11e886fe54: Pushed
e4a72a779215: Pushed
f39189475ed5: Pushed
2e8ebb2b408a: Pushed
15d83f966159: Pushed
160b539ba9c2: Pushed
bedc7db8444f: Pushed
262398613c23: Pushed
5bc7edccbdb7: Pushed
bff9fe6e429c: Layer already exists
bde2377ec88f: Pushed
7c245b2fe4f1: Layer already exists
f9e18e59a565: Layer already exists
26a504e63be4: Layer already exists
8bf42db0de72: Layer already exists
31892cc314cb: Layer already exists
11936051f93b: Layer already exists
a9ba48bdfb7b: Pushed
2505906f4755: Pushed
20220115123043: digest: sha256:ab2f4d2492b6a7231900065e109fd520deaa4c38e3a406eae1cd58ef9785ea8f size: 4520

> Task :sdks:java:testing:load-tests:run
Jan 15, 2022 12:32:49 PM org.apache.beam.runners.dataflow.options.DataflowPipelineOptions$StagingLocationFactory create
INFO: No stagingLocation provided, falling back to gcpTempLocation
Jan 15, 2022 12:32:50 PM org.apache.beam.runners.dataflow.DataflowRunner fromOptions
INFO: PipelineOptions.filesToStage was not specified. Defaulting to files from the classpath: will stage 203 files. Enable logging at DEBUG level to see which files will be staged.
Jan 15, 2022 12:32:51 PM org.apache.beam.sdk.Pipeline validate
WARNING: The following transforms do not have stable unique names: Window.Into()
Jan 15, 2022 12:32:51 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Executing pipeline on the Dataflow Service, which will have billing implications related to Google Compute Engine usage and other Google Cloud Services.
Jan 15, 2022 12:32:53 PM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Uploading 203 files from PipelineOptions.filesToStage to staging location to prepare for execution.
Jan 15, 2022 12:32:54 PM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Staging files complete: 203 files cached, 0 files newly uploaded in 0 seconds
Jan 15, 2022 12:32:54 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Staging portable pipeline proto to gs://temp-storage-for-perf-tests/loadtests/staging/
Jan 15, 2022 12:32:54 PM org.apache.beam.runners.dataflow.util.PackageUtil tryStagePackage
INFO: Uploading <113691 bytes, hash 748dd753147541b48d994b69c7544e1070c40503642698519c93b956088a845c> to gs://temp-storage-for-perf-tests/loadtests/staging/pipeline-dI3XUxR1QbSNmUtpx1ROEHDEBQNkJphRnJO5VgiKhFw.pb
Jan 15, 2022 12:32:56 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read input/DataflowRunner.StreamingUnboundedRead.ReadWithIds as step s1
Jan 15, 2022 12:32:56 PM org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource split
INFO: Split into 20 bundles of sizes: [org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@71c17a57, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@640ab13c, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1e0a864d, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@440e3ce6, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3e67f5f2, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@fd53053, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4527f70a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@707b1a44, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7132a9dc, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@57435801, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2da66a44, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@527fc8e, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@61bfc9bf, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2c7106d9, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@329bad59, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@862f408, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@178f268a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@33ecbd6c, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5c723f2d, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@432f521f]
Jan 15, 2022 12:32:56 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read input/StripIds as step s2
Jan 15, 2022 12:32:56 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect start time metrics (input) as step s3
Jan 15, 2022 12:32:56 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Window.Into()/Window.Assign as step s4
Jan 15, 2022 12:32:56 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read co-input/DataflowRunner.StreamingUnboundedRead.ReadWithIds as step s5
Jan 15, 2022 12:32:56 PM org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource split
INFO: Split into 20 bundles of sizes: [org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3ac3f6f, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1abebef3, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@18f55704, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@67cefd84, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5fbe155, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6add8e3f, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@58a2b917, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@48904d5a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@12bbfc54, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1491344a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@59b65dce, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1386313f, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5e922647, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@433c6abb, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@288f173f, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@a22c4d8, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@45cd7bc5, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6504a875, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@35e26d05, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@29fa6b65]
Jan 15, 2022 12:32:56 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read co-input/StripIds as step s6
Jan 15, 2022 12:32:56 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect start time metrics (co-input) as step s7
Jan 15, 2022 12:32:56 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Window.Into()2/Window.Assign as step s8
Jan 15, 2022 12:32:56 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/MakeUnionTable0 as step s9
Jan 15, 2022 12:32:56 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/MakeUnionTable1 as step s10
Jan 15, 2022 12:32:56 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/Flatten as step s11
Jan 15, 2022 12:32:56 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/GBK as step s12
Jan 15, 2022 12:32:56 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/ConstructCoGbkResultFn as step s13
Jan 15, 2022 12:32:56 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Ungroup and reiterate as step s14
Jan 15, 2022 12:32:56 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect total bytes as step s15
Jan 15, 2022 12:32:56 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect end time metrics as step s16
Jan 15, 2022 12:32:56 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Dataflow SDK version: 2.37.0-SNAPSHOT
Jan 15, 2022 12:32:57 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To access the Dataflow monitoring console, please navigate to https://console.cloud.google.com/dataflow/jobs/us-central1/2022-01-15_04_32_56-17676102161441551514?project=apache-beam-testing
Jan 15, 2022 12:32:57 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Submitted job: 2022-01-15_04_32_56-17676102161441551514
Jan 15, 2022 12:32:57 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To cancel the job using the 'gcloud' tool, run:
> gcloud dataflow jobs --project=apache-beam-testing cancel --region=us-central1 2022-01-15_04_32_56-17676102161441551514
Jan 15, 2022 12:33:04 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
WARNING: 2022-01-15T12:33:04.045Z: The workflow name is not a valid Cloud Label. Labels applied to Cloud resources (such as GCE Instances) for monitoring will be labeled with this modified job name: load0tests0java170dataflow0v20streaming0cogbk01-jenkins-01-eipm. For the best monitoring experience, please name your job with a valid Cloud Label. For details, see: https://cloud.google.com/compute/docs/labeling-resources#restrictions
Jan 15, 2022 12:33:10 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-15T12:33:08.117Z: Worker configuration: e2-standard-2 in us-central1-b.
Jan 15, 2022 12:33:10 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-15T12:33:08.688Z: Expanding SplittableParDo operations into optimizable parts.
Jan 15, 2022 12:33:10 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-15T12:33:08.724Z: Expanding CollectionToSingleton operations into optimizable parts.
Jan 15, 2022 12:33:10 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-15T12:33:08.790Z: Expanding CoGroupByKey operations into optimizable parts.
Jan 15, 2022 12:33:10 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-15T12:33:08.865Z: Expanding SplittableProcessKeyed operations into optimizable parts.
Jan 15, 2022 12:33:10 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-15T12:33:08.894Z: Expanding GroupByKey operations into streaming Read/Write steps
Jan 15, 2022 12:33:10 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-15T12:33:08.961Z: Lifting ValueCombiningMappingFns into MergeBucketsMappingFns
Jan 15, 2022 12:33:10 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-15T12:33:09.133Z: Fusing adjacent ParDo, Read, Write, and Flatten operations
Jan 15, 2022 12:33:10 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-15T12:33:09.162Z: Unzipping flatten CoGroupByKey-Flatten for input CoGroupByKey-MakeUnionTable0-ParMultiDo-ConstructUnionTable-.output
Jan 15, 2022 12:33:10 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-15T12:33:09.195Z: Fusing unzipped copy of CoGroupByKey/GBK/WriteStream, through flatten CoGroupByKey/Flatten, into producer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable)
Jan 15, 2022 12:33:10 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-15T12:33:09.229Z: Fusing consumer CoGroupByKey/GBK/WriteStream into CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable)
Jan 15, 2022 12:33:10 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-15T12:33:09.262Z: Fusing consumer Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read input/Impulse
Jan 15, 2022 12:33:10 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-15T12:33:09.299Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
Jan 15, 2022 12:33:10 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-15T12:33:09.326Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
Jan 15, 2022 12:33:10 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-15T12:33:09.360Z: Fusing consumer Read input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
Jan 15, 2022 12:33:10 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-15T12:33:09.389Z: Fusing consumer Collect start time metrics (input)/ParMultiDo(TimeMonitor) into Read input/ParDo(StripIds)/ParMultiDo(StripIds)
Jan 15, 2022 12:33:10 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-15T12:33:09.422Z: Fusing consumer Window.Into()/Window.Assign into Collect start time metrics (input)/ParMultiDo(TimeMonitor)
Jan 15, 2022 12:33:10 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-15T12:33:09.456Z: Fusing consumer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable) into Window.Into()/Window.Assign
Jan 15, 2022 12:33:10 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-15T12:33:09.490Z: Fusing consumer Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read co-input/Impulse
Jan 15, 2022 12:33:10 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-15T12:33:09.521Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
Jan 15, 2022 12:33:10 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-15T12:33:09.557Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
Jan 15, 2022 12:33:10 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-15T12:33:09.589Z: Fusing consumer Read co-input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
Jan 15, 2022 12:33:10 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-15T12:33:09.624Z: Fusing consumer Collect start time metrics (co-input)/ParMultiDo(TimeMonitor) into Read co-input/ParDo(StripIds)/ParMultiDo(StripIds)
Jan 15, 2022 12:33:10 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-15T12:33:09.661Z: Fusing consumer Window.Into()2/Window.Assign into Collect start time metrics (co-input)/ParMultiDo(TimeMonitor)
Jan 15, 2022 12:33:10 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-15T12:33:09.723Z: Fusing consumer CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable) into Window.Into()2/Window.Assign
Jan 15, 2022 12:33:10 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-15T12:33:09.752Z: Fusing consumer CoGroupByKey/GBK/MergeBuckets into CoGroupByKey/GBK/ReadStream
Jan 15, 2022 12:33:10 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-15T12:33:09.785Z: Fusing consumer CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult) into CoGroupByKey/GBK/MergeBuckets
Jan 15, 2022 12:33:10 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-15T12:33:09.819Z: Fusing consumer Ungroup and reiterate/ParMultiDo(UngroupAndReiterate) into CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult)
Jan 15, 2022 12:33:10 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-15T12:33:09.852Z: Fusing consumer Collect total bytes/ParMultiDo(ByteMonitor) into Ungroup and reiterate/ParMultiDo(UngroupAndReiterate)
Jan 15, 2022 12:33:10 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-15T12:33:09.888Z: Fusing consumer Collect end time metrics/ParMultiDo(TimeMonitor) into Collect total bytes/ParMultiDo(ByteMonitor)
Jan 15, 2022 12:33:10 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-15T12:33:10.276Z: Starting 5 ****s in us-central1-b...
Jan 15, 2022 12:33:21 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-15T12:33:19.069Z: Your project already contains 100 Dataflow-created metric descriptors, so new user metrics of the form custom.googleapis.com/* will not be created. However, all user metrics are also available in the metric dataflow.googleapis.com/job/user_counter. If you rely on the custom metrics, you can delete old / unused metric descriptors. See https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.list and https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.delete
Jan 15, 2022 12:33:50 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-15T12:33:49.457Z: Autoscaling: Raised the number of ****s to 5 so that the pipeline can catch up with its backlog and keep up with its input rate.
Jan 15, 2022 12:34:46 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-15T12:34:45.393Z: Workers have started successfully.
Jan 15, 2022 12:34:46 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-15T12:34:45.471Z: Workers have started successfully.
Jan 15, 2022 4:00:39 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-15T16:00:38.417Z: Cancel request is committed for workflow job: 2022-01-15_04_32_56-17676102161441551514.
Jan 15, 2022 4:00:39 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-15T16:00:38.491Z: Cleaning up.
Jan 15, 2022 4:00:39 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-15T16:00:38.552Z: Stopping **** pool...
Jan 15, 2022 4:00:39 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-15T16:00:38.595Z: Stopping **** pool...
Jan 15, 2022 4:03:00 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-15T16:03:00.237Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Jan 15, 2022 4:03:00 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-15T16:03:00.285Z: Worker pool stopped.
Jan 15, 2022 4:03:06 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-01-15_04_32_56-17676102161441551514 finished with status CANCELLED.
Load test results for test (ID): df408673-d0c7-4fc8-bb2b-dd6f6de0861f and timestamp: 2022-01-15T12:32:50.725000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                 12303.035
dataflow_v2_java17_total_bytes_count             3.18893904E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220115123043
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:ab2f4d2492b6a7231900065e109fd520deaa4c38e3a406eae1cd58ef9785ea8f
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220115123043]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:ab2f4d2492b6a7231900065e109fd520deaa4c38e3a406eae1cd58ef9785ea8f]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220115123043] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:ab2f4d2492b6a7231900065e109fd520deaa4c38e3a406eae1cd58ef9785ea8f])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:ab2f4d2492b6a7231900065e109fd520deaa4c38e3a406eae1cd58ef9785ea8f
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:ab2f4d2492b6a7231900065e109fd520deaa4c38e3a406eae1cd58ef9785ea8f
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:ab2f4d2492b6a7231900065e109fd520deaa4c38e3a406eae1cd58ef9785ea8f].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 32m 49s
109 actionable tasks: 72 executed, 33 from cache, 4 up-to-date

Publishing build scan...
https://scans.gradle.com/s/ukblg24k7abm6

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #29

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/29/display/redirect?page=changes>

Changes:

[blais] python sdk examples: Fixed typo in wordcount example.

[noreply] [BEAM-13480] Increase pipeline timeout for

[noreply] Stronger typing inference for CoGBK. (#16465)

[noreply] [BEAM-12464] Change ProtoSchemaTranslator beam schema creation to match

[noreply] Introduce the notion of a JoinIndex for fewer shuffles. (#16101)

[noreply] Merge pull request #16467 from [BEAM-12164]: SpannerIO

[noreply] Merge pull request #16385 from [BEAM-13535] [Playground] add cancel

[noreply] Merge pull request #16485 from [BEAM-13486] [Playground] For unit tests

[heejong] [BEAM-13455] Remove duplicated artifacts when using multiple

[noreply] [BEAM-12572] Run java examples on multiple runners (#16450)


------------------------------------------
[...truncated 1.78 MB...]
	at java.io.RandomAccessFile.write(RandomAccessFile.java:525)
	at org.gradle.internal.io.RandomAccessFileOutputStream.write(RandomAccessFileOutputStream.java:46)
	at java.io.BufferedOutputStream.flushBuffer(BufferedOutputStream.java:82)
	at java.io.BufferedOutputStream.flush(BufferedOutputStream.java:140)
	at java.io.FilterOutputStream.flush(FilterOutputStream.java:140)
	at org.gradle.cache.internal.btree.ByteOutput.done(ByteOutput.java:62)
	at org.gradle.cache.internal.btree.FileBackedBlockStore$BlockImpl.write(FileBackedBlockStore.java:223)
	at org.gradle.cache.internal.btree.FileBackedBlockStore.write(FileBackedBlockStore.java:136)
	... 23 more
Failed to execute org.gradle.cache.internal.AsyncCacheAccessDecoratedCache$$Lambda$222/1355592988@4d0024d1.
org.gradle.api.UncheckedIOException: Could not add entry '<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/build/target/java_third_party_licenses/kafka-clients-2.2.2.jar/common/message/README.md'> to cache fileHashes.bin (<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/.gradle/7.3.2/fileHashes/fileHashes.bin).>
	at org.gradle.cache.internal.btree.BTreePersistentIndexedCache.put(BTreePersistentIndexedCache.java:162)
	at org.gradle.cache.internal.DefaultMultiProcessSafePersistentIndexedCache.lambda$put$1(DefaultMultiProcessSafePersistentIndexedCache.java:67)
	at org.gradle.cache.internal.DefaultFileLockManager$DefaultFileLock.doWriteAction(DefaultFileLockManager.java:216)
	at org.gradle.cache.internal.DefaultFileLockManager$DefaultFileLock.writeFile(DefaultFileLockManager.java:206)
	at org.gradle.cache.internal.DefaultCacheAccess$UnitOfWorkFileAccess.writeFile(DefaultCacheAccess.java:449)
	at org.gradle.cache.internal.DefaultMultiProcessSafePersistentIndexedCache.put(DefaultMultiProcessSafePersistentIndexedCache.java:67)
	at org.gradle.cache.internal.AsyncCacheAccessDecoratedCache.lambda$putLater$1(AsyncCacheAccessDecoratedCache.java:56)
	at org.gradle.internal.concurrent.ExecutorPolicy$CatchAndRecordFailures.onExecute(ExecutorPolicy.java:64)
	at org.gradle.cache.internal.CacheAccessWorker$2.run(CacheAccessWorker.java:185)
	at org.gradle.internal.Factories$1.create(Factories.java:31)
	at org.gradle.cache.internal.DefaultCacheAccess.useCache(DefaultCacheAccess.java:220)
	at org.gradle.cache.internal.DefaultCacheAccess.useCache(DefaultCacheAccess.java:201)
	at org.gradle.cache.internal.CacheAccessWorker.flushOperations(CacheAccessWorker.java:175)
	at org.gradle.cache.internal.CacheAccessWorker.run(CacheAccessWorker.java:145)
	at org.gradle.internal.concurrent.ExecutorPolicy$CatchAndRecordFailures.onExecute(ExecutorPolicy.java:64)
	at org.gradle.internal.concurrent.ManagedExecutorImpl$1.run(ManagedExecutorImpl.java:48)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
	at org.gradle.internal.concurrent.ThreadFactoryImpl$ManagedThreadRunnable.run(ThreadFactoryImpl.java:61)
	at java.lang.Thread.run(Thread.java:748)
Caused by: org.gradle.api.UncheckedIOException: java.io.IOException: No space left on device
	at org.gradle.cache.internal.btree.FileBackedBlockStore.write(FileBackedBlockStore.java:140)
	at org.gradle.cache.internal.btree.CachingBlockStore.flush(CachingBlockStore.java:64)
	at org.gradle.cache.internal.btree.FreeListBlockStore.flush(FreeListBlockStore.java:103)
	at org.gradle.cache.internal.btree.StateCheckBlockStore.flush(StateCheckBlockStore.java:85)
	at org.gradle.cache.internal.btree.BTreePersistentIndexedCache.put(BTreePersistentIndexedCache.java:160)
	... 19 more
Caused by: java.io.IOException: No space left on device
	at java.io.RandomAccessFile.writeBytes(Native Method)
	at java.io.RandomAccessFile.write(RandomAccessFile.java:525)
	at org.gradle.internal.io.RandomAccessFileOutputStream.write(RandomAccessFileOutputStream.java:46)
	at java.io.BufferedOutputStream.flushBuffer(BufferedOutputStream.java:82)
	at java.io.BufferedOutputStream.flush(BufferedOutputStream.java:140)
	at java.io.FilterOutputStream.flush(FilterOutputStream.java:140)
	at org.gradle.cache.internal.btree.ByteOutput.done(ByteOutput.java:62)
	at org.gradle.cache.internal.btree.FileBackedBlockStore$BlockImpl.write(FileBackedBlockStore.java:223)
	at org.gradle.cache.internal.btree.FileBackedBlockStore.write(FileBackedBlockStore.java:136)
	... 23 more
Failed to execute org.gradle.cache.internal.AsyncCacheAccessDecoratedCache$$Lambda$222/1355592988@7cc1e8fc.
org.gradle.api.UncheckedIOException: Could not add entry '<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/build/target/java_third_party_licenses/zkclient-0.10.jar/LICENSE'> to cache fileHashes.bin (<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/.gradle/7.3.2/fileHashes/fileHashes.bin).>
	at org.gradle.cache.internal.btree.BTreePersistentIndexedCache.put(BTreePersistentIndexedCache.java:162)
	at org.gradle.cache.internal.DefaultMultiProcessSafePersistentIndexedCache.lambda$put$1(DefaultMultiProcessSafePersistentIndexedCache.java:67)
	at org.gradle.cache.internal.DefaultFileLockManager$DefaultFileLock.doWriteAction(DefaultFileLockManager.java:216)
	at org.gradle.cache.internal.DefaultFileLockManager$DefaultFileLock.writeFile(DefaultFileLockManager.java:206)
	at org.gradle.cache.internal.DefaultCacheAccess$UnitOfWorkFileAccess.writeFile(DefaultCacheAccess.java:449)
	at org.gradle.cache.internal.DefaultMultiProcessSafePersistentIndexedCache.put(DefaultMultiProcessSafePersistentIndexedCache.java:67)
	at org.gradle.cache.internal.AsyncCacheAccessDecoratedCache.lambda$putLater$1(AsyncCacheAccessDecoratedCache.java:56)
	at org.gradle.internal.concurrent.ExecutorPolicy$CatchAndRecordFailures.onExecute(ExecutorPolicy.java:64)
	at org.gradle.cache.internal.CacheAccessWorker$2.run(CacheAccessWorker.java:185)
	at org.gradle.internal.Factories$1.create(Factories.java:31)
	at org.gradle.cache.internal.DefaultCacheAccess.useCache(DefaultCacheAccess.java:220)
	at org.gradle.cache.internal.DefaultCacheAccess.useCache(DefaultCacheAccess.java:201)
	at org.gradle.cache.internal.CacheAccessWorker.flushOperations(CacheAccessWorker.java:175)
	at org.gradle.cache.internal.CacheAccessWorker.run(CacheAccessWorker.java:145)
	at org.gradle.internal.concurrent.ExecutorPolicy$CatchAndRecordFailures.onExecute(ExecutorPolicy.java:64)
	at org.gradle.internal.concurrent.ManagedExecutorImpl$1.run(ManagedExecutorImpl.java:48)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
	at org.gradle.internal.concurrent.ThreadFactoryImpl$ManagedThreadRunnable.run(ThreadFactoryImpl.java:61)
	at java.lang.Thread.run(Thread.java:748)
Caused by: org.gradle.api.UncheckedIOException: java.io.IOException: No space left on device
	at org.gradle.cache.internal.btree.FileBackedBlockStore.write(FileBackedBlockStore.java:140)
	at org.gradle.cache.internal.btree.CachingBlockStore.flush(CachingBlockStore.java:64)
	at org.gradle.cache.internal.btree.FreeListBlockStore.flush(FreeListBlockStore.java:103)
	at org.gradle.cache.internal.btree.StateCheckBlockStore.flush(StateCheckBlockStore.java:85)
	at org.gradle.cache.internal.btree.BTreePersistentIndexedCache.put(BTreePersistentIndexedCache.java:160)
	... 19 more
Caused by: java.io.IOException: No space left on device
	at java.io.RandomAccessFile.writeBytes(Native Method)
	at java.io.RandomAccessFile.write(RandomAccessFile.java:525)
	at org.gradle.internal.io.RandomAccessFileOutputStream.write(RandomAccessFileOutputStream.java:46)
	at java.io.BufferedOutputStream.flushBuffer(BufferedOutputStream.java:82)
	at java.io.BufferedOutputStream.flush(BufferedOutputStream.java:140)
	at java.io.FilterOutputStream.flush(FilterOutputStream.java:140)
	at org.gradle.cache.internal.btree.ByteOutput.done(ByteOutput.java:62)
	at org.gradle.cache.internal.btree.FileBackedBlockStore$BlockImpl.write(FileBackedBlockStore.java:223)
	at org.gradle.cache.internal.btree.FileBackedBlockStore.write(FileBackedBlockStore.java:136)
	... 23 more
Failed to execute org.gradle.cache.internal.AsyncCacheAccessDecoratedCache$$Lambda$222/1355592988@704a69c1.
org.gradle.api.UncheckedIOException: Could not add entry '<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/build/target/java_third_party_licenses/google-cloud-core-http-2.3.3.jar/LICENSE'> to cache fileHashes.bin (<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/.gradle/7.3.2/fileHashes/fileHashes.bin).>
	at org.gradle.cache.internal.btree.BTreePersistentIndexedCache.put(BTreePersistentIndexedCache.java:162)
	at org.gradle.cache.internal.DefaultMultiProcessSafePersistentIndexedCache.lambda$put$1(DefaultMultiProcessSafePersistentIndexedCache.java:67)
	at org.gradle.cache.internal.DefaultFileLockManager$DefaultFileLock.doWriteAction(DefaultFileLockManager.java:216)
	at org.gradle.cache.internal.DefaultFileLockManager$DefaultFileLock.writeFile(DefaultFileLockManager.java:206)
	at org.gradle.cache.internal.DefaultCacheAccess$UnitOfWorkFileAccess.writeFile(DefaultCacheAccess.java:449)
	at org.gradle.cache.internal.DefaultMultiProcessSafePersistentIndexedCache.put(DefaultMultiProcessSafePersistentIndexedCache.java:67)
	at org.gradle.cache.internal.AsyncCacheAccessDecoratedCache.lambda$putLater$1(AsyncCacheAccessDecoratedCache.java:56)
	at org.gradle.internal.concurrent.ExecutorPolicy$CatchAndRecordFailures.onExecute(ExecutorPolicy.java:64)
	at org.gradle.cache.internal.CacheAccessWorker$2.run(CacheAccessWorker.java:185)
	at org.gradle.internal.Factories$1.create(Factories.java:31)
	at org.gradle.cache.internal.DefaultCacheAccess.useCache(DefaultCacheAccess.java:220)
	at org.gradle.cache.internal.DefaultCacheAccess.useCache(DefaultCacheAccess.java:201)
	at org.gradle.cache.internal.CacheAccessWorker.flushOperations(CacheAccessWorker.java:175)
	at org.gradle.cache.internal.CacheAccessWorker.run(CacheAccessWorker.java:145)
	at org.gradle.internal.concurrent.ExecutorPolicy$CatchAndRecordFailures.onExecute(ExecutorPolicy.java:64)
	at org.gradle.internal.concurrent.ManagedExecutorImpl$1.run(ManagedExecutorImpl.java:48)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
	at org.gradle.internal.concurrent.ThreadFactoryImpl$ManagedThreadRunnable.run(ThreadFactoryImpl.java:61)
	at java.lang.Thread.run(Thread.java:748)
Caused by: org.gradle.api.UncheckedIOException: java.io.IOException: No space left on device
	at org.gradle.cache.internal.btree.FileBackedBlockStore.write(FileBackedBlockStore.java:140)
	at org.gradle.cache.internal.btree.CachingBlockStore.flush(CachingBlockStore.java:64)
	at org.gradle.cache.internal.btree.FreeListBlockStore.flush(FreeListBlockStore.java:103)
	at org.gradle.cache.internal.btree.StateCheckBlockStore.flush(StateCheckBlockStore.java:85)
	at org.gradle.cache.internal.btree.BTreePersistentIndexedCache.put(BTreePersistentIndexedCache.java:160)
	... 19 more
Caused by: java.io.IOException: No space left on device
	at java.io.RandomAccessFile.writeBytes(Native Method)
	at java.io.RandomAccessFile.write(RandomAccessFile.java:525)
	at org.gradle.internal.io.RandomAccessFileOutputStream.write(RandomAccessFileOutputStream.java:46)
	at java.io.BufferedOutputStream.flushBuffer(BufferedOutputStream.java:82)
	at java.io.BufferedOutputStream.flush(BufferedOutputStream.java:140)
	at java.io.FilterOutputStream.flush(FilterOutputStream.java:140)
	at org.gradle.cache.internal.btree.ByteOutput.done(ByteOutput.java:62)
	at org.gradle.cache.internal.btree.FileBackedBlockStore$BlockImpl.write(FileBackedBlockStore.java:223)
	at org.gradle.cache.internal.btree.FileBackedBlockStore.write(FileBackedBlockStore.java:136)
	... 23 more
Failed to execute org.gradle.cache.internal.AsyncCacheAccessDecoratedCache$$Lambda$222/1355592988@3fa3a21c.
org.gradle.api.UncheckedIOException: Could not add entry '<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/sdks/java/container/build/target/java_third_party_licenses/google-oauth-client-1.32.1.jar/LICENSE'> to cache fileHashes.bin (<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/.gradle/7.3.2/fileHashes/fileHashes.bin).>
	at org.gradle.cache.internal.btree.BTreePersistentIndexedCache.put(BTreePersistentIndexedCache.java:162)
	at org.gradle.cache.internal.DefaultMultiProcessSafePersistentIndexedCache.lambda$put$1(DefaultMultiProcessSafePersistentIndexedCache.java:67)
	at org.gradle.cache.internal.DefaultFileLockManager$DefaultFileLock.doWriteAction(DefaultFileLockManager.java:216)
	at org.gradle.cache.internal.DefaultFileLockManager$DefaultFileLock.writeFile(DefaultFileLockManager.java:206)
	at org.gradle.cache.internal.DefaultCacheAccess$UnitOfWorkFileAccess.writeFile(DefaultCacheAccess.java:449)
	at org.gradle.cache.internal.DefaultMultiProcessSafePersistentIndexedCache.put(DefaultMultiProcessSafePersistentIndexedCache.java:67)
	at org.gradle.cache.internal.AsyncCacheAccessDecoratedCache.lambda$putLater$1(AsyncCacheAccessDecoratedCache.java:56)
	at org.gradle.internal.concurrent.ExecutorPolicy$CatchAndRecordFailures.onExecute(ExecutorPolicy.java:64)
	at org.gradle.cache.internal.CacheAccessWorker$2.run(CacheAccessWorker.java:185)
	at org.gradle.internal.Factories$1.create(Factories.java:31)
	at org.gradle.cache.internal.DefaultCacheAccess.useCache(DefaultCacheAccess.java:220)
	at org.gradle.cache.internal.DefaultCacheAccess.useCache(DefaultCacheAccess.java:201)
	at org.gradle.cache.internal.CacheAccessWorker.flushOperations(CacheAccessWorker.java:175)
	at org.gradle.cache.internal.CacheAccessWorker.run(CacheAccessWorker.java:145)
	at org.gradle.internal.concurrent.ExecutorPolicy$CatchAndRecordFailures.onExecute(ExecutorPolicy.java:64)
	at org.gradle.internal.concurrent.ManagedExecutorImpl$1.run(ManagedExecutorImpl.java:48)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
	at org.gradle.internal.concurrent.ThreadFactoryImpl$ManagedThreadRunnable.run(ThreadFactoryImpl.java:61)
	at java.lang.Thread.run(Thread.java:748)
Caused by: org.gradle.api.UncheckedIOException: java.io.IOException: No space left on device
	at org.gradle.cache.internal.btree.FileBackedBlockStore.write(FileBackedBlockStore.java:140)
	at org.gradle.cache.internal.btree.CachingBlockStore.flush(CachingBlockStore.java:64)
	at org.gradle.cache.internal.btree.FreeListBlockStore.flush(FreeListBlockStore.java:103)
	at org.gradle.cache.internal.btree.StateCheckBlockStore.flush(StateCheckBlockStore.java:85)
	at org.gradle.cache.internal.btree.BTreePersistentIndexedCache.put(BTreePersistentIndexedCache.java:160)
	... 19 more
Caused by: java.io.IOException: No space left on device
	at java.io.RandomAccessFile.writeBytes(Native Method)
	at java.io.RandomAccessFile.write(RandomAccessFile.java:525)
	at org.gradle.internal.io.RandomAccessFileOutputStream.write(RandomAccessFileOutputStream.java:46)
	at java.io.BufferedOutputStream.flushBuffer(BufferedOutputStream.java:82)
	at java.io.BufferedOutputStream.flush(BufferedOutputStream.java:140)
	at java.io.FilterOutputStream.flush(FilterOutputStream.java:140)
	at org.gradle.cache.internal.btree.ByteOutput.done(ByteOutput.java:62)
	at org.gradle.cache.internal.btree.FileBackedBlockStore$BlockImpl.write(FileBackedBlockStore.java:223)
	at org.gradle.cache.internal.btree.FileBackedBlockStore.write(FileBackedBlockStore.java:136)
	... 23 more

> Task :sdks:java:container:pullLicenses FAILED

FAILURE: Build completed with 2 failures.

1: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':release:go-licenses:java:dockerRun'.
> Process 'command 'docker'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.
==============================================================================

2: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':sdks:java:container:pullLicenses'.
> Process 'command './license_scripts/license_script.sh'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.
==============================================================================

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 45s
102 actionable tasks: 67 executed, 31 from cache, 4 up-to-date

A build scan cannot be produced as an error occurred spooling the build data.
Please report this problem via https://gradle.com/help/plugin and include the following via copy/paste:

----------
Gradle version: 7.3.2
Plugin version: 3.4.1

java.lang.IllegalStateException: Could not close the event spooler due to previous errors.
	at com.gradle.scan.plugin.internal.f.c.c.a(SourceFile:128)
	at com.gradle.scan.plugin.internal.q.a$a.a(SourceFile:31)
	at com.gradle.scan.plugin.internal.q.a$a.a(SourceFile:20)
	at com.gradle.scan.plugin.internal.q.a.c(SourceFile:67)
----------


FAILURE: Build failed with an exception.

* What went wrong:
Could not add entry ':sdks:java:container:pullLicenses' to cache executionHistory.bin (<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/.gradle/7.3.2/executionHistory/executionHistory.bin).>
> java.io.IOException: No space left on device

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 45s
Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #28

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/28/display/redirect?page=changes>

Changes:

[mmack] [BEAM-13638] Datatype of timestamp fields in SqsMessage for AWS IOs for

[noreply] Merge pull request #16469 from [BEAM-13623][Playground] [Bugfix] During

[noreply] Merge pull request #16149 from [BEAM-13113] [Playground] playground

[noreply] Merge pull request #16363 from [BEAM-13557] [Playground] show code

[noreply] Merge pull request #16374 from [BEAM-13398][Playground] Split LifeCycle

[noreply] [BEAM-13616][BEAM-13646] Update vendored calcite 1.28.0 with protobuf

[chamikaramj] Adds several example multi-language Python pipelines

[noreply] Merge pull request #16325 from [BEAM-13471] [Playground] Tag existing

[noreply] [BEAM-13399] Move service liveness polling to Runner type (#16487)


------------------------------------------
[...truncated 49.05 KB...]
bff9fe6e429c: Waiting
65b0ffa223e1: Waiting
9e98e8586004: Waiting
ab7be8c61871: Waiting
a04b069a3119: Waiting
8bf42db0de72: Waiting
26a504e63be4: Waiting
f9e18e59a565: Waiting
387fb76dedf0: Pushed
a1a050b760ec: Pushed
349a89be8394: Pushed
120b183bde69: Pushed
9e98e8586004: Pushed
aec7efa0bb7d: Pushed
8dfc6125a02f: Pushed
65b0ffa223e1: Pushed
d3b21b2f6c8f: Pushed
a04b069a3119: Pushed
e2030dae3737: Pushed
bff9fe6e429c: Layer already exists
7c245b2fe4f1: Layer already exists
f9e18e59a565: Layer already exists
26a504e63be4: Layer already exists
31892cc314cb: Layer already exists
8bf42db0de72: Layer already exists
11936051f93b: Layer already exists
206655f2b859: Pushed
ab7be8c61871: Pushed
20220113123036: digest: sha256:69464a7b079c1827fd8fe954ee7166397747e3993d2f8159ea98ae31752957f6 size: 4520

> Task :sdks:java:testing:load-tests:run
Jan 13, 2022 12:32:34 PM org.apache.beam.runners.dataflow.options.DataflowPipelineOptions$StagingLocationFactory create
INFO: No stagingLocation provided, falling back to gcpTempLocation
Jan 13, 2022 12:32:34 PM org.apache.beam.runners.dataflow.DataflowRunner fromOptions
INFO: PipelineOptions.filesToStage was not specified. Defaulting to files from the classpath: will stage 203 files. Enable logging at DEBUG level to see which files will be staged.
Jan 13, 2022 12:32:35 PM org.apache.beam.sdk.Pipeline validate
WARNING: The following transforms do not have stable unique names: Window.Into()
Jan 13, 2022 12:32:35 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Executing pipeline on the Dataflow Service, which will have billing implications related to Google Compute Engine usage and other Google Cloud Services.
Jan 13, 2022 12:32:37 PM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Uploading 203 files from PipelineOptions.filesToStage to staging location to prepare for execution.
Jan 13, 2022 12:32:38 PM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Staging files complete: 203 files cached, 0 files newly uploaded in 0 seconds
Jan 13, 2022 12:32:38 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Staging portable pipeline proto to gs://temp-storage-for-perf-tests/loadtests/staging/
Jan 13, 2022 12:32:38 PM org.apache.beam.runners.dataflow.util.PackageUtil tryStagePackage
INFO: Uploading <113691 bytes, hash 7270882687f8993aa505d9009cfe5011d59e42959d8be5ae6b4e531d7bfe1640> to gs://temp-storage-for-perf-tests/loadtests/staging/pipeline-cnCIJof4mTqlBdkAnP5QEdWeQpWdi-Wua05THXv-FkA.pb
Jan 13, 2022 12:32:39 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read input/DataflowRunner.StreamingUnboundedRead.ReadWithIds as step s1
Jan 13, 2022 12:32:39 PM org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource split
INFO: Split into 20 bundles of sizes: [org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4aa21f9d, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@71c17a57, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@640ab13c, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1e0a864d, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@440e3ce6, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3e67f5f2, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@fd53053, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4527f70a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@707b1a44, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7132a9dc, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@57435801, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2da66a44, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@527fc8e, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@61bfc9bf, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2c7106d9, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@329bad59, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@862f408, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@178f268a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@33ecbd6c, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5c723f2d]
Jan 13, 2022 12:32:40 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read input/StripIds as step s2
Jan 13, 2022 12:32:40 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect start time metrics (input) as step s3
Jan 13, 2022 12:32:40 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Window.Into()/Window.Assign as step s4
Jan 13, 2022 12:32:40 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read co-input/DataflowRunner.StreamingUnboundedRead.ReadWithIds as step s5
Jan 13, 2022 12:32:40 PM org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource split
INFO: Split into 20 bundles of sizes: [org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@24eb65e3, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3ac3f6f, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1abebef3, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@18f55704, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@67cefd84, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5fbe155, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6add8e3f, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@58a2b917, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@48904d5a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@12bbfc54, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1491344a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@59b65dce, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1386313f, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5e922647, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@433c6abb, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@288f173f, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@a22c4d8, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@45cd7bc5, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6504a875, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@35e26d05]
Jan 13, 2022 12:32:40 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read co-input/StripIds as step s6
Jan 13, 2022 12:32:40 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect start time metrics (co-input) as step s7
Jan 13, 2022 12:32:40 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Window.Into()2/Window.Assign as step s8
Jan 13, 2022 12:32:40 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/MakeUnionTable0 as step s9
Jan 13, 2022 12:32:40 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/MakeUnionTable1 as step s10
Jan 13, 2022 12:32:40 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/Flatten as step s11
Jan 13, 2022 12:32:40 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/GBK as step s12
Jan 13, 2022 12:32:40 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/ConstructCoGbkResultFn as step s13
Jan 13, 2022 12:32:40 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Ungroup and reiterate as step s14
Jan 13, 2022 12:32:40 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect total bytes as step s15
Jan 13, 2022 12:32:40 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect end time metrics as step s16
Jan 13, 2022 12:32:40 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Dataflow SDK version: 2.37.0-SNAPSHOT
Jan 13, 2022 12:32:40 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To access the Dataflow monitoring console, please navigate to https://console.cloud.google.com/dataflow/jobs/us-central1/2022-01-13_04_32_40-1316229289575014301?project=apache-beam-testing
Jan 13, 2022 12:32:40 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Submitted job: 2022-01-13_04_32_40-1316229289575014301
Jan 13, 2022 12:32:40 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To cancel the job using the 'gcloud' tool, run:
> gcloud dataflow jobs --project=apache-beam-testing cancel --region=us-central1 2022-01-13_04_32_40-1316229289575014301
Jan 13, 2022 12:32:50 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
WARNING: 2022-01-13T12:32:49.326Z: The workflow name is not a valid Cloud Label. Labels applied to Cloud resources (such as GCE Instances) for monitoring will be labeled with this modified job name: load0tests0java170dataflow0v20streaming0cogbk01-jenkins-01-vgr8. For the best monitoring experience, please name your job with a valid Cloud Label. For details, see: https://cloud.google.com/compute/docs/labeling-resources#restrictions
Jan 13, 2022 12:32:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-13T12:32:53.719Z: Worker configuration: e2-standard-2 in us-central1-b.
Jan 13, 2022 12:32:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-13T12:32:54.459Z: Expanding SplittableParDo operations into optimizable parts.
Jan 13, 2022 12:32:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-13T12:32:54.492Z: Expanding CollectionToSingleton operations into optimizable parts.
Jan 13, 2022 12:32:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-13T12:32:54.569Z: Expanding CoGroupByKey operations into optimizable parts.
Jan 13, 2022 12:32:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-13T12:32:54.636Z: Expanding SplittableProcessKeyed operations into optimizable parts.
Jan 13, 2022 12:32:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-13T12:32:54.668Z: Expanding GroupByKey operations into streaming Read/Write steps
Jan 13, 2022 12:32:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-13T12:32:54.734Z: Lifting ValueCombiningMappingFns into MergeBucketsMappingFns
Jan 13, 2022 12:32:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-13T12:32:54.836Z: Fusing adjacent ParDo, Read, Write, and Flatten operations
Jan 13, 2022 12:32:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-13T12:32:54.860Z: Unzipping flatten CoGroupByKey-Flatten for input CoGroupByKey-MakeUnionTable0-ParMultiDo-ConstructUnionTable-.output
Jan 13, 2022 12:32:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-13T12:32:54.894Z: Fusing unzipped copy of CoGroupByKey/GBK/WriteStream, through flatten CoGroupByKey/Flatten, into producer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable)
Jan 13, 2022 12:32:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-13T12:32:54.917Z: Fusing consumer CoGroupByKey/GBK/WriteStream into CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable)
Jan 13, 2022 12:32:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-13T12:32:54.968Z: Fusing consumer Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read input/Impulse
Jan 13, 2022 12:32:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-13T12:32:55.028Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
Jan 13, 2022 12:32:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-13T12:32:55.061Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
Jan 13, 2022 12:32:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-13T12:32:55.094Z: Fusing consumer Read input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
Jan 13, 2022 12:32:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-13T12:32:55.128Z: Fusing consumer Collect start time metrics (input)/ParMultiDo(TimeMonitor) into Read input/ParDo(StripIds)/ParMultiDo(StripIds)
Jan 13, 2022 12:32:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-13T12:32:55.161Z: Fusing consumer Window.Into()/Window.Assign into Collect start time metrics (input)/ParMultiDo(TimeMonitor)
Jan 13, 2022 12:32:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-13T12:32:55.195Z: Fusing consumer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable) into Window.Into()/Window.Assign
Jan 13, 2022 12:32:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-13T12:32:55.226Z: Fusing consumer Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read co-input/Impulse
Jan 13, 2022 12:32:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-13T12:32:55.259Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
Jan 13, 2022 12:32:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-13T12:32:55.332Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
Jan 13, 2022 12:32:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-13T12:32:55.358Z: Fusing consumer Read co-input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
Jan 13, 2022 12:32:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-13T12:32:55.394Z: Fusing consumer Collect start time metrics (co-input)/ParMultiDo(TimeMonitor) into Read co-input/ParDo(StripIds)/ParMultiDo(StripIds)
Jan 13, 2022 12:32:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-13T12:32:55.433Z: Fusing consumer Window.Into()2/Window.Assign into Collect start time metrics (co-input)/ParMultiDo(TimeMonitor)
Jan 13, 2022 12:32:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-13T12:32:55.466Z: Fusing consumer CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable) into Window.Into()2/Window.Assign
Jan 13, 2022 12:32:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-13T12:32:55.496Z: Fusing consumer CoGroupByKey/GBK/MergeBuckets into CoGroupByKey/GBK/ReadStream
Jan 13, 2022 12:32:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-13T12:32:55.531Z: Fusing consumer CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult) into CoGroupByKey/GBK/MergeBuckets
Jan 13, 2022 12:32:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-13T12:32:55.565Z: Fusing consumer Ungroup and reiterate/ParMultiDo(UngroupAndReiterate) into CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult)
Jan 13, 2022 12:32:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-13T12:32:55.589Z: Fusing consumer Collect total bytes/ParMultiDo(ByteMonitor) into Ungroup and reiterate/ParMultiDo(UngroupAndReiterate)
Jan 13, 2022 12:32:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-13T12:32:55.655Z: Fusing consumer Collect end time metrics/ParMultiDo(TimeMonitor) into Collect total bytes/ParMultiDo(ByteMonitor)
Jan 13, 2022 12:32:56 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-13T12:32:56.069Z: Starting 5 ****s in us-central1-b...
Jan 13, 2022 12:33:22 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-13T12:33:22.473Z: Your project already contains 100 Dataflow-created metric descriptors, so new user metrics of the form custom.googleapis.com/* will not be created. However, all user metrics are also available in the metric dataflow.googleapis.com/job/user_counter. If you rely on the custom metrics, you can delete old / unused metric descriptors. See https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.list and https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.delete
Jan 13, 2022 12:33:48 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-13T12:33:48.084Z: Autoscaling: Raised the number of ****s to 5 so that the pipeline can catch up with its backlog and keep up with its input rate.
Jan 13, 2022 12:34:45 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-13T12:34:44.234Z: Workers have started successfully.
Jan 13, 2022 12:34:45 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-13T12:34:44.262Z: Workers have started successfully.
Jan 13, 2022 4:00:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-13T16:00:38.800Z: Cancel request is committed for workflow job: 2022-01-13_04_32_40-1316229289575014301.
Jan 13, 2022 4:00:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-13T16:00:38.888Z: Cleaning up.
Jan 13, 2022 4:00:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-13T16:00:38.957Z: Stopping **** pool...
Jan 13, 2022 4:00:41 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-13T16:00:39.011Z: Stopping **** pool...
Jan 13, 2022 4:03:00 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-13T16:02:59.404Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Jan 13, 2022 4:03:00 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-13T16:02:59.439Z: Worker pool stopped.
Jan 13, 2022 4:03:06 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-01-13_04_32_40-1316229289575014301 finished with status CANCELLED.
Load test results for test (ID): b6e89fdb-8519-4005-9eb7-6bb9d002f6f9 and timestamp: 2022-01-13T12:32:35.008000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                 12232.132
dataflow_v2_java17_total_bytes_count             2.06816431E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220113123036
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:69464a7b079c1827fd8fe954ee7166397747e3993d2f8159ea98ae31752957f6
Deleted: sha256:a6efe6a5cc82f00458d4ba6386c4efe422e7486688ecb7fc1dd193f46b01ba5d
Deleted: sha256:47b7e59111d1df91ef5fba0ea0615d64848efd0c4efe9996a71155babb47b017
Deleted: sha256:ac288338321333c4f01ebba3984b1aa1055463f331e117b21f511622a84ed36b
Deleted: sha256:d8c14b2d3780907e2756fa8556e9bbad5333a86c931c382a81cb99caf3aa27e9
Deleted: sha256:ced9cd3e7c60b966cb94f48316fe0a9aa10385a89a3c9261badde5832e6fb732
Deleted: sha256:cee362cdf1cf452eff03dcbd954a710393899ceca955c0d515ace70a931c9f97
Deleted: sha256:a6b7d49dce0d0976e663fc2784c43a1ad4aba3b0d8f0232c8a58634c11d208d7
Deleted: sha256:fc51448fe9cbcb0fe398a4f4ac0a5547290485daac1cf635a42c0abd5f18b80c
Deleted: sha256:145bba74afa4a5faae8d02f39a84e44697febbd1f1e0f56a090a406dd0fc01e3
Deleted: sha256:4d9a92ae399baa908668bf85e7ada51808ef29d00241a7d69b0ee68d56a74075
Deleted: sha256:d257d6da93f68a229bcd498b0c185c82fe1b6bfc7d6d2bfd1bdea050aa0f2945
Deleted: sha256:70f403cd5ea73635916f51a0aa8296f8e9c837125501f0d50a8e19ca0ee33816
Deleted: sha256:c7523a5469d3eb4891782b677d237add59b71e21c0c504a671b4038ed45ce243
Deleted: sha256:b442a3182b10204753dcc81ae6c9be9ae27b7136f912a7aebcb373298f616a6d
Deleted: sha256:230cabcc1da6a15008c8b475b0402b49807752a4941e4f6d6bd8ddc5ff2af394
Deleted: sha256:2b2e906ed6947485db4603d1121fed40f77bf495509d2e0583208badbcfd5a73
Deleted: sha256:e14c6d3ca291ccb2d0c4b86e8814d6d890ab692d4c5c5e515cf206415e424f64
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220113123036]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:69464a7b079c1827fd8fe954ee7166397747e3993d2f8159ea98ae31752957f6]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220113123036] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:69464a7b079c1827fd8fe954ee7166397747e3993d2f8159ea98ae31752957f6])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:69464a7b079c1827fd8fe954ee7166397747e3993d2f8159ea98ae31752957f6
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:69464a7b079c1827fd8fe954ee7166397747e3993d2f8159ea98ae31752957f6
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:69464a7b079c1827fd8fe954ee7166397747e3993d2f8159ea98ae31752957f6].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 32m 47s
109 actionable tasks: 72 executed, 33 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/wzuxeb3r6acp6

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #27

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/27/display/redirect?page=changes>

Changes:

[ningkang0957] [BEAM-13602] Prevented metrics gathering from failing bigtable io

[ningkang0957] make the code more pythonic

[mmack] [BEAM-13243][BEAM-8374] Add support for missing PublishResponse fields

[Robert Bradshaw] Optional args and kwargs for named external transforms.

[mmack] [adhoc] Run spotlessApply on java examples to fix master

[noreply] [BEAM-13628] Update SideInputCache to use full Transform and

[noreply] [BEAM-13432] Skip ExpansionService creation in Job Server (#16222)

[noreply] [BEAM-13616] Initial files for vendored gRPC 1.43.2 (#16460)


------------------------------------------
[...truncated 50.50 KB...]
e2e623018604: Pushed
f9e18e59a565: Layer already exists
26a504e63be4: Layer already exists
8bf42db0de72: Layer already exists
31892cc314cb: Layer already exists
11936051f93b: Layer already exists
8f8becdaaef7: Pushed
df1de1123875: Pushed
20220112123035: digest: sha256:eae73b74c7cee1be2d90dec0adb44e27fc9798b467f822dccfb3e0407214c58c size: 4520

> Task :sdks:java:testing:load-tests:run
Jan 12, 2022 12:32:46 PM org.apache.beam.runners.dataflow.options.DataflowPipelineOptions$StagingLocationFactory create
INFO: No stagingLocation provided, falling back to gcpTempLocation
Jan 12, 2022 12:32:47 PM org.apache.beam.runners.dataflow.DataflowRunner fromOptions
INFO: PipelineOptions.filesToStage was not specified. Defaulting to files from the classpath: will stage 203 files. Enable logging at DEBUG level to see which files will be staged.
Jan 12, 2022 12:32:48 PM org.apache.beam.sdk.Pipeline validate
WARNING: The following transforms do not have stable unique names: Window.Into()
Jan 12, 2022 12:32:48 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Executing pipeline on the Dataflow Service, which will have billing implications related to Google Compute Engine usage and other Google Cloud Services.
Jan 12, 2022 12:32:51 PM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Uploading 203 files from PipelineOptions.filesToStage to staging location to prepare for execution.
Jan 12, 2022 12:32:51 PM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Staging files complete: 203 files cached, 0 files newly uploaded in 0 seconds
Jan 12, 2022 12:32:51 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Staging portable pipeline proto to gs://temp-storage-for-perf-tests/loadtests/staging/
Jan 12, 2022 12:32:51 PM org.apache.beam.runners.dataflow.util.PackageUtil tryStagePackage
INFO: Uploading <113691 bytes, hash 267cc14bab79894a3f26f01848e8d4032cda162347e6e27ac07fff95978bb769> to gs://temp-storage-for-perf-tests/loadtests/staging/pipeline-JnzBS6t5iUo_JvAYSOjUAyzaFiNH5uJ6wH__lZeLt2k.pb
Jan 12, 2022 12:32:53 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read input/DataflowRunner.StreamingUnboundedRead.ReadWithIds as step s1
Jan 12, 2022 12:32:53 PM org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource split
INFO: Split into 20 bundles of sizes: [org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4aa21f9d, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@71c17a57, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@640ab13c, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1e0a864d, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@440e3ce6, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3e67f5f2, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@fd53053, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4527f70a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@707b1a44, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7132a9dc, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@57435801, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2da66a44, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@527fc8e, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@61bfc9bf, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2c7106d9, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@329bad59, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@862f408, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@178f268a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@33ecbd6c, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5c723f2d]
Jan 12, 2022 12:32:53 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read input/StripIds as step s2
Jan 12, 2022 12:32:53 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect start time metrics (input) as step s3
Jan 12, 2022 12:32:53 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Window.Into()/Window.Assign as step s4
Jan 12, 2022 12:32:53 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read co-input/DataflowRunner.StreamingUnboundedRead.ReadWithIds as step s5
Jan 12, 2022 12:32:53 PM org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource split
INFO: Split into 20 bundles of sizes: [org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@24eb65e3, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3ac3f6f, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1abebef3, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@18f55704, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@67cefd84, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5fbe155, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6add8e3f, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@58a2b917, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@48904d5a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@12bbfc54, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1491344a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@59b65dce, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1386313f, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5e922647, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@433c6abb, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@288f173f, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@a22c4d8, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@45cd7bc5, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6504a875, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@35e26d05]
Jan 12, 2022 12:32:53 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read co-input/StripIds as step s6
Jan 12, 2022 12:32:53 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect start time metrics (co-input) as step s7
Jan 12, 2022 12:32:53 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Window.Into()2/Window.Assign as step s8
Jan 12, 2022 12:32:53 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/MakeUnionTable0 as step s9
Jan 12, 2022 12:32:53 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/MakeUnionTable1 as step s10
Jan 12, 2022 12:32:53 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/Flatten as step s11
Jan 12, 2022 12:32:53 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/GBK as step s12
Jan 12, 2022 12:32:53 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/ConstructCoGbkResultFn as step s13
Jan 12, 2022 12:32:53 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Ungroup and reiterate as step s14
Jan 12, 2022 12:32:53 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect total bytes as step s15
Jan 12, 2022 12:32:53 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect end time metrics as step s16
Jan 12, 2022 12:32:53 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Dataflow SDK version: 2.37.0-SNAPSHOT
Jan 12, 2022 12:32:54 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To access the Dataflow monitoring console, please navigate to https://console.cloud.google.com/dataflow/jobs/us-central1/2022-01-12_04_32_54-4753480158470970014?project=apache-beam-testing
Jan 12, 2022 12:32:54 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Submitted job: 2022-01-12_04_32_54-4753480158470970014
Jan 12, 2022 12:32:54 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To cancel the job using the 'gcloud' tool, run:
> gcloud dataflow jobs --project=apache-beam-testing cancel --region=us-central1 2022-01-12_04_32_54-4753480158470970014
Jan 12, 2022 12:33:03 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
WARNING: 2022-01-12T12:33:01.222Z: The workflow name is not a valid Cloud Label. Labels applied to Cloud resources (such as GCE Instances) for monitoring will be labeled with this modified job name: load0tests0java170dataflow0v20streaming0cogbk01-jenkins-01-plt9. For the best monitoring experience, please name your job with a valid Cloud Label. For details, see: https://cloud.google.com/compute/docs/labeling-resources#restrictions
Jan 12, 2022 12:33:07 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-12T12:33:06.345Z: Worker configuration: e2-standard-2 in us-central1-f.
Jan 12, 2022 12:33:07 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-12T12:33:06.957Z: Expanding SplittableParDo operations into optimizable parts.
Jan 12, 2022 12:33:07 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-12T12:33:06.999Z: Expanding CollectionToSingleton operations into optimizable parts.
Jan 12, 2022 12:33:07 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-12T12:33:07.083Z: Expanding CoGroupByKey operations into optimizable parts.
Jan 12, 2022 12:33:07 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-12T12:33:07.171Z: Expanding SplittableProcessKeyed operations into optimizable parts.
Jan 12, 2022 12:33:07 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-12T12:33:07.213Z: Expanding GroupByKey operations into streaming Read/Write steps
Jan 12, 2022 12:33:07 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-12T12:33:07.284Z: Lifting ValueCombiningMappingFns into MergeBucketsMappingFns
Jan 12, 2022 12:33:07 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-12T12:33:07.382Z: Fusing adjacent ParDo, Read, Write, and Flatten operations
Jan 12, 2022 12:33:07 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-12T12:33:07.428Z: Unzipping flatten CoGroupByKey-Flatten for input CoGroupByKey-MakeUnionTable0-ParMultiDo-ConstructUnionTable-.output
Jan 12, 2022 12:33:07 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-12T12:33:07.454Z: Fusing unzipped copy of CoGroupByKey/GBK/WriteStream, through flatten CoGroupByKey/Flatten, into producer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable)
Jan 12, 2022 12:33:07 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-12T12:33:07.478Z: Fusing consumer CoGroupByKey/GBK/WriteStream into CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable)
Jan 12, 2022 12:33:07 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-12T12:33:07.514Z: Fusing consumer Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read input/Impulse
Jan 12, 2022 12:33:07 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-12T12:33:07.548Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
Jan 12, 2022 12:33:07 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-12T12:33:07.584Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
Jan 12, 2022 12:33:07 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-12T12:33:07.624Z: Fusing consumer Read input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
Jan 12, 2022 12:33:07 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-12T12:33:07.673Z: Fusing consumer Collect start time metrics (input)/ParMultiDo(TimeMonitor) into Read input/ParDo(StripIds)/ParMultiDo(StripIds)
Jan 12, 2022 12:33:07 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-12T12:33:07.702Z: Fusing consumer Window.Into()/Window.Assign into Collect start time metrics (input)/ParMultiDo(TimeMonitor)
Jan 12, 2022 12:33:07 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-12T12:33:07.732Z: Fusing consumer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable) into Window.Into()/Window.Assign
Jan 12, 2022 12:33:07 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-12T12:33:07.758Z: Fusing consumer Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read co-input/Impulse
Jan 12, 2022 12:33:09 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-12T12:33:07.784Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
Jan 12, 2022 12:33:09 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-12T12:33:07.812Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
Jan 12, 2022 12:33:09 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-12T12:33:07.834Z: Fusing consumer Read co-input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
Jan 12, 2022 12:33:09 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-12T12:33:07.856Z: Fusing consumer Collect start time metrics (co-input)/ParMultiDo(TimeMonitor) into Read co-input/ParDo(StripIds)/ParMultiDo(StripIds)
Jan 12, 2022 12:33:09 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-12T12:33:07.885Z: Fusing consumer Window.Into()2/Window.Assign into Collect start time metrics (co-input)/ParMultiDo(TimeMonitor)
Jan 12, 2022 12:33:09 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-12T12:33:07.909Z: Fusing consumer CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable) into Window.Into()2/Window.Assign
Jan 12, 2022 12:33:09 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-12T12:33:07.937Z: Fusing consumer CoGroupByKey/GBK/MergeBuckets into CoGroupByKey/GBK/ReadStream
Jan 12, 2022 12:33:09 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-12T12:33:07.975Z: Fusing consumer CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult) into CoGroupByKey/GBK/MergeBuckets
Jan 12, 2022 12:33:09 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-12T12:33:08.002Z: Fusing consumer Ungroup and reiterate/ParMultiDo(UngroupAndReiterate) into CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult)
Jan 12, 2022 12:33:09 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-12T12:33:08.036Z: Fusing consumer Collect total bytes/ParMultiDo(ByteMonitor) into Ungroup and reiterate/ParMultiDo(UngroupAndReiterate)
Jan 12, 2022 12:33:09 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-12T12:33:08.071Z: Fusing consumer Collect end time metrics/ParMultiDo(TimeMonitor) into Collect total bytes/ParMultiDo(ByteMonitor)
Jan 12, 2022 12:33:09 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-12T12:33:08.466Z: Starting 5 ****s in us-central1-f...
Jan 12, 2022 12:33:32 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-12T12:33:32.201Z: Your project already contains 100 Dataflow-created metric descriptors, so new user metrics of the form custom.googleapis.com/* will not be created. However, all user metrics are also available in the metric dataflow.googleapis.com/job/user_counter. If you rely on the custom metrics, you can delete old / unused metric descriptors. See https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.list and https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.delete
Jan 12, 2022 12:33:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-12T12:33:54.238Z: Autoscaling: Raised the number of ****s to 5 so that the pipeline can catch up with its backlog and keep up with its input rate.
Jan 12, 2022 12:34:55 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-12T12:34:52.809Z: Workers have started successfully.
Jan 12, 2022 12:34:55 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-12T12:34:52.843Z: Workers have started successfully.
Jan 12, 2022 1:54:11 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
SEVERE: 2022-01-12T13:54:09.793Z: Staged package gcsio-2.2.4-p9i4-xCmDBUow6YptIUTZYGtZZhNat4tocR78u1Q8kQ.jar at location 'gs://temp-storage-for-perf-tests/loadtests/staging/gcsio-2.2.4-p9i4-xCmDBUow6YptIUTZYGtZZhNat4tocR78u1Q8kQ.jar' is inaccessible.
Jan 12, 2022 1:54:11 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
SEVERE: 2022-01-12T13:54:10.143Z: Staged package google-api-services-iamcredentials-v1-rev20210326-1.32.1-rzLlTs22sMZOFz3yEOfAB70ZZw2Gl99hr42n_k-f8-M.jar at location 'gs://temp-storage-for-perf-tests/loadtests/staging/google-api-services-iamcredentials-v1-rev20210326-1.32.1-rzLlTs22sMZOFz3yEOfAB70ZZw2Gl99hr42n_k-f8-M.jar' is inaccessible.
Jan 12, 2022 1:54:14 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
SEVERE: 2022-01-12T13:54:12.186Z: Staged package util-2.2.4-r-7-okFosTGJSr5au-fNqmC_EgCN_DyG3X-dgI2RZFA.jar at location 'gs://temp-storage-for-perf-tests/loadtests/staging/util-2.2.4-r-7-okFosTGJSr5au-fNqmC_EgCN_DyG3X-dgI2RZFA.jar' is inaccessible.
Jan 12, 2022 1:54:14 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
WARNING: 2022-01-12T13:54:12.249Z: One or more access checks for temp location or staged files failed. Please refer to other error messages for details. For more information on security and permissions, please see https://cloud.google.com/dataflow/security-and-permissions.
Jan 12, 2022 1:57:12 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
WARNING: 2022-01-12T13:57:11.860Z: One or more access checks for temp location or staged files failed. Please refer to other error messages for details. For more information on security and permissions, please see https://cloud.google.com/dataflow/security-and-permissions.
Jan 12, 2022 2:00:11 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
SEVERE: 2022-01-12T14:00:09.988Z: Staged package gcsio-2.2.4-p9i4-xCmDBUow6YptIUTZYGtZZhNat4tocR78u1Q8kQ.jar at location 'gs://temp-storage-for-perf-tests/loadtests/staging/gcsio-2.2.4-p9i4-xCmDBUow6YptIUTZYGtZZhNat4tocR78u1Q8kQ.jar' is inaccessible.
Jan 12, 2022 2:00:11 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
SEVERE: 2022-01-12T14:00:10.129Z: Staged package google-api-services-iamcredentials-v1-rev20210326-1.32.1-rzLlTs22sMZOFz3yEOfAB70ZZw2Gl99hr42n_k-f8-M.jar at location 'gs://temp-storage-for-perf-tests/loadtests/staging/google-api-services-iamcredentials-v1-rev20210326-1.32.1-rzLlTs22sMZOFz3yEOfAB70ZZw2Gl99hr42n_k-f8-M.jar' is inaccessible.
Jan 12, 2022 2:00:12 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
SEVERE: 2022-01-12T14:00:12.521Z: Staged package util-2.2.4-r-7-okFosTGJSr5au-fNqmC_EgCN_DyG3X-dgI2RZFA.jar at location 'gs://temp-storage-for-perf-tests/loadtests/staging/util-2.2.4-r-7-okFosTGJSr5au-fNqmC_EgCN_DyG3X-dgI2RZFA.jar' is inaccessible.
Jan 12, 2022 2:00:13 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
WARNING: 2022-01-12T14:00:12.599Z: One or more access checks for temp location or staged files failed. Please refer to other error messages for details. For more information on security and permissions, please see https://cloud.google.com/dataflow/security-and-permissions.
Jan 12, 2022 2:03:14 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
WARNING: 2022-01-12T14:03:12.102Z: One or more access checks for temp location or staged files failed. Please refer to other error messages for details. For more information on security and permissions, please see https://cloud.google.com/dataflow/security-and-permissions.
Jan 12, 2022 4:00:45 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-12T16:00:44.502Z: Cancel request is committed for workflow job: 2022-01-12_04_32_54-4753480158470970014.
Jan 12, 2022 4:00:45 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-12T16:00:44.537Z: Cleaning up.
Jan 12, 2022 4:00:45 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-12T16:00:44.617Z: Stopping **** pool...
Jan 12, 2022 4:00:45 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-12T16:00:44.661Z: Stopping **** pool...
Jan 12, 2022 4:03:06 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-12T16:03:05.472Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Jan 12, 2022 4:03:06 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-12T16:03:05.516Z: Worker pool stopped.
Jan 12, 2022 4:03:11 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-01-12_04_32_54-4753480158470970014 finished with status CANCELLED.
Load test results for test (ID): e7d660e1-79ae-4204-a064-2367f3eb10ce and timestamp: 2022-01-12T12:32:47.718000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                 12317.688
dataflow_v2_java17_total_bytes_count             2.69268555E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220112123035
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:eae73b74c7cee1be2d90dec0adb44e27fc9798b467f822dccfb3e0407214c58c
Deleted: sha256:72b23024dd8c0e02a625b1aaa377d70ddace4ab9c4dd5d0c49a9bfc58a951bd3
Deleted: sha256:9f8e53140fcf27cda74579f0834935ef37a0bca5235e54a33658bfafa050d2bb
Deleted: sha256:7a76b4010998c4b7f62f85723f5523611e6abb8bc173d560fba7210ee84d0c79
Deleted: sha256:11636dcdac63f52177a29570c7bbee66b69cc256dd45a71255bd9754044c3bf5
Deleted: sha256:bd5984eacdc892d5e759bafac7c23f52281fc3df1d31c548d8e37f73a33bee77
Deleted: sha256:8edf0fad1212b28182be1271b4b2198682433fe48c3fe62d6f446b75c42b7e23
Deleted: sha256:910bc4e069ef99337f2f6d6b2df0fc12d9c67f8884260498b562d843c3753223
Deleted: sha256:70d2895f43a07cc54ae87093e993217a99ff7e4a803c4cff5af2b15936e5c527
Deleted: sha256:073cb81c592be58ebcecbbc6e143d2c4b957fb8f80d684a1dd0528c6092c9a60
Deleted: sha256:f26b7ba5a01c346eca9a07a29b4829662821f3bbe64068e1e7522f045ac9c0e4
Deleted: sha256:f55fb3dbe6451854acbf990224d82dc944dede0ca5200bf356677f0de1695061
Deleted: sha256:27b5ee9c93772f41df4526678e00b318b5739da4e14ad15b4c7d806c3f16c8b5
Deleted: sha256:f577276746d1c543957a0ab560fcbd2a32060c54588fece9613f028f982490f7
Deleted: sha256:ab2dc75e00b2d1936ce4bdb8f438c151740982efe6ab2270e3c1d514464504be
Deleted: sha256:c35bb8289492e86aad84169ab3c3edd09fd159caf3b4e2f9b0cc92b217bdaaf1
Deleted: sha256:0374273f6af699e49e39e52ef4e377e9d7fbd06b27390bd4e6ae21a249c79be0
Deleted: sha256:9e1a2009be7d17d27b4e7ff97043a50a32ac03d48e9db837d7fa7bdf440420ee
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220112123035]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:eae73b74c7cee1be2d90dec0adb44e27fc9798b467f822dccfb3e0407214c58c]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220112123035] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:eae73b74c7cee1be2d90dec0adb44e27fc9798b467f822dccfb3e0407214c58c])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:eae73b74c7cee1be2d90dec0adb44e27fc9798b467f822dccfb3e0407214c58c
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:eae73b74c7cee1be2d90dec0adb44e27fc9798b467f822dccfb3e0407214c58c
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:eae73b74c7cee1be2d90dec0adb44e27fc9798b467f822dccfb3e0407214c58c].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 32m 54s
109 actionable tasks: 73 executed, 32 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/pfe2afcxmchua

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #26

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/26/display/redirect?page=changes>

Changes:

[relax] don't close streams in finalize

[noreply] Loosen typing extensions bound

[Alexey Romanenko] [BEAM-4868] Bump com.amazonaws to 1.12.135

[noreply] [BEAM-12092] Bump jedis to version 4.0.1 (#16287)

[noreply] [BEAM-13534] Add automated port polling to expansion service runner if

[noreply] Merge pull request #16344 from [BEAM-13536][Playground][Bugfix] CI step

[noreply] Merge pull request #16359 from [BEAM-13545][Playground] Add

[noreply] Merge pull request #16384 from [BEAM-13308] [Playground] Getting

[noreply] Merge pull request #16306 from [BEAM-13447] [Playground] Add filling of

[noreply] Merge pull request #16361 from [BEAM-13543][Playground] Add logic of

[noreply] [BEAM-12562] Dataframe pipe implementation (#16256)

[noreply] Merge pull request #16338 from [BEAM-13528][Playground] Add liveness

[noreply] [BEAM-13626] Remap expanded outputs after merging. (#16471)

[noreply] Merge pull request #16147 from [BEAM-13359] [Playground] Tag existing

[noreply] [BEAM-3221] Improve documentation in model pipeline protos (#16474)

[noreply] [BEAM-13614] Add OnWindowExpiration support to the Java SDK harness and

[noreply] Merge pull request #16156 from [BEAM-13391] Fix temporary file format in


------------------------------------------
[...truncated 49.70 KB...]
26a504e63be4: Waiting
31892cc314cb: Waiting
e29a09801a7d: Waiting
04c4da4e5aa4: Waiting
cc0f4c8d4b83: Waiting
3e094c340ade: Waiting
a5be962eb406: Waiting
c779b5a100e9: Waiting
468c4b6130ba: Pushed
1c7e7a7e57c4: Pushed
4aa8b04d294b: Pushed
6f7978e1999d: Pushed
88e741cbfdbc: Pushed
1dabc3157eb4: Pushed
3e094c340ade: Pushed
04c4da4e5aa4: Pushed
e29a09801a7d: Pushed
bff9fe6e429c: Layer already exists
7c245b2fe4f1: Layer already exists
f9e18e59a565: Layer already exists
26a504e63be4: Layer already exists
8bf42db0de72: Layer already exists
31892cc314cb: Layer already exists
11936051f93b: Layer already exists
cc0f4c8d4b83: Pushed
a5be962eb406: Pushed
6576bc269d69: Pushed
c779b5a100e9: Pushed
20220111123216: digest: sha256:042af543950b044c4636231aebd35b8c26ce7a993188224632086cba1f9e8c73 size: 4520

> Task :sdks:java:testing:load-tests:run
Jan 11, 2022 12:34:08 PM org.apache.beam.runners.dataflow.options.DataflowPipelineOptions$StagingLocationFactory create
INFO: No stagingLocation provided, falling back to gcpTempLocation
Jan 11, 2022 12:34:08 PM org.apache.beam.runners.dataflow.DataflowRunner fromOptions
INFO: PipelineOptions.filesToStage was not specified. Defaulting to files from the classpath: will stage 203 files. Enable logging at DEBUG level to see which files will be staged.
Jan 11, 2022 12:34:09 PM org.apache.beam.sdk.Pipeline validate
WARNING: The following transforms do not have stable unique names: Window.Into()
Jan 11, 2022 12:34:09 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Executing pipeline on the Dataflow Service, which will have billing implications related to Google Compute Engine usage and other Google Cloud Services.
Jan 11, 2022 12:34:11 PM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Uploading 203 files from PipelineOptions.filesToStage to staging location to prepare for execution.
Jan 11, 2022 12:34:12 PM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Staging files complete: 203 files cached, 0 files newly uploaded in 0 seconds
Jan 11, 2022 12:34:12 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Staging portable pipeline proto to gs://temp-storage-for-perf-tests/loadtests/staging/
Jan 11, 2022 12:34:12 PM org.apache.beam.runners.dataflow.util.PackageUtil tryStagePackage
INFO: Uploading <113691 bytes, hash 63d28af4bc9d824168992964dc02dc52bb8740aca80233f0d75c9d529879c181> to gs://temp-storage-for-perf-tests/loadtests/staging/pipeline-Y9KK9LydgkFomSlk3ALcUruHQKyoAjPw11ydUph5wYE.pb
Jan 11, 2022 12:34:14 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read input/DataflowRunner.StreamingUnboundedRead.ReadWithIds as step s1
Jan 11, 2022 12:34:14 PM org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource split
INFO: Split into 20 bundles of sizes: [org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4aa21f9d, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@71c17a57, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@640ab13c, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1e0a864d, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@440e3ce6, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3e67f5f2, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@fd53053, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4527f70a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@707b1a44, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7132a9dc, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@57435801, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2da66a44, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@527fc8e, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@61bfc9bf, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2c7106d9, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@329bad59, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@862f408, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@178f268a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@33ecbd6c, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5c723f2d]
Jan 11, 2022 12:34:14 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read input/StripIds as step s2
Jan 11, 2022 12:34:14 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect start time metrics (input) as step s3
Jan 11, 2022 12:34:14 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Window.Into()/Window.Assign as step s4
Jan 11, 2022 12:34:14 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read co-input/DataflowRunner.StreamingUnboundedRead.ReadWithIds as step s5
Jan 11, 2022 12:34:14 PM org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource split
INFO: Split into 20 bundles of sizes: [org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@24eb65e3, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3ac3f6f, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1abebef3, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@18f55704, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@67cefd84, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5fbe155, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6add8e3f, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@58a2b917, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@48904d5a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@12bbfc54, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1491344a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@59b65dce, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1386313f, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5e922647, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@433c6abb, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@288f173f, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@a22c4d8, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@45cd7bc5, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6504a875, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@35e26d05]
Jan 11, 2022 12:34:14 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read co-input/StripIds as step s6
Jan 11, 2022 12:34:14 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect start time metrics (co-input) as step s7
Jan 11, 2022 12:34:14 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Window.Into()2/Window.Assign as step s8
Jan 11, 2022 12:34:14 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/MakeUnionTable0 as step s9
Jan 11, 2022 12:34:14 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/MakeUnionTable1 as step s10
Jan 11, 2022 12:34:14 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/Flatten as step s11
Jan 11, 2022 12:34:14 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/GBK as step s12
Jan 11, 2022 12:34:14 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/ConstructCoGbkResultFn as step s13
Jan 11, 2022 12:34:14 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Ungroup and reiterate as step s14
Jan 11, 2022 12:34:14 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect total bytes as step s15
Jan 11, 2022 12:34:14 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect end time metrics as step s16
Jan 11, 2022 12:34:14 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Dataflow SDK version: 2.37.0-SNAPSHOT
Jan 11, 2022 12:34:17 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To access the Dataflow monitoring console, please navigate to https://console.cloud.google.com/dataflow/jobs/us-central1/2022-01-11_04_34_14-17177141928381302403?project=apache-beam-testing
Jan 11, 2022 12:34:17 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Submitted job: 2022-01-11_04_34_14-17177141928381302403
Jan 11, 2022 12:34:17 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To cancel the job using the 'gcloud' tool, run:
> gcloud dataflow jobs --project=apache-beam-testing cancel --region=us-central1 2022-01-11_04_34_14-17177141928381302403
Jan 11, 2022 12:34:27 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
WARNING: 2022-01-11T12:34:23.810Z: The workflow name is not a valid Cloud Label. Labels applied to Cloud resources (such as GCE Instances) for monitoring will be labeled with this modified job name: load0tests0java170dataflow0v20streaming0cogbk01-jenkins-01-9ey8. For the best monitoring experience, please name your job with a valid Cloud Label. For details, see: https://cloud.google.com/compute/docs/labeling-resources#restrictions
Jan 11, 2022 12:34:32 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-11T12:34:31.937Z: Worker configuration: e2-standard-2 in us-central1-f.
Jan 11, 2022 12:34:33 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-11T12:34:32.577Z: Expanding SplittableParDo operations into optimizable parts.
Jan 11, 2022 12:34:33 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-11T12:34:32.619Z: Expanding CollectionToSingleton operations into optimizable parts.
Jan 11, 2022 12:34:33 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-11T12:34:32.676Z: Expanding CoGroupByKey operations into optimizable parts.
Jan 11, 2022 12:34:33 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-11T12:34:32.775Z: Expanding SplittableProcessKeyed operations into optimizable parts.
Jan 11, 2022 12:34:33 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-11T12:34:32.821Z: Expanding GroupByKey operations into streaming Read/Write steps
Jan 11, 2022 12:34:33 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-11T12:34:32.881Z: Lifting ValueCombiningMappingFns into MergeBucketsMappingFns
Jan 11, 2022 12:34:33 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-11T12:34:32.978Z: Fusing adjacent ParDo, Read, Write, and Flatten operations
Jan 11, 2022 12:34:33 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-11T12:34:33.037Z: Unzipping flatten CoGroupByKey-Flatten for input CoGroupByKey-MakeUnionTable0-ParMultiDo-ConstructUnionTable-.output
Jan 11, 2022 12:34:33 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-11T12:34:33.073Z: Fusing unzipped copy of CoGroupByKey/GBK/WriteStream, through flatten CoGroupByKey/Flatten, into producer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable)
Jan 11, 2022 12:34:33 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-11T12:34:33.100Z: Fusing consumer CoGroupByKey/GBK/WriteStream into CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable)
Jan 11, 2022 12:34:33 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-11T12:34:33.134Z: Fusing consumer Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read input/Impulse
Jan 11, 2022 12:34:33 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-11T12:34:33.166Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
Jan 11, 2022 12:34:33 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-11T12:34:33.192Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
Jan 11, 2022 12:34:33 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-11T12:34:33.221Z: Fusing consumer Read input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
Jan 11, 2022 12:34:33 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-11T12:34:33.256Z: Fusing consumer Collect start time metrics (input)/ParMultiDo(TimeMonitor) into Read input/ParDo(StripIds)/ParMultiDo(StripIds)
Jan 11, 2022 12:34:33 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-11T12:34:33.286Z: Fusing consumer Window.Into()/Window.Assign into Collect start time metrics (input)/ParMultiDo(TimeMonitor)
Jan 11, 2022 12:34:33 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-11T12:34:33.319Z: Fusing consumer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable) into Window.Into()/Window.Assign
Jan 11, 2022 12:34:33 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-11T12:34:33.355Z: Fusing consumer Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read co-input/Impulse
Jan 11, 2022 12:34:33 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-11T12:34:33.405Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
Jan 11, 2022 12:34:33 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-11T12:34:33.479Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
Jan 11, 2022 12:34:33 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-11T12:34:33.530Z: Fusing consumer Read co-input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
Jan 11, 2022 12:34:33 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-11T12:34:33.570Z: Fusing consumer Collect start time metrics (co-input)/ParMultiDo(TimeMonitor) into Read co-input/ParDo(StripIds)/ParMultiDo(StripIds)
Jan 11, 2022 12:34:33 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-11T12:34:33.617Z: Fusing consumer Window.Into()2/Window.Assign into Collect start time metrics (co-input)/ParMultiDo(TimeMonitor)
Jan 11, 2022 12:34:33 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-11T12:34:33.657Z: Fusing consumer CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable) into Window.Into()2/Window.Assign
Jan 11, 2022 12:34:33 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-11T12:34:33.689Z: Fusing consumer CoGroupByKey/GBK/MergeBuckets into CoGroupByKey/GBK/ReadStream
Jan 11, 2022 12:34:33 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-11T12:34:33.744Z: Fusing consumer CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult) into CoGroupByKey/GBK/MergeBuckets
Jan 11, 2022 12:34:33 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-11T12:34:33.783Z: Fusing consumer Ungroup and reiterate/ParMultiDo(UngroupAndReiterate) into CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult)
Jan 11, 2022 12:34:33 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-11T12:34:33.821Z: Fusing consumer Collect total bytes/ParMultiDo(ByteMonitor) into Ungroup and reiterate/ParMultiDo(UngroupAndReiterate)
Jan 11, 2022 12:34:33 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-11T12:34:33.848Z: Fusing consumer Collect end time metrics/ParMultiDo(TimeMonitor) into Collect total bytes/ParMultiDo(ByteMonitor)
Jan 11, 2022 12:34:36 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-11T12:34:34.282Z: Starting 5 ****s in us-central1-f...
Jan 11, 2022 12:35:08 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-11T12:35:06.513Z: Your project already contains 100 Dataflow-created metric descriptors, so new user metrics of the form custom.googleapis.com/* will not be created. However, all user metrics are also available in the metric dataflow.googleapis.com/job/user_counter. If you rely on the custom metrics, you can delete old / unused metric descriptors. See https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.list and https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.delete
Jan 11, 2022 12:35:26 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-11T12:35:25.081Z: Autoscaling: Raised the number of ****s to 5 so that the pipeline can catch up with its backlog and keep up with its input rate.
Jan 11, 2022 12:36:24 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-11T12:36:23.012Z: Workers have started successfully.
Jan 11, 2022 12:36:24 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-11T12:36:23.043Z: Workers have started successfully.
Jan 11, 2022 4:00:38 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-11T16:00:37.874Z: Cancel request is committed for workflow job: 2022-01-11_04_34_14-17177141928381302403.
Jan 11, 2022 4:00:38 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-11T16:00:37.952Z: Cleaning up.
Jan 11, 2022 4:00:38 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-11T16:00:38.031Z: Stopping **** pool...
Jan 11, 2022 4:00:38 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-11T16:00:38.094Z: Stopping **** pool...
Jan 11, 2022 4:03:09 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-11T16:03:08.132Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Jan 11, 2022 4:03:09 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-11T16:03:08.168Z: Worker pool stopped.
Jan 11, 2022 4:03:13 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-01-11_04_34_14-17177141928381302403 finished with status CANCELLED.
Load test results for test (ID): 72a39f32-a0ad-420f-b7e4-09bbbf9f2f06 and timestamp: 2022-01-11T12:34:09.021000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                  12213.03
dataflow_v2_java17_total_bytes_count             2.16478782E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220111123216
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:042af543950b044c4636231aebd35b8c26ce7a993188224632086cba1f9e8c73
Deleted: sha256:7d6956f029465d6705691f8012cb9c6394d72bb41a343ec33e50c4e4dd9fc627
Deleted: sha256:4222b680a65f4a09c4e8f1bc84d436d28d03bf62c908c0332764693d77426d0f
Deleted: sha256:0b494c7933df536ed3fcfa6b8d94ab23ea696959177be872ef0398f7222f0a77
Deleted: sha256:81ee4047cf0065e3fb45b82b9a4a9287f26be5cc9abc465334de4002372e0fb5
Deleted: sha256:ba7d411e9043c9e87fdfac8c9026a91a8399134befedeae222ae5bbecccd70b1
Deleted: sha256:fb47083a656863c83a8e73619ef08cd2b626eaf0a7aa9e6e93a705e16bd38cc9
Deleted: sha256:256563dab1799c6cacf4e7b3ee0f17e3d23d0db44f1a08044593203fc563cc21
Deleted: sha256:40b28b3303b1e2f8bba4a679d42b4e50441fde28d52af0108d9172f4acad5ec8
Deleted: sha256:b7259671c945080e02f78d586844aee198755dd9b7abef1ff7efadd4d55b8eda
Deleted: sha256:5d63a5ce7432ef96e9c7aa2a7825ea9d55bcbfac3393808388723d02fb217bd0
Deleted: sha256:3bc695bd64c31cc9940cbb147dd039ee2098566f498af97577fdb83cc74e0ff4
Deleted: sha256:50f4e8ec363eb72ba6b22f4287e37a59d90e4303b62f162ebe471c68c45e690e
Deleted: sha256:f8870a05ef6a5b992fef774888a1a14b05039a8e0f703bfc4d145d0fde03abf2
Deleted: sha256:efca6ad31a191c87ec8a868d8adfe679e01ad64f27e2d7b3f5a3ade3eeb133cb
Deleted: sha256:c63d58b6f1b108cdfdb3f4fe204be530561f0c3cef52ab8d26fe2f0a05b8bde1
Deleted: sha256:ccf7ee1a3526d2ff4d6e20c27f02848b4c962eae9d1b31235a5b958987d6d986
Deleted: sha256:c7ea9f07f5c3b156ff7929dd9f75d24bedb2dfaad6778aa27f48612e9c9660f1
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220111123216]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:042af543950b044c4636231aebd35b8c26ce7a993188224632086cba1f9e8c73]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220111123216] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:042af543950b044c4636231aebd35b8c26ce7a993188224632086cba1f9e8c73])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:042af543950b044c4636231aebd35b8c26ce7a993188224632086cba1f9e8c73
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:042af543950b044c4636231aebd35b8c26ce7a993188224632086cba1f9e8c73
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:042af543950b044c4636231aebd35b8c26ce7a993188224632086cba1f9e8c73].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 31m 14s
109 actionable tasks: 72 executed, 33 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/s3rt6rroinbuo

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #25

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/25/display/redirect?page=changes>

Changes:

[Alexey Romanenko] [BEAM-8727] Bump software.amazon.awssdk to 2.17.106


------------------------------------------
[...truncated 51.17 KB...]
INFO: Adding Collect start time metrics (input) as step s3
Jan 10, 2022 12:39:59 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Window.Into()/Window.Assign as step s4
Jan 10, 2022 12:39:59 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read co-input/DataflowRunner.StreamingUnboundedRead.ReadWithIds as step s5
Jan 10, 2022 12:39:59 PM org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource split
INFO: Split into 20 bundles of sizes: [org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4a37191a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5854a18, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@d5556bf, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@791c12e3, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@b112b13, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@24eb65e3, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3ac3f6f, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1abebef3, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@18f55704, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@67cefd84, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5fbe155, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6add8e3f, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@58a2b917, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@48904d5a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@12bbfc54, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1491344a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@59b65dce, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1386313f, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5e922647, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@433c6abb]
Jan 10, 2022 12:39:59 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read co-input/StripIds as step s6
Jan 10, 2022 12:39:59 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect start time metrics (co-input) as step s7
Jan 10, 2022 12:39:59 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Window.Into()2/Window.Assign as step s8
Jan 10, 2022 12:39:59 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/MakeUnionTable0 as step s9
Jan 10, 2022 12:39:59 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/MakeUnionTable1 as step s10
Jan 10, 2022 12:39:59 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/Flatten as step s11
Jan 10, 2022 12:39:59 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/GBK as step s12
Jan 10, 2022 12:39:59 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/ConstructCoGbkResultFn as step s13
Jan 10, 2022 12:39:59 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Ungroup and reiterate as step s14
Jan 10, 2022 12:39:59 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect total bytes as step s15
Jan 10, 2022 12:39:59 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect end time metrics as step s16
Jan 10, 2022 12:39:59 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Dataflow SDK version: 2.37.0-SNAPSHOT
Jan 10, 2022 12:40:00 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To access the Dataflow monitoring console, please navigate to https://console.cloud.google.com/dataflow/jobs/us-central1/2022-01-10_04_39_59-15736926293421518288?project=apache-beam-testing
Jan 10, 2022 12:40:00 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Submitted job: 2022-01-10_04_39_59-15736926293421518288
Jan 10, 2022 12:40:00 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To cancel the job using the 'gcloud' tool, run:
> gcloud dataflow jobs --project=apache-beam-testing cancel --region=us-central1 2022-01-10_04_39_59-15736926293421518288
Jan 10, 2022 12:40:08 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
WARNING: 2022-01-10T12:40:06.796Z: The workflow name is not a valid Cloud Label. Labels applied to Cloud resources (such as GCE Instances) for monitoring will be labeled with this modified job name: load0tests0java170dataflow0v20streaming0cogbk01-jenkins-01-slwf. For the best monitoring experience, please name your job with a valid Cloud Label. For details, see: https://cloud.google.com/compute/docs/labeling-resources#restrictions
Jan 10, 2022 12:40:12 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-10T12:40:11.072Z: Worker configuration: e2-standard-2 in us-central1-f.
Jan 10, 2022 12:40:12 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-10T12:40:11.838Z: Expanding SplittableParDo operations into optimizable parts.
Jan 10, 2022 12:40:12 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-10T12:40:11.869Z: Expanding CollectionToSingleton operations into optimizable parts.
Jan 10, 2022 12:40:12 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-10T12:40:11.919Z: Expanding CoGroupByKey operations into optimizable parts.
Jan 10, 2022 12:40:12 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-10T12:40:12.036Z: Expanding SplittableProcessKeyed operations into optimizable parts.
Jan 10, 2022 12:40:12 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-10T12:40:12.089Z: Expanding GroupByKey operations into streaming Read/Write steps
Jan 10, 2022 12:40:15 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-10T12:40:12.147Z: Lifting ValueCombiningMappingFns into MergeBucketsMappingFns
Jan 10, 2022 12:40:15 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-10T12:40:12.247Z: Fusing adjacent ParDo, Read, Write, and Flatten operations
Jan 10, 2022 12:40:15 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-10T12:40:12.284Z: Unzipping flatten CoGroupByKey-Flatten for input CoGroupByKey-MakeUnionTable0-ParMultiDo-ConstructUnionTable-.output
Jan 10, 2022 12:40:15 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-10T12:40:12.307Z: Fusing unzipped copy of CoGroupByKey/GBK/WriteStream, through flatten CoGroupByKey/Flatten, into producer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable)
Jan 10, 2022 12:40:15 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-10T12:40:12.332Z: Fusing consumer CoGroupByKey/GBK/WriteStream into CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable)
Jan 10, 2022 12:40:15 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-10T12:40:12.365Z: Fusing consumer Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read input/Impulse
Jan 10, 2022 12:40:15 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-10T12:40:12.397Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
Jan 10, 2022 12:40:15 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-10T12:40:12.433Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
Jan 10, 2022 12:40:15 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-10T12:40:12.470Z: Fusing consumer Read input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
Jan 10, 2022 12:40:15 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-10T12:40:12.491Z: Fusing consumer Collect start time metrics (input)/ParMultiDo(TimeMonitor) into Read input/ParDo(StripIds)/ParMultiDo(StripIds)
Jan 10, 2022 12:40:15 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-10T12:40:12.523Z: Fusing consumer Window.Into()/Window.Assign into Collect start time metrics (input)/ParMultiDo(TimeMonitor)
Jan 10, 2022 12:40:15 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-10T12:40:12.570Z: Fusing consumer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable) into Window.Into()/Window.Assign
Jan 10, 2022 12:40:15 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-10T12:40:12.597Z: Fusing consumer Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read co-input/Impulse
Jan 10, 2022 12:40:15 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-10T12:40:12.618Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
Jan 10, 2022 12:40:15 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-10T12:40:12.655Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
Jan 10, 2022 12:40:15 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-10T12:40:12.679Z: Fusing consumer Read co-input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
Jan 10, 2022 12:40:15 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-10T12:40:12.708Z: Fusing consumer Collect start time metrics (co-input)/ParMultiDo(TimeMonitor) into Read co-input/ParDo(StripIds)/ParMultiDo(StripIds)
Jan 10, 2022 12:40:15 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-10T12:40:12.740Z: Fusing consumer Window.Into()2/Window.Assign into Collect start time metrics (co-input)/ParMultiDo(TimeMonitor)
Jan 10, 2022 12:40:15 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-10T12:40:12.775Z: Fusing consumer CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable) into Window.Into()2/Window.Assign
Jan 10, 2022 12:40:15 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-10T12:40:12.807Z: Fusing consumer CoGroupByKey/GBK/MergeBuckets into CoGroupByKey/GBK/ReadStream
Jan 10, 2022 12:40:15 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-10T12:40:12.840Z: Fusing consumer CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult) into CoGroupByKey/GBK/MergeBuckets
Jan 10, 2022 12:40:15 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-10T12:40:12.868Z: Fusing consumer Ungroup and reiterate/ParMultiDo(UngroupAndReiterate) into CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult)
Jan 10, 2022 12:40:15 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-10T12:40:12.907Z: Fusing consumer Collect total bytes/ParMultiDo(ByteMonitor) into Ungroup and reiterate/ParMultiDo(UngroupAndReiterate)
Jan 10, 2022 12:40:15 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-10T12:40:12.941Z: Fusing consumer Collect end time metrics/ParMultiDo(TimeMonitor) into Collect total bytes/ParMultiDo(ByteMonitor)
Jan 10, 2022 12:40:15 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-10T12:40:13.324Z: Starting 5 ****s in us-central1-f...
Jan 10, 2022 12:40:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-10T12:40:43.001Z: Your project already contains 100 Dataflow-created metric descriptors, so new user metrics of the form custom.googleapis.com/* will not be created. However, all user metrics are also available in the metric dataflow.googleapis.com/job/user_counter. If you rely on the custom metrics, you can delete old / unused metric descriptors. See https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.list and https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.delete
Jan 10, 2022 12:41:02 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-10T12:41:01.762Z: Autoscaling: Raised the number of ****s to 5 so that the pipeline can catch up with its backlog and keep up with its input rate.
Jan 10, 2022 12:41:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-10T12:41:54.396Z: Workers have started successfully.
Jan 10, 2022 12:41:54 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-10T12:41:54.423Z: Workers have started successfully.
Jan 10, 2022 12:52:14 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
SEVERE: 2022-01-10T12:52:14.180Z: Staged package beam-vendor-grpc-1_36_0-0.2-2X-EhRNYA0GgyvffOgWohko1OyyyPyTMuiHHvaPG9es.jar at location 'gs://temp-storage-for-perf-tests/loadtests/staging/beam-vendor-grpc-1_36_0-0.2-2X-EhRNYA0GgyvffOgWohko1OyyyPyTMuiHHvaPG9es.jar' is inaccessible.
Jan 10, 2022 12:52:16 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
SEVERE: 2022-01-10T12:52:14.349Z: Staged package cldrdata-k07I6K9W3X5KTQbcDIEsqM0LXyM18f0eR6IaJw-P_xk.jar at location 'gs://temp-storage-for-perf-tests/loadtests/staging/cldrdata-k07I6K9W3X5KTQbcDIEsqM0LXyM18f0eR6IaJw-P_xk.jar' is inaccessible.
Jan 10, 2022 12:52:16 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
SEVERE: 2022-01-10T12:52:14.604Z: Staged package dnsns-RGhCDg3GVOQVC2r6ka2N0hmI4eqQH6VobuoAnQ74MnE.jar at location 'gs://temp-storage-for-perf-tests/loadtests/staging/dnsns-RGhCDg3GVOQVC2r6ka2N0hmI4eqQH6VobuoAnQ74MnE.jar' is inaccessible.
Jan 10, 2022 12:52:16 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
SEVERE: 2022-01-10T12:52:14.683Z: Staged package flatbuffers-java-1.12.0-P4wIi03QSphYch8uFiUIyU2w3Yb5YeMG7mPvLtqHG_c.jar at location 'gs://temp-storage-for-perf-tests/loadtests/staging/flatbuffers-java-1.12.0-P4wIi03QSphYch8uFiUIyU2w3Yb5YeMG7mPvLtqHG_c.jar' is inaccessible.
Jan 10, 2022 12:52:16 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
SEVERE: 2022-01-10T12:52:15.819Z: Staged package icedtea-sound-Bdoi2wYa757-fzq5vconCy4SSQ22ZaOq7yuC98fKPs8.jar at location 'gs://temp-storage-for-perf-tests/loadtests/staging/icedtea-sound-Bdoi2wYa757-fzq5vconCy4SSQ22ZaOq7yuC98fKPs8.jar' is inaccessible.
Jan 10, 2022 12:52:16 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
SEVERE: 2022-01-10T12:52:15.873Z: Staged package jaccess-CMbK-IOdQPLKHEqCuDnE-yBk-VpbtVT-hgjbHRUGO78.jar at location 'gs://temp-storage-for-perf-tests/loadtests/staging/jaccess-CMbK-IOdQPLKHEqCuDnE-yBk-VpbtVT-hgjbHRUGO78.jar' is inaccessible.
Jan 10, 2022 12:52:16 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
SEVERE: 2022-01-10T12:52:16.099Z: Staged package joda-time-2.10.10-3Y58khhaZ40be5M_MSCbYgPI_6kemIBHWhvgNGuWF-M.jar at location 'gs://temp-storage-for-perf-tests/loadtests/staging/joda-time-2.10.10-3Y58khhaZ40be5M_MSCbYgPI_6kemIBHWhvgNGuWF-M.jar' is inaccessible.
Jan 10, 2022 12:52:16 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
SEVERE: 2022-01-10T12:52:16.284Z: Staged package localedata-ae5Z0L6ak4922fztWeWy7ajiWXdG3ubNrwerJRFoFj0.jar at location 'gs://temp-storage-for-perf-tests/loadtests/staging/localedata-ae5Z0L6ak4922fztWeWy7ajiWXdG3ubNrwerJRFoFj0.jar' is inaccessible.
Jan 10, 2022 12:52:16 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
SEVERE: 2022-01-10T12:52:16.378Z: Staged package nashorn-XHtz_UehGpYcLTOrATrTnMNVUgEVa_ttoWkPxnVfqTo.jar at location 'gs://temp-storage-for-perf-tests/loadtests/staging/nashorn-XHtz_UehGpYcLTOrATrTnMNVUgEVa_ttoWkPxnVfqTo.jar' is inaccessible.
Jan 10, 2022 12:52:17 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
WARNING: 2022-01-10T12:52:17.088Z: One or more access checks for temp location or staged files failed. Please refer to other error messages for details. For more information on security and permissions, please see https://cloud.google.com/dataflow/security-and-permissions.
Jan 10, 2022 4:00:44 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-10T16:00:41.684Z: Cancel request is committed for workflow job: 2022-01-10_04_39_59-15736926293421518288.
Jan 10, 2022 4:00:44 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-10T16:00:41.751Z: Cleaning up.
Jan 10, 2022 4:00:44 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-10T16:00:41.823Z: Stopping **** pool...
Jan 10, 2022 4:00:44 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-10T16:00:41.865Z: Stopping **** pool...
Jan 10, 2022 4:03:12 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-10T16:03:11.241Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Jan 10, 2022 4:03:12 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-10T16:03:11.278Z: Worker pool stopped.
Jan 10, 2022 4:03:17 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-01-10_04_39_59-15736926293421518288 finished with status CANCELLED.
Load test results for test (ID): f09c43a2-02c4-49b8-8571-570b88a08584 and timestamp: 2022-01-10T12:39:53.438000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                 11890.045
dataflow_v2_java17_total_bytes_count             1.30727934E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220110123752
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:0fea859cd5bf6fe50d01a96741742d148f794e15cb75d3e1f9c94f260a4094d5
Deleted: sha256:ba7b50fbf320168acc9cd1335c324607fe611547419a57ad3bb8c2dc7428def5
Deleted: sha256:b80d2947d1a42da9fce6bd47333a60c270f327243f52c72cc0c78b8d4b49b6f8
Deleted: sha256:21543f48256fd3649f9827a4cf74ac7154aa98481df64892724b7309f96a0967
Deleted: sha256:42e8eb07cbc5b3c54018e6eac1140836381f1e8fbcb2789fab7e98c04486372c
Deleted: sha256:98876176a02d3a299f0424072cd58220573a5bc6ed4cf17520b7d64722724c75
Deleted: sha256:deab8d669ddafde69893ebfb23f4fc2a57b92a25bfaf846c5b3a99e27ed8cccd
Deleted: sha256:198b9e6f3f609fb8ce33ebaad57e788239261b36ca078940ed47d02d8b77422f
Deleted: sha256:4c01b2a95ef6bc5c20113b09ab96de947f4ed3b32897d010747d91708d2033f6
Deleted: sha256:4a4b3bf0c1798ce8f384262618fd52286b1e6e8eda5964164342656bc5b3a2ef
Deleted: sha256:ca8d12747da8e754dc807801f70064f83f3de952977835e16fd47312efc85083
Deleted: sha256:a04b27336887dd41ab0af3d8c360898126d8a63db4a4e09582ef5e59be0d79d8
Deleted: sha256:5cefdac6abaa403a457f872b1a53213899691e132916e308a445012cada7d71e
Deleted: sha256:66fba344762eb9a89467aeab762fdaff7c9800acc64eff703c372f829da8017f
Deleted: sha256:9fd7507f6a2758cb6966f7d9bb61fe4f4f40ef031086d4617043833f593bd5ea
Deleted: sha256:e3ef3358e8a613328c5ff09ca1955b545d900e137a82d13b7fbc89e468479242
Deleted: sha256:44c7f15f6042aec2125ce0728c235d52f3ca6a0b0a16c2da296a376b53e4868b
Deleted: sha256:f58e6f387dee1bc883afdb26c505e1d00ad531f12bcd074a5ca094b50c2c5beb
Deleted: sha256:04ed0c861a6f3e2c1152f08d6333c9828b12b2008eb6689b5e67b1911e92fcb0
Deleted: sha256:f317e398d9fae2097fa1aa469efe74aa3de7e78d49ad1c2aef46b43f38f388ff
Deleted: sha256:cb8a6871b5ee1bf01f36311e4be672ae1df8f1156c29e822f323f27b8309f360
Deleted: sha256:6164beb0344d987d9b8ce394d8d009eb752d1cb5058b4ea4f1e73cda4dfdb701
Deleted: sha256:2cf1ebaa63f4197ab705502f53a9b155cfaac7a5234529d7cf491d16a78a3467
Deleted: sha256:eaae951a72883eb231511a6917151217debb193ef39cd079a7e5f1ed679f08ea
Deleted: sha256:52f253e799595766774650af8f4a2ea2067b040d64a8a5de7ab0348b93b6cd8c
Deleted: sha256:7069679fc4d196e1faf5dc59be1c79485d88de81081df9ac95a4ddf483d3630e
Deleted: sha256:8ed8c0a29b65c7cd379fbc0026fc94e18386887a2c548caf4dc270e99a40c69b
Deleted: sha256:179c32628b941f4b58e5e1727ddaed1084a0debcf8c54039c827eeaa686c1edb
Deleted: sha256:4f0082d55aef90f65355b94411d9968b100f734f8c55537277d6b8a6d8a47dc2
Deleted: sha256:e3466fd734587afc4dd60a113755765e5adebc32d18d7061db125d52b7189d87
Deleted: sha256:1b4f529dcb2ffe7a1d4c457a798d3af7b0125a1a7b890545084e080aa52fb290
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220110123752]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:0fea859cd5bf6fe50d01a96741742d148f794e15cb75d3e1f9c94f260a4094d5]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220110123752] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:0fea859cd5bf6fe50d01a96741742d148f794e15cb75d3e1f9c94f260a4094d5])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:63aa69ba85114b9a7e55d1e8cf4c7c5ef2151b9ae2729504823103f9c2e7130c
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:63aa69ba85114b9a7e55d1e8cf4c7c5ef2151b9ae2729504823103f9c2e7130c
ERROR: (gcloud.container.images.delete) Not found: response: {'status': '404', 'content-length': '168', 'x-xss-protection': '0', 'transfer-encoding': 'chunked', 'server': 'Docker Registry', '-content-encoding': 'gzip', 'docker-distribution-api-version': 'registry/2.0', 'cache-control': 'private', 'date': 'Mon, 10 Jan 2022 16:03:20 GMT', 'x-frame-options': 'SAMEORIGIN', 'content-type': 'application/json'}
Failed to compute blob liveness for manifest: 'sha256:63aa69ba85114b9a7e55d1e8cf4c7c5ef2151b9ae2729504823103f9c2e7130c': None

> Task :runners:google-cloud-dataflow-java:cleanUpDockerImages FAILED

FAILURE: Build completed with 2 failures.

1: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.
==============================================================================

2: Task failed with an exception.
-----------
* Where:
Build file '<https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/ws/src/runners/google-cloud-dataflow-java/build.gradle'> line: 290

* What went wrong:
Execution failed for task ':runners:google-cloud-dataflow-java:cleanUpDockerImages'.
> Process 'command './scripts/cleanup_untagged_gcr_images.sh'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.
==============================================================================

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 25m 41s
109 actionable tasks: 75 executed, 30 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/w72xgajshzb5g

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17 #24

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_LoadTests_Java_CoGBK_Dataflow_V2_Streaming_Java17/24/display/redirect>

Changes:


------------------------------------------
[...truncated 49.38 KB...]
595295217575: Pushed
4053b0d7663e: Pushed
e95d6ee8311f: Pushed
ddab198eb079: Pushed
47e9c0170cd0: Pushed
bff9fe6e429c: Layer already exists
7c245b2fe4f1: Layer already exists
f9e18e59a565: Layer already exists
26a504e63be4: Layer already exists
8bf42db0de72: Layer already exists
31892cc314cb: Layer already exists
11936051f93b: Layer already exists
9f8fc49502b5: Pushed
6b450a91a620: Pushed
fe38aa21368f: Pushed
20220109123034: digest: sha256:dc8a416058bbf175f474b30b88e89a77b820271a5c24ced4de2ca97a2972e5de size: 4520

> Task :sdks:java:testing:load-tests:run
Jan 09, 2022 12:32:18 PM org.apache.beam.runners.dataflow.options.DataflowPipelineOptions$StagingLocationFactory create
INFO: No stagingLocation provided, falling back to gcpTempLocation
Jan 09, 2022 12:32:18 PM org.apache.beam.runners.dataflow.DataflowRunner fromOptions
INFO: PipelineOptions.filesToStage was not specified. Defaulting to files from the classpath: will stage 203 files. Enable logging at DEBUG level to see which files will be staged.
Jan 09, 2022 12:32:19 PM org.apache.beam.sdk.Pipeline validate
WARNING: The following transforms do not have stable unique names: Window.Into()
Jan 09, 2022 12:32:19 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Executing pipeline on the Dataflow Service, which will have billing implications related to Google Compute Engine usage and other Google Cloud Services.
Jan 09, 2022 12:32:21 PM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Uploading 203 files from PipelineOptions.filesToStage to staging location to prepare for execution.
Jan 09, 2022 12:32:22 PM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Staging files complete: 203 files cached, 0 files newly uploaded in 0 seconds
Jan 09, 2022 12:32:22 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Staging portable pipeline proto to gs://temp-storage-for-perf-tests/loadtests/staging/
Jan 09, 2022 12:32:22 PM org.apache.beam.runners.dataflow.util.PackageUtil tryStagePackage
INFO: Uploading <113691 bytes, hash 159acf6b4c33624d91700c3a78c2097562e9fd17348f85779bbdb2530ae139ea> to gs://temp-storage-for-perf-tests/loadtests/staging/pipeline-FZrPa0wzYk2RcAw6eMIJdWLp_Rc0j4V3m72yUwrhOeo.pb
Jan 09, 2022 12:32:24 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read input/DataflowRunner.StreamingUnboundedRead.ReadWithIds as step s1
Jan 09, 2022 12:32:24 PM org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource split
INFO: Split into 20 bundles of sizes: [org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@507d64aa, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@37045b48, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@60b34931, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4aa21f9d, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@71c17a57, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@640ab13c, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1e0a864d, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@440e3ce6, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3e67f5f2, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@fd53053, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@4527f70a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@707b1a44, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@7132a9dc, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@57435801, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2da66a44, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@527fc8e, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@61bfc9bf, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@2c7106d9, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@329bad59, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@862f408]
Jan 09, 2022 12:32:24 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read input/StripIds as step s2
Jan 09, 2022 12:32:24 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect start time metrics (input) as step s3
Jan 09, 2022 12:32:24 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Window.Into()/Window.Assign as step s4
Jan 09, 2022 12:32:24 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read co-input/DataflowRunner.StreamingUnboundedRead.ReadWithIds as step s5
Jan 09, 2022 12:32:24 PM org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource split
INFO: Split into 20 bundles of sizes: [org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@d5556bf, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@791c12e3, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@b112b13, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@24eb65e3, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@3ac3f6f, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1abebef3, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@18f55704, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@67cefd84, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5fbe155, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@6add8e3f, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@58a2b917, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@48904d5a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@12bbfc54, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1491344a, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@59b65dce, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@1386313f, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@5e922647, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@433c6abb, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@288f173f, org.apache.beam.sdk.io.synthetic.SyntheticUnboundedSource@a22c4d8]
Jan 09, 2022 12:32:24 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read co-input/StripIds as step s6
Jan 09, 2022 12:32:24 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect start time metrics (co-input) as step s7
Jan 09, 2022 12:32:24 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Window.Into()2/Window.Assign as step s8
Jan 09, 2022 12:32:24 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/MakeUnionTable0 as step s9
Jan 09, 2022 12:32:24 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/MakeUnionTable1 as step s10
Jan 09, 2022 12:32:24 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/Flatten as step s11
Jan 09, 2022 12:32:24 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/GBK as step s12
Jan 09, 2022 12:32:24 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding CoGroupByKey/ConstructCoGbkResultFn as step s13
Jan 09, 2022 12:32:24 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Ungroup and reiterate as step s14
Jan 09, 2022 12:32:24 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect total bytes as step s15
Jan 09, 2022 12:32:24 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect end time metrics as step s16
Jan 09, 2022 12:32:24 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Dataflow SDK version: 2.37.0-SNAPSHOT
Jan 09, 2022 12:32:25 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To access the Dataflow monitoring console, please navigate to https://console.cloud.google.com/dataflow/jobs/us-central1/2022-01-09_04_32_24-11917580719529466718?project=apache-beam-testing
Jan 09, 2022 12:32:25 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Submitted job: 2022-01-09_04_32_24-11917580719529466718
Jan 09, 2022 12:32:25 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To cancel the job using the 'gcloud' tool, run:
> gcloud dataflow jobs --project=apache-beam-testing cancel --region=us-central1 2022-01-09_04_32_24-11917580719529466718
Jan 09, 2022 12:32:31 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
WARNING: 2022-01-09T12:32:31.486Z: The workflow name is not a valid Cloud Label. Labels applied to Cloud resources (such as GCE Instances) for monitoring will be labeled with this modified job name: load0tests0java170dataflow0v20streaming0cogbk01-jenkins-01-fw8g. For the best monitoring experience, please name your job with a valid Cloud Label. For details, see: https://cloud.google.com/compute/docs/labeling-resources#restrictions
Jan 09, 2022 12:32:36 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-09T12:32:35.425Z: Worker configuration: e2-standard-2 in us-central1-f.
Jan 09, 2022 12:32:36 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-09T12:32:35.987Z: Expanding SplittableParDo operations into optimizable parts.
Jan 09, 2022 12:32:36 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-09T12:32:36.008Z: Expanding CollectionToSingleton operations into optimizable parts.
Jan 09, 2022 12:32:36 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-09T12:32:36.077Z: Expanding CoGroupByKey operations into optimizable parts.
Jan 09, 2022 12:32:36 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-09T12:32:36.142Z: Expanding SplittableProcessKeyed operations into optimizable parts.
Jan 09, 2022 12:32:36 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-09T12:32:36.159Z: Expanding GroupByKey operations into streaming Read/Write steps
Jan 09, 2022 12:32:36 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-09T12:32:36.218Z: Lifting ValueCombiningMappingFns into MergeBucketsMappingFns
Jan 09, 2022 12:32:36 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-09T12:32:36.300Z: Fusing adjacent ParDo, Read, Write, and Flatten operations
Jan 09, 2022 12:32:36 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-09T12:32:36.334Z: Unzipping flatten CoGroupByKey-Flatten for input CoGroupByKey-MakeUnionTable0-ParMultiDo-ConstructUnionTable-.output
Jan 09, 2022 12:32:36 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-09T12:32:36.387Z: Fusing unzipped copy of CoGroupByKey/GBK/WriteStream, through flatten CoGroupByKey/Flatten, into producer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable)
Jan 09, 2022 12:32:36 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-09T12:32:36.427Z: Fusing consumer CoGroupByKey/GBK/WriteStream into CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable)
Jan 09, 2022 12:32:36 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-09T12:32:36.481Z: Fusing consumer Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read input/Impulse
Jan 09, 2022 12:32:39 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-09T12:32:36.510Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
Jan 09, 2022 12:32:39 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-09T12:32:36.576Z: Fusing consumer Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
Jan 09, 2022 12:32:39 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-09T12:32:36.612Z: Fusing consumer Read input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
Jan 09, 2022 12:32:39 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-09T12:32:36.633Z: Fusing consumer Collect start time metrics (input)/ParMultiDo(TimeMonitor) into Read input/ParDo(StripIds)/ParMultiDo(StripIds)
Jan 09, 2022 12:32:39 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-09T12:32:36.665Z: Fusing consumer Window.Into()/Window.Assign into Collect start time metrics (input)/ParMultiDo(TimeMonitor)
Jan 09, 2022 12:32:39 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-09T12:32:36.704Z: Fusing consumer CoGroupByKey/MakeUnionTable0/ParMultiDo(ConstructUnionTable) into Window.Into()/Window.Assign
Jan 09, 2022 12:32:39 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-09T12:32:36.732Z: Fusing consumer Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource) into Read co-input/Impulse
Jan 09, 2022 12:32:39 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-09T12:32:36.765Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction into Read co-input/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource)
Jan 09, 2022 12:32:39 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-09T12:32:36.813Z: Fusing consumer Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/SplitWithSizing into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/PairWithRestriction
Jan 09, 2022 12:32:39 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-09T12:32:36.859Z: Fusing consumer Read co-input/ParDo(StripIds)/ParMultiDo(StripIds) into Read-co-input-ParDo-UnboundedSourceAsSDFWrapper--ParMultiDo-UnboundedSourceAsSDFWrapper-/ProcessElementAndRestrictionWithSizing
Jan 09, 2022 12:32:39 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-09T12:32:36.881Z: Fusing consumer Collect start time metrics (co-input)/ParMultiDo(TimeMonitor) into Read co-input/ParDo(StripIds)/ParMultiDo(StripIds)
Jan 09, 2022 12:32:39 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-09T12:32:36.914Z: Fusing consumer Window.Into()2/Window.Assign into Collect start time metrics (co-input)/ParMultiDo(TimeMonitor)
Jan 09, 2022 12:32:39 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-09T12:32:36.938Z: Fusing consumer CoGroupByKey/MakeUnionTable1/ParMultiDo(ConstructUnionTable) into Window.Into()2/Window.Assign
Jan 09, 2022 12:32:39 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-09T12:32:36.971Z: Fusing consumer CoGroupByKey/GBK/MergeBuckets into CoGroupByKey/GBK/ReadStream
Jan 09, 2022 12:32:39 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-09T12:32:37.004Z: Fusing consumer CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult) into CoGroupByKey/GBK/MergeBuckets
Jan 09, 2022 12:32:39 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-09T12:32:37.027Z: Fusing consumer Ungroup and reiterate/ParMultiDo(UngroupAndReiterate) into CoGroupByKey/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult)
Jan 09, 2022 12:32:39 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-09T12:32:37.058Z: Fusing consumer Collect total bytes/ParMultiDo(ByteMonitor) into Ungroup and reiterate/ParMultiDo(UngroupAndReiterate)
Jan 09, 2022 12:32:39 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-09T12:32:37.094Z: Fusing consumer Collect end time metrics/ParMultiDo(TimeMonitor) into Collect total bytes/ParMultiDo(ByteMonitor)
Jan 09, 2022 12:32:39 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-09T12:32:37.416Z: Starting 5 ****s in us-central1-f...
Jan 09, 2022 12:32:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-09T12:32:42.039Z: Your project already contains 100 Dataflow-created metric descriptors, so new user metrics of the form custom.googleapis.com/* will not be created. However, all user metrics are also available in the metric dataflow.googleapis.com/job/user_counter. If you rely on the custom metrics, you can delete old / unused metric descriptors. See https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.list and https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.delete
Jan 09, 2022 12:33:26 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-09T12:33:26.501Z: Autoscaling: Raised the number of ****s to 5 so that the pipeline can catch up with its backlog and keep up with its input rate.
Jan 09, 2022 12:34:19 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-09T12:34:18.854Z: Workers have started successfully.
Jan 09, 2022 12:34:19 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-09T12:34:18.894Z: Workers have started successfully.
Jan 09, 2022 4:00:39 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-09T16:00:37.328Z: Cancel request is committed for workflow job: 2022-01-09_04_32_24-11917580719529466718.
Jan 09, 2022 4:00:39 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-09T16:00:37.380Z: Cleaning up.
Jan 09, 2022 4:00:39 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-09T16:00:37.664Z: Stopping **** pool...
Jan 09, 2022 4:00:39 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-09T16:00:37.740Z: Stopping **** pool...
Jan 09, 2022 4:03:01 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-09T16:03:00.582Z: Autoscaling: Reduced the number of ****s to 0 based on low average **** CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate.
Jan 09, 2022 4:03:01 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2022-01-09T16:03:00.622Z: Worker pool stopped.
Jan 09, 2022 4:03:08 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2022-01-09_04_32_24-11917580719529466718 finished with status CANCELLED.
Load test results for test (ID): 18e201bc-459a-498b-80f0-5d6ef874f545 and timestamp: 2022-01-09T12:32:19.272000000Z:
                 Metric:                    Value:
dataflow_v2_java17_runtime_sec                 12322.264
dataflow_v2_java17_total_bytes_count             2.95557406E10
Exception in thread "main" java.lang.RuntimeException: Invalid job state: CANCELLED.
	at org.apache.beam.sdk.loadtests.JobFailure.handleFailure(JobFailure.java:51)
	at org.apache.beam.sdk.loadtests.LoadTest.run(LoadTest.java:139)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.run(CoGroupByKeyLoadTest.java:62)
	at org.apache.beam.sdk.loadtests.CoGroupByKeyLoadTest.main(CoGroupByKeyLoadTest.java:157)

> Task :sdks:java:testing:load-tests:run FAILED

> Task :runners:google-cloud-dataflow-java:cleanUpDockerImages
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220109123034
Untagged: us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:dc8a416058bbf175f474b30b88e89a77b820271a5c24ced4de2ca97a2972e5de
Deleted: sha256:238f8d2ec862a157768e1d900952de2b9bfe470d95cba54e0144824bfad8b837
Deleted: sha256:fe7e1623ea9fe262cc5b981c6e37d8be6ec12e39f47cd319b90d5300a83ab415
Deleted: sha256:da0a3ba28d1d21683f2c114b37265b139a28cba81ba60cecaf539cca193f61af
Deleted: sha256:e0da82a172952d0f1d937c8013e68ab29327ec38bd80c31777a705deaa36d942
Deleted: sha256:9ac8c5eac2dd073bab86ee0dedeb20b58df5cdf558ed998542b9320e67359b5b
Deleted: sha256:0035d7db3adea533b042cae698311323a143cc2960f9598055d0e24c1e3fd228
Deleted: sha256:aca5826c5f074028497cb3007a535aa43eac103606d17f2ac50480884bad2721
Deleted: sha256:73deeb0751ac7652fea232c87cfa28e759a85dc84cc9529261991e903276683c
Deleted: sha256:6bbbcfdc0fbbbf247c410fef9c9b1f4727336a303799afd9255ec7835b91b37d
Deleted: sha256:b335f65608380dd37775cc3d54bd9c7f95912bff04facbf418f9459ef1f7f1da
Deleted: sha256:3358ebb786359b7b941fbc0b7b22230bcb8c832ce462e217161df4dff3fa8cfb
Deleted: sha256:58a21325125737f563e33f4b246740023195b232ae7f0c453119fe764aac31b0
Deleted: sha256:f130c76e1192f40e955162029101013e875c7a2cb296a96504486d3aa16e0087
Deleted: sha256:639628655536510d99a436009d48dc39fa1196e8f82c0fe4dd72169d962f9b77
Deleted: sha256:d73ee7dc43426c970345e556789d0312d2fd26b55bd025566a78a8fde54e7f2e
Deleted: sha256:b72e816ed35c723e415e9d19229345e6c755fb4dcf557e638c9ce60e4238d125
Deleted: sha256:7d8f9f6db5c41e170c9fac7c1626fd524709cbe9b69d0d51a9c6c79027754028
Deleted: sha256:c4889ccb5a4f57cf7cb29c06d8b6d18f0c4830cd60e0a7b94fe20fab3c8ab151
Deleted: sha256:491de74ad05ee94c11fac0ca450c2be388920b21dab2ebd60acfbdc041352282
Deleted: sha256:ec1ee9b1c191e34f2cdd0ab087cbf7ee085f05283b9f6308e890c129572c4b2d
Deleted: sha256:c75b363ad59535a8f9f1dee8b663976cc5aea17b069a62f7a25ca839380b4839
Deleted: sha256:8471e84e6b7f4b796417ba7ee413bc613327b52b61e5b462150d732c9a56f526
Deleted: sha256:5561031081ef7674bf4daaeb32c499978f850d412a59d02494dac60a46324e00
Deleted: sha256:8852115ab74b07a7c50d98b4f84a3b2d3e920ea67f224880bfc7e31fefb2206f
Deleted: sha256:5c69ebaa5e0449162650bab7dacaef8042e6c552591397402b846e1b0551ef93
Deleted: sha256:f554783bbd07ebc3ae037f658773ab3718830017beea2bdf93f99c94c6f61b37
Deleted: sha256:35677d095624e2000e3ca074d7b38df262760d15203c69ed3efa03683c35e0b4
Deleted: sha256:42c0385dfd80b1efbd9399c1e50ef9e5bd69830ad2dbdf533388629b32020b9e
Deleted: sha256:189977f28854373c2953fb405cc3e2b09e05dc6d16972cc8365fca3b33cf648b
Deleted: sha256:f381d93f970160384ae8269a03e7b853e5c8af76bb4e60b9aa9ade02f4fee40f
Tag: [us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220109123034]
- referencing digest: [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:dc8a416058bbf175f474b30b88e89a77b820271a5c24ced4de2ca97a2972e5de]

Deleted [[us.gcr.io/apache-beam-testing/java-postcommit-it/java:20220109123034] (referencing [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:dc8a416058bbf175f474b30b88e89a77b820271a5c24ced4de2ca97a2972e5de])].
Removing untagged image us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:dc8a416058bbf175f474b30b88e89a77b820271a5c24ced4de2ca97a2972e5de
Digests:
- us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:dc8a416058bbf175f474b30b88e89a77b820271a5c24ced4de2ca97a2972e5de
Deleted [us.gcr.io/apache-beam-testing/java-postcommit-it/java@sha256:dc8a416058bbf175f474b30b88e89a77b820271a5c24ced4de2ca97a2972e5de].

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:testing:load-tests:run'.
> Process 'command '/usr/lib/jvm/java-8-openjdk-amd64/bin/java'' finished with non-zero exit value 1

* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
> Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0.

You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins.

See https://docs.gradle.org/7.3.2/userguide/command_line_interface.html#sec:command_line_warnings

Execution optimizations have been disabled for 1 invalid unit(s) of work during this build to ensure correctness.
Please consult deprecation warnings for more details.

BUILD FAILED in 3h 32m 51s
109 actionable tasks: 72 executed, 33 from cache, 4 up-to-date

Publishing build scan...
https://gradle.com/s/zrxzxmy7rkg7m

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org