You are viewing a plain text version of this content. The canonical link for it is here.
Posted to builds@beam.apache.org by Apache Jenkins Server <je...@builds.apache.org> on 2020/12/17 14:13:18 UTC
beam_BiqQueryIO_Streaming_Performance_Test_Java - Build # 1752 -
Aborted!
beam_BiqQueryIO_Streaming_Performance_Test_Java - Build # 1752 - Aborted:
Check console output at https://ci-beam.apache.org/job/beam_BiqQueryIO_Streaming_Performance_Test_Java/1752/ to view the results.
Jenkins build is back to normal :
beam_BiqQueryIO_Streaming_Performance_Test_Java #1755
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_BiqQueryIO_Streaming_Performance_Test_Java/1755/display/redirect?page=changes>
---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org
Build failed in Jenkins:
beam_BiqQueryIO_Streaming_Performance_Test_Java #1754
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_BiqQueryIO_Streaming_Performance_Test_Java/1754/display/redirect?page=changes>
Changes:
[Pablo Estrada] Cleanup directory with SVN repo for PMC members
[Pablo Estrada] Adding combiner packing optimization for Portable Runner
[Pablo Estrada] fix formatting
[Udi Meiri] [BEAM-10658] Update website to recommend the right version
[noreply] Rewrite container docs for custom containers (#13420)
[noreply] [BEAM-11458] Upgrade SamzRunner to Samza 1.5 (#13550)
------------------------------------------
[...truncated 276.34 KB...]
Build cache key for task ':runners:google-cloud-dataflow-java:compileTestJava' is 88d824e0765aa77406221f2a79fb5092
Task ':runners:google-cloud-dataflow-java:compileTestJava' is not up-to-date because:
No history is available.
Watching 1358 directories to track changes
Watching 1358 directories to track changes
Watching 1358 directories to track changes
Watching 1369 directories to track changes
Watching 1370 directories to track changes
Watching 1371 directories to track changes
Loaded cache entry for task ':runners:google-cloud-dataflow-java:compileTestJava' with cache key 88d824e0765aa77406221f2a79fb5092
:runners:google-cloud-dataflow-java:compileTestJava (Thread[Execution **** for ':' Thread 5,5,main]) completed. Took 0.211 secs.
:runners:google-cloud-dataflow-java:testClasses (Thread[Execution **** for ':' Thread 5,5,main]) started.
> Task :runners:google-cloud-dataflow-java:testClasses UP-TO-DATE
Skipping task ':runners:google-cloud-dataflow-java:testClasses' as it has no actions.
:runners:google-cloud-dataflow-java:testClasses (Thread[Execution **** for ':' Thread 5,5,main]) completed. Took 0.0 secs.
:runners:google-cloud-dataflow-java:testJar (Thread[Execution **** for ':' Thread 5,5,main]) started.
> Task :runners:google-cloud-dataflow-java:testJar
Watching 1371 directories to track changes
Could not read file path '<https://ci-beam.apache.org/job/beam_BiqQueryIO_Streaming_Performance_Test_Java/ws/src/runners/google-cloud-dataflow-java/build/resources/test'.>
Watching 1371 directories to track changes
Watching 1372 directories to track changes
Caching disabled for task ':runners:google-cloud-dataflow-java:testJar' because:
Caching has not been enabled for the task
Task ':runners:google-cloud-dataflow-java:testJar' is not up-to-date because:
No history is available.
Watching 1372 directories to track changes
file or directory '<https://ci-beam.apache.org/job/beam_BiqQueryIO_Streaming_Performance_Test_Java/ws/src/runners/google-cloud-dataflow-java/build/resources/test',> not found
Watching 1372 directories to track changes
:runners:google-cloud-dataflow-java:testJar (Thread[Execution **** for ':' Thread 5,5,main]) completed. Took 0.036 secs.
:sdks:java:io:bigquery-io-perf-tests:integrationTest (Thread[Execution **** for ':' Thread 5,5,main]) started.
Gradle Test Executor 1 started executing tests.
> Task :sdks:java:io:bigquery-io-perf-tests:integrationTest
Watching 1372 directories to track changes
Watching 1372 directories to track changes
Watching 1372 directories to track changes
Watching 1372 directories to track changes
Watching 1372 directories to track changes
Custom actions are attached to task ':sdks:java:io:bigquery-io-perf-tests:integrationTest'.
Build cache key for task ':sdks:java:io:bigquery-io-perf-tests:integrationTest' is c0a38310f92a667436ad308b94fd3e01
Task ':sdks:java:io:bigquery-io-perf-tests:integrationTest' is not up-to-date because:
Task.upToDateWhen is false.
Watching 1372 directories to track changes
Watching 1372 directories to track changes
Starting process 'Gradle Test Executor 1'. Working directory: <https://ci-beam.apache.org/job/beam_BiqQueryIO_Streaming_Performance_Test_Java/ws/src/sdks/java/io/bigquery-io-perf-tests> Command: /usr/lib/jvm/java-8-openjdk-amd64/bin/java -DbeamTestPipelineOptions=["--project=apache-beam-testing","--tempLocation=gs://temp-storage-for-perf-tests/loadtests","--tempRoot=gs://temp-storage-for-perf-tests/loadtests","--writeMethod=STREAMING_INSERTS","--writeFormat=JSON","--testBigQueryDataset=beam_performance","--testBigQueryTable=bqio_write_10GB_java_stream_1218000641","--metricsBigQueryDataset=beam_performance","--metricsBigQueryTable=bqio_10GB_results_java_stream","--influxMeasurement=bqio_10GB_results_java_stream","--sourceOptions={\"numRecords\":\"10485760\",\"keySizeBytes\":\"1\",\"valueSizeBytes\":\"1024\"}","--runner=DataflowRunner","--maxNumWorkers=5","--numWorkers=5","--autoscalingAlgorithm=NONE","--influxDatabase=beam_test_metrics","--influxHost=http://10.128.0.96:8086","--****HarnessContainerImage=","--dataflowWorkerJar=<https://ci-beam.apache.org/job/beam_BiqQueryIO_Streaming_Performance_Test_Java/ws/src/runners/google-cloud-dataflow-java/****/legacy-****/build/libs/beam-runners-google-cloud-dataflow-java-legacy-****-2.28.0-SNAPSHOT.jar","--region=us-central1"]> -Djava.security.manager=****.org.gradle.process.internal.****.child.BootstrapSecurityManager -Dorg.gradle.native=false -Xmx2g -Dfile.encoding=UTF-8 -Duser.country=US -Duser.language=en -Duser.variant -ea -cp /home/jenkins/.gradle/caches/6.7/****Main/gradle-****.jar ****.org.gradle.process.internal.****.GradleWorkerMain 'Gradle Test Executor 1'
Successfully started process 'Gradle Test Executor 1'
org.apache.beam.sdk.bigqueryioperftests.BigQueryIOIT STANDARD_ERROR
SLF4J: Class path contains multiple SLF4J bindings.
SLF4J: Found binding in [jar:<https://ci-beam.apache.org/job/beam_BiqQueryIO_Streaming_Performance_Test_Java/ws/src/runners/google-cloud-dataflow-java/****/legacy-****/build/libs/beam-runners-google-cloud-dataflow-java-legacy-****-2.28.0-SNAPSHOT.jar!/org/slf4j/impl/StaticLoggerBinder.class]>
SLF4J: Found binding in [jar:file:/home/jenkins/.gradle/caches/modules-2/files-2.1/org.slf4j/slf4j-jdk14/1.7.30/d35953dd2fe54ebe39fdf18cfd82fe6eb35b25ed/slf4j-jdk14-1.7.30.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation.
SLF4J: Actual binding is of type [org.slf4j.impl.JDK14LoggerFactory]
org.apache.beam.sdk.bigqueryioperftests.BigQueryIOIT > testWriteThenRead STANDARD_ERROR
Dec 18, 2020 12:33:43 AM org.apache.beam.runners.dataflow.options.DataflowPipelineOptions$StagingLocationFactory create
INFO: No stagingLocation provided, falling back to gcpTempLocation
Dec 18, 2020 12:33:44 AM org.apache.beam.runners.dataflow.DataflowRunner fromOptions
INFO: PipelineOptions.filesToStage was not specified. Defaulting to files from the classpath: will stage 210 files. Enable logging at DEBUG level to see which files will be staged.
Dec 18, 2020 12:33:45 AM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Executing pipeline on the Dataflow Service, which will have billing implications related to Google Compute Engine usage and other Google Cloud Services.
Dec 18, 2020 12:33:47 AM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Uploading 211 files from PipelineOptions.filesToStage to staging location to prepare for execution.
Dec 18, 2020 12:33:48 AM org.apache.beam.runners.dataflow.util.PackageUtil$PackageAttributes forFileToStage
INFO: Staging custom dataflow-****.jar as beam-runners-google-cloud-dataflow-java-legacy-****-2.28.0-SNAPSHOT-SY51_ntnFm08YbRxJRsH0MUwRdL3yw8xnIg9nfCgmLo.jar
Dec 18, 2020 12:33:48 AM org.apache.beam.runners.dataflow.util.PackageUtil tryStagePackage
INFO: Uploading <https://ci-beam.apache.org/job/beam_BiqQueryIO_Streaming_Performance_Test_Java/ws/src/sdks/java/core/build/libs/beam-sdks-java-core-2.28.0-SNAPSHOT-unshaded.jar> to gs://temp-storage-for-perf-tests/loadtests/staging/beam-sdks-java-core-2.28.0-SNAPSHOT-unshaded-MO8rLDMgw2_59BHguyxx5fiOg26hO7Uli4fJzOLgjYs.jar
Dec 18, 2020 12:33:48 AM org.apache.beam.runners.dataflow.util.PackageUtil tryStagePackage
INFO: Uploading /tmp/test4290360351359473335.zip to gs://temp-storage-for-perf-tests/loadtests/staging/test-VvVWc48F8h_nPeUPRdGNaqpM8yUazC9wJ-_SOw-Cq4M.jar
Dec 18, 2020 12:33:48 AM org.apache.beam.runners.dataflow.util.PackageUtil tryStagePackage
INFO: Uploading <https://ci-beam.apache.org/job/beam_BiqQueryIO_Streaming_Performance_Test_Java/ws/src/sdks/java/core/build/libs/beam-sdks-java-core-2.28.0-SNAPSHOT.jar> to gs://temp-storage-for-perf-tests/loadtests/staging/beam-sdks-java-core-2.28.0-SNAPSHOT-uoWIWf0suN938J4VvkuuBJZdafdeOmc32dr1cK29daA.jar
Dec 18, 2020 12:33:49 AM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Staging files complete: 208 files cached, 3 files newly uploaded in 1 seconds
Dec 18, 2020 12:33:49 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read from source as step s1
Dec 18, 2020 12:33:49 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Gather time as step s2
Dec 18, 2020 12:33:49 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Map records as step s3
Dec 18, 2020 12:33:49 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Write to BQ/PrepareWrite/ParDo(Anonymous) as step s4
Dec 18, 2020 12:33:49 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Write to BQ/StreamingInserts/CreateTables/ParDo(CreateTables) as step s5
Dec 18, 2020 12:33:49 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Write to BQ/StreamingInserts/StreamingWriteTables/ShardTableWrites as step s6
Dec 18, 2020 12:33:49 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Write to BQ/StreamingInserts/StreamingWriteTables/TagWithUniqueIds as step s7
Dec 18, 2020 12:33:49 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Write to BQ/StreamingInserts/StreamingWriteTables/Reshuffle/Window.Into()/Window.Assign as step s8
Dec 18, 2020 12:33:49 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Write to BQ/StreamingInserts/StreamingWriteTables/Reshuffle/GroupByKey as step s9
Dec 18, 2020 12:33:49 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Write to BQ/StreamingInserts/StreamingWriteTables/Reshuffle/ExpandIterable as step s10
Dec 18, 2020 12:33:49 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Write to BQ/StreamingInserts/StreamingWriteTables/GlobalWindow/Window.Assign as step s11
Dec 18, 2020 12:33:49 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Write to BQ/StreamingInserts/StreamingWriteTables/StreamingWrite as step s12
Dec 18, 2020 12:33:49 AM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Staging pipeline description to gs://temp-storage-for-perf-tests/loadtests/staging/
Dec 18, 2020 12:33:49 AM org.apache.beam.runners.dataflow.util.PackageUtil tryStagePackage
INFO: Uploading <109326 bytes, hash 8b8300ddc458151f9e86e60825a769fdf47a9d63e23061b7de01e09b608f04ff> to gs://temp-storage-for-perf-tests/loadtests/staging/pipeline-i4MA3cRYFR-ehuYIJadp_fR6nWPiMGG33gHgm2CPBP8.pb
Dec 18, 2020 12:33:50 AM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Dataflow SDK version: 2.28.0-SNAPSHOT
Dec 18, 2020 12:33:51 AM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To access the Dataflow monitoring console, please navigate to https://console.cloud.google.com/dataflow/jobs/us-central1/2020-12-17_16_33_50-10517056353046153571?project=apache-beam-testing
Dec 18, 2020 12:33:51 AM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Submitted job: 2020-12-17_16_33_50-10517056353046153571
Dec 18, 2020 12:33:51 AM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To cancel the job using the 'gcloud' tool, run:
> gcloud dataflow jobs --project=apache-beam-testing cancel --region=us-central1 2020-12-17_16_33_50-10517056353046153571
Dec 18, 2020 12:33:52 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
WARNING: 2020-12-18T00:33:50.132Z: The requested max number of ****s (5) is ignored as autoscaling is explicitly disabled (autoscalingAlgorithm=NONE).
Dec 18, 2020 12:34:02 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2020-12-18T00:34:01.168Z: Worker configuration: n1-standard-1 in us-central1-f.
Dec 18, 2020 12:34:02 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2020-12-18T00:34:02.049Z: Expanding CoGroupByKey operations into optimizable parts.
Dec 18, 2020 12:34:02 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2020-12-18T00:34:02.133Z: Expanding GroupByKey operations into optimizable parts.
Dec 18, 2020 12:34:02 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2020-12-18T00:34:02.168Z: Lifting ValueCombiningMappingFns into MergeBucketsMappingFns
Dec 18, 2020 12:34:02 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2020-12-18T00:34:02.300Z: Fusing adjacent ParDo, Read, Write, and Flatten operations
Dec 18, 2020 12:34:02 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2020-12-18T00:34:02.329Z: Fusing consumer Gather time into Read from source
Dec 18, 2020 12:34:02 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2020-12-18T00:34:02.354Z: Fusing consumer Map records into Gather time
Dec 18, 2020 12:34:02 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2020-12-18T00:34:02.393Z: Fusing consumer Write to BQ/PrepareWrite/ParDo(Anonymous) into Map records
Dec 18, 2020 12:34:02 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2020-12-18T00:34:02.420Z: Fusing consumer Write to BQ/StreamingInserts/CreateTables/ParDo(CreateTables) into Write to BQ/PrepareWrite/ParDo(Anonymous)
Dec 18, 2020 12:34:02 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2020-12-18T00:34:02.446Z: Fusing consumer Write to BQ/StreamingInserts/StreamingWriteTables/ShardTableWrites into Write to BQ/StreamingInserts/CreateTables/ParDo(CreateTables)
Dec 18, 2020 12:34:02 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2020-12-18T00:34:02.482Z: Fusing consumer Write to BQ/StreamingInserts/StreamingWriteTables/TagWithUniqueIds into Write to BQ/StreamingInserts/StreamingWriteTables/ShardTableWrites
Dec 18, 2020 12:34:02 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2020-12-18T00:34:02.507Z: Fusing consumer Write to BQ/StreamingInserts/StreamingWriteTables/Reshuffle/Window.Into()/Window.Assign into Write to BQ/StreamingInserts/StreamingWriteTables/TagWithUniqueIds
Dec 18, 2020 12:34:02 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2020-12-18T00:34:02.540Z: Fusing consumer Write to BQ/StreamingInserts/StreamingWriteTables/Reshuffle/GroupByKey/Reify into Write to BQ/StreamingInserts/StreamingWriteTables/Reshuffle/Window.Into()/Window.Assign
Dec 18, 2020 12:34:02 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2020-12-18T00:34:02.593Z: Fusing consumer Write to BQ/StreamingInserts/StreamingWriteTables/Reshuffle/GroupByKey/Write into Write to BQ/StreamingInserts/StreamingWriteTables/Reshuffle/GroupByKey/Reify
Dec 18, 2020 12:34:02 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2020-12-18T00:34:02.619Z: Fusing consumer Write to BQ/StreamingInserts/StreamingWriteTables/Reshuffle/GroupByKey/GroupByWindow into Write to BQ/StreamingInserts/StreamingWriteTables/Reshuffle/GroupByKey/Read
Dec 18, 2020 12:34:02 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2020-12-18T00:34:02.663Z: Fusing consumer Write to BQ/StreamingInserts/StreamingWriteTables/Reshuffle/ExpandIterable into Write to BQ/StreamingInserts/StreamingWriteTables/Reshuffle/GroupByKey/GroupByWindow
Dec 18, 2020 12:34:02 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2020-12-18T00:34:02.694Z: Fusing consumer Write to BQ/StreamingInserts/StreamingWriteTables/GlobalWindow/Window.Assign into Write to BQ/StreamingInserts/StreamingWriteTables/Reshuffle/ExpandIterable
Dec 18, 2020 12:34:02 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2020-12-18T00:34:02.724Z: Fusing consumer Write to BQ/StreamingInserts/StreamingWriteTables/StreamingWrite into Write to BQ/StreamingInserts/StreamingWriteTables/GlobalWindow/Window.Assign
Dec 18, 2020 12:34:05 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2020-12-18T00:34:03.109Z: Executing operation Write to BQ/StreamingInserts/StreamingWriteTables/Reshuffle/GroupByKey/Create
Dec 18, 2020 12:34:05 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2020-12-18T00:34:03.192Z: Starting 5 ****s in us-central1-f...
Dec 18, 2020 12:34:05 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2020-12-18T00:34:03.269Z: Finished operation Write to BQ/StreamingInserts/StreamingWriteTables/Reshuffle/GroupByKey/Create
Dec 18, 2020 12:34:05 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2020-12-18T00:34:03.423Z: Executing operation Read from source+Gather time+Map records+Write to BQ/PrepareWrite/ParDo(Anonymous)+Write to BQ/StreamingInserts/CreateTables/ParDo(CreateTables)+Write to BQ/StreamingInserts/StreamingWriteTables/ShardTableWrites+Write to BQ/StreamingInserts/StreamingWriteTables/TagWithUniqueIds+Write to BQ/StreamingInserts/StreamingWriteTables/Reshuffle/Window.Into()/Window.Assign+Write to BQ/StreamingInserts/StreamingWriteTables/Reshuffle/GroupByKey/Reify+Write to BQ/StreamingInserts/StreamingWriteTables/Reshuffle/GroupByKey/Write
Dec 18, 2020 12:34:32 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2020-12-18T00:34:31.581Z: Autoscaling: Raised the number of ****s to 2 based on the rate of progress in the currently running stage(s).
Dec 18, 2020 12:34:32 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2020-12-18T00:34:31.685Z: Resized **** pool to 2, though goal was 5. This could be a quota issue.
Dec 18, 2020 12:34:32 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2020-12-18T00:34:31.921Z: Your project already contains 100 Dataflow-created metric descriptors, so new user metrics of the form custom.googleapis.com/* will not be created. However, all user metrics are also available in the metric dataflow.googleapis.com/job/user_counter. If you rely on the custom metrics, you can delete old / unused metric descriptors. See https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.list and https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.delete
Dec 18, 2020 12:34:42 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2020-12-18T00:34:41.970Z: Autoscaling: Raised the number of ****s to 5 based on the rate of progress in the currently running stage(s).
Dec 18, 2020 12:34:55 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2020-12-18T00:34:54.923Z: Workers have started successfully.
Dec 18, 2020 12:34:55 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2020-12-18T00:34:54.956Z: Workers have started successfully.
Dec 18, 2020 12:37:06 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2020-12-18T00:37:06.016Z: Finished operation Read from source+Gather time+Map records+Write to BQ/PrepareWrite/ParDo(Anonymous)+Write to BQ/StreamingInserts/CreateTables/ParDo(CreateTables)+Write to BQ/StreamingInserts/StreamingWriteTables/ShardTableWrites+Write to BQ/StreamingInserts/StreamingWriteTables/TagWithUniqueIds+Write to BQ/StreamingInserts/StreamingWriteTables/Reshuffle/Window.Into()/Window.Assign+Write to BQ/StreamingInserts/StreamingWriteTables/Reshuffle/GroupByKey/Reify+Write to BQ/StreamingInserts/StreamingWriteTables/Reshuffle/GroupByKey/Write
Dec 18, 2020 12:37:06 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2020-12-18T00:37:06.088Z: Executing operation Write to BQ/StreamingInserts/StreamingWriteTables/Reshuffle/GroupByKey/Close
Dec 18, 2020 12:37:06 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2020-12-18T00:37:06.155Z: Finished operation Write to BQ/StreamingInserts/StreamingWriteTables/Reshuffle/GroupByKey/Close
Dec 18, 2020 12:37:06 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2020-12-18T00:37:06.254Z: Executing operation Write to BQ/StreamingInserts/StreamingWriteTables/Reshuffle/GroupByKey/Read+Write to BQ/StreamingInserts/StreamingWriteTables/Reshuffle/GroupByKey/GroupByWindow+Write to BQ/StreamingInserts/StreamingWriteTables/Reshuffle/ExpandIterable+Write to BQ/StreamingInserts/StreamingWriteTables/GlobalWindow/Window.Assign+Write to BQ/StreamingInserts/StreamingWriteTables/StreamingWrite
org.apache.beam.sdk.bigqueryioperftests.BigQueryIOIT > testWriteThenRead SKIPPED
Watching 1374 directories to track changes
Watching 1376 directories to track changes
Watching 1377 directories to track changes
> Task :sdks:java:io:bigquery-io-perf-tests:integrationTest FAILED
:sdks:java:io:bigquery-io-perf-tests:integrationTest (Thread[Execution **** for ':' Thread 5,5,main]) completed. Took 8 mins 33.12 secs.
FAILURE: Build failed with an exception.
* What went wrong:
Execution failed for task ':sdks:java:io:bigquery-io-perf-tests:integrationTest'.
> Process 'Gradle Test Executor 1' finished with non-zero exit value 143
This problem might be caused by incorrect test process configuration.
Please refer to the test execution section in the User Manual at https://docs.gradle.org/6.7/userguide/java_testing.html#sec:test_execution
* Try:
Run with --stacktrace option to get the stack trace. Run with --debug option to get more log output. Run with --scan to get full insights.
* Get more help at https://help.gradle.org
Deprecated Gradle features were used in this build, making it incompatible with Gradle 7.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/6.7/userguide/command_line_interface.html#sec:command_line_warnings
BUILD FAILED in 8m 54s
109 actionable tasks: 66 executed, 43 from cache
Watching 1377 directories to track changes
Publishing build scan...
The message received from the daemon indicates that the daemon has disappeared.
Build request sent: Build{id=2740741a-7942-4349-915c-7c0aa04a15ae, currentDir=<https://ci-beam.apache.org/job/beam_BiqQueryIO_Streaming_Performance_Test_Java/ws/src}>
Attempting to read last messages from the daemon log...
Daemon pid: 31225
log file: /home/jenkins/.gradle/daemon/6.7/daemon-31225.out.log
----- Last 20 lines from daemon log file - daemon-31225.out.log -----
Execution failed for task ':sdks:java:io:bigquery-io-perf-tests:integrationTest'.
> Process 'Gradle Test Executor 1' finished with non-zero exit value 143
This problem might be caused by incorrect test process configuration.
Please refer to the test execution section in the User Manual at https://docs.gradle.org/6.7/userguide/java_testing.html#sec:test_execution
* Try:
Run with --stacktrace option to get the stack trace. Run with --debug option to get more log output. Run with --scan to get full insights.
* Get more help at https://help.gradle.org
Deprecated Gradle features were used in this build, making it incompatible with Gradle 7.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/6.7/userguide/command_line_interface.html#sec:command_line_warnings
BUILD FAILED in 8m 54s
109 actionable tasks: 66 executed, 43 from cache
Watching 1377 directories to track changes
Publishing build scan...
Daemon vm is shutting down... The daemon has exited normally or was terminated in response to a user interrupt.
----- End of the daemon log -----
FAILURE: Build failed with an exception.
* What went wrong:
Gradle build daemon disappeared unexpectedly (it may have been killed or may have crashed)
* Try:
Run with --stacktrace option to get the stack trace. Run with --debug option to get more log output. Run with --scan to get full insights.
* Get more help at https://help.gradle.org
Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure
---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org
Build failed in Jenkins:
beam_BiqQueryIO_Streaming_Performance_Test_Java #1753
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_BiqQueryIO_Streaming_Performance_Test_Java/1753/display/redirect?page=changes>
Changes:
[noreply] [Beam-11002] Fixes BufferOverflowException in XMLReader (#13513)
------------------------------------------
[...truncated 276.54 KB...]
Task ':runners:google-cloud-dataflow-java:compileTestJava' is not up-to-date because:
No history is available.
Watching 1355 directories to track changes
Watching 1355 directories to track changes
Watching 1355 directories to track changes
Watching 1366 directories to track changes
Watching 1367 directories to track changes
Watching 1368 directories to track changes
Loaded cache entry for task ':runners:google-cloud-dataflow-java:compileTestJava' with cache key 88d824e0765aa77406221f2a79fb5092
:runners:google-cloud-dataflow-java:compileTestJava (Thread[Daemon **** Thread 2,5,main]) completed. Took 0.19 secs.
:runners:google-cloud-dataflow-java:testClasses (Thread[Daemon **** Thread 2,5,main]) started.
> Task :runners:google-cloud-dataflow-java:testClasses UP-TO-DATE
Skipping task ':runners:google-cloud-dataflow-java:testClasses' as it has no actions.
:runners:google-cloud-dataflow-java:testClasses (Thread[Daemon **** Thread 2,5,main]) completed. Took 0.0 secs.
:runners:google-cloud-dataflow-java:testJar (Thread[Execution **** for ':' Thread 8,5,main]) started.
> Task :runners:google-cloud-dataflow-java:testJar
Watching 1368 directories to track changes
Could not read file path '<https://ci-beam.apache.org/job/beam_BiqQueryIO_Streaming_Performance_Test_Java/ws/src/runners/google-cloud-dataflow-java/build/resources/test'.>
Watching 1368 directories to track changes
Watching 1369 directories to track changes
Caching disabled for task ':runners:google-cloud-dataflow-java:testJar' because:
Caching has not been enabled for the task
Task ':runners:google-cloud-dataflow-java:testJar' is not up-to-date because:
No history is available.
Watching 1369 directories to track changes
file or directory '<https://ci-beam.apache.org/job/beam_BiqQueryIO_Streaming_Performance_Test_Java/ws/src/runners/google-cloud-dataflow-java/build/resources/test',> not found
Watching 1369 directories to track changes
:runners:google-cloud-dataflow-java:testJar (Thread[Execution **** for ':' Thread 8,5,main]) completed. Took 0.035 secs.
:sdks:java:io:bigquery-io-perf-tests:integrationTest (Thread[Execution **** for ':' Thread 8,5,main]) started.
Gradle Test Executor 1 started executing tests.
> Task :sdks:java:io:bigquery-io-perf-tests:integrationTest
Watching 1369 directories to track changes
Watching 1369 directories to track changes
Watching 1369 directories to track changes
Watching 1369 directories to track changes
Watching 1369 directories to track changes
Custom actions are attached to task ':sdks:java:io:bigquery-io-perf-tests:integrationTest'.
Build cache key for task ':sdks:java:io:bigquery-io-perf-tests:integrationTest' is c0a38310f92a667436ad308b94fd3e01
Task ':sdks:java:io:bigquery-io-perf-tests:integrationTest' is not up-to-date because:
Task.upToDateWhen is false.
Watching 1369 directories to track changes
Watching 1369 directories to track changes
Starting process 'Gradle Test Executor 1'. Working directory: <https://ci-beam.apache.org/job/beam_BiqQueryIO_Streaming_Performance_Test_Java/ws/src/sdks/java/io/bigquery-io-perf-tests> Command: /usr/lib/jvm/java-8-openjdk-amd64/bin/java -DbeamTestPipelineOptions=["--project=apache-beam-testing","--tempLocation=gs://temp-storage-for-perf-tests/loadtests","--tempRoot=gs://temp-storage-for-perf-tests/loadtests","--writeMethod=STREAMING_INSERTS","--writeFormat=JSON","--testBigQueryDataset=beam_performance","--testBigQueryTable=bqio_write_10GB_java_stream_1217150414","--metricsBigQueryDataset=beam_performance","--metricsBigQueryTable=bqio_10GB_results_java_stream","--influxMeasurement=bqio_10GB_results_java_stream","--sourceOptions={\"numRecords\":\"10485760\",\"keySizeBytes\":\"1\",\"valueSizeBytes\":\"1024\"}","--runner=DataflowRunner","--maxNumWorkers=5","--numWorkers=5","--autoscalingAlgorithm=NONE","--influxDatabase=beam_test_metrics","--influxHost=http://10.128.0.96:8086","--****HarnessContainerImage=","--dataflowWorkerJar=<https://ci-beam.apache.org/job/beam_BiqQueryIO_Streaming_Performance_Test_Java/ws/src/runners/google-cloud-dataflow-java/****/legacy-****/build/libs/beam-runners-google-cloud-dataflow-java-legacy-****-2.28.0-SNAPSHOT.jar","--region=us-central1"]> -Djava.security.manager=****.org.gradle.process.internal.****.child.BootstrapSecurityManager -Dorg.gradle.native=false -Xmx2g -Dfile.encoding=UTF-8 -Duser.country=US -Duser.language=en -Duser.variant -ea -cp /home/jenkins/.gradle/caches/6.7/****Main/gradle-****.jar ****.org.gradle.process.internal.****.GradleWorkerMain 'Gradle Test Executor 1'
Successfully started process 'Gradle Test Executor 1'
org.apache.beam.sdk.bigqueryioperftests.BigQueryIOIT STANDARD_ERROR
SLF4J: Class path contains multiple SLF4J bindings.
SLF4J: Found binding in [jar:<https://ci-beam.apache.org/job/beam_BiqQueryIO_Streaming_Performance_Test_Java/ws/src/runners/google-cloud-dataflow-java/****/legacy-****/build/libs/beam-runners-google-cloud-dataflow-java-legacy-****-2.28.0-SNAPSHOT.jar!/org/slf4j/impl/StaticLoggerBinder.class]>
SLF4J: Found binding in [jar:file:/home/jenkins/.gradle/caches/modules-2/files-2.1/org.slf4j/slf4j-jdk14/1.7.30/d35953dd2fe54ebe39fdf18cfd82fe6eb35b25ed/slf4j-jdk14-1.7.30.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation.
SLF4J: Actual binding is of type [org.slf4j.impl.JDK14LoggerFactory]
org.apache.beam.sdk.bigqueryioperftests.BigQueryIOIT > testWriteThenRead STANDARD_ERROR
Dec 17, 2020 6:33:41 PM org.apache.beam.runners.dataflow.options.DataflowPipelineOptions$StagingLocationFactory create
INFO: No stagingLocation provided, falling back to gcpTempLocation
Dec 17, 2020 6:33:42 PM org.apache.beam.runners.dataflow.DataflowRunner fromOptions
INFO: PipelineOptions.filesToStage was not specified. Defaulting to files from the classpath: will stage 210 files. Enable logging at DEBUG level to see which files will be staged.
Dec 17, 2020 6:33:44 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Executing pipeline on the Dataflow Service, which will have billing implications related to Google Compute Engine usage and other Google Cloud Services.
Dec 17, 2020 6:33:45 PM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Uploading 211 files from PipelineOptions.filesToStage to staging location to prepare for execution.
Dec 17, 2020 6:33:45 PM org.apache.beam.runners.dataflow.util.PackageUtil tryStagePackage
INFO: Uploading /tmp/test3367858460463909507.zip to gs://temp-storage-for-perf-tests/loadtests/staging/test-LsoHMllRUv9r4tqGjDXHa4zh4i_OniNhxLRZ0NeuA1o.jar
Dec 17, 2020 6:33:45 PM org.apache.beam.runners.dataflow.util.PackageUtil$PackageAttributes forFileToStage
INFO: Staging custom dataflow-****.jar as beam-runners-google-cloud-dataflow-java-legacy-****-2.28.0-SNAPSHOT-SY51_ntnFm08YbRxJRsH0MUwRdL3yw8xnIg9nfCgmLo.jar
Dec 17, 2020 6:33:45 PM org.apache.beam.runners.dataflow.util.PackageUtil tryStagePackage
INFO: Uploading <https://ci-beam.apache.org/job/beam_BiqQueryIO_Streaming_Performance_Test_Java/ws/src/sdks/java/core/build/libs/beam-sdks-java-core-2.28.0-SNAPSHOT-unshaded.jar> to gs://temp-storage-for-perf-tests/loadtests/staging/beam-sdks-java-core-2.28.0-SNAPSHOT-unshaded-s4U8UJqimYWvisJsRpPCMd40lCCCxSlqMh9eKQ-zaQ0.jar
Dec 17, 2020 6:33:46 PM org.apache.beam.runners.dataflow.util.PackageUtil tryStagePackage
INFO: Uploading <https://ci-beam.apache.org/job/beam_BiqQueryIO_Streaming_Performance_Test_Java/ws/src/sdks/java/core/build/libs/beam-sdks-java-core-2.28.0-SNAPSHOT.jar> to gs://temp-storage-for-perf-tests/loadtests/staging/beam-sdks-java-core-2.28.0-SNAPSHOT-PgRMKvwyBWMZCTGia6egvlOhG1IAap9uUu3izcIwrTI.jar
Dec 17, 2020 6:33:46 PM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Staging files complete: 208 files cached, 3 files newly uploaded in 1 seconds
Dec 17, 2020 6:33:46 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read from source as step s1
Dec 17, 2020 6:33:46 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Gather time as step s2
Dec 17, 2020 6:33:46 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Map records as step s3
Dec 17, 2020 6:33:46 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Write to BQ/PrepareWrite/ParDo(Anonymous) as step s4
Dec 17, 2020 6:33:46 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Write to BQ/StreamingInserts/CreateTables/ParDo(CreateTables) as step s5
Dec 17, 2020 6:33:46 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Write to BQ/StreamingInserts/StreamingWriteTables/ShardTableWrites as step s6
Dec 17, 2020 6:33:46 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Write to BQ/StreamingInserts/StreamingWriteTables/TagWithUniqueIds as step s7
Dec 17, 2020 6:33:46 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Write to BQ/StreamingInserts/StreamingWriteTables/Reshuffle/Window.Into()/Window.Assign as step s8
Dec 17, 2020 6:33:46 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Write to BQ/StreamingInserts/StreamingWriteTables/Reshuffle/GroupByKey as step s9
Dec 17, 2020 6:33:46 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Write to BQ/StreamingInserts/StreamingWriteTables/Reshuffle/ExpandIterable as step s10
Dec 17, 2020 6:33:46 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Write to BQ/StreamingInserts/StreamingWriteTables/GlobalWindow/Window.Assign as step s11
Dec 17, 2020 6:33:46 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Write to BQ/StreamingInserts/StreamingWriteTables/StreamingWrite as step s12
Dec 17, 2020 6:33:46 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Staging pipeline description to gs://temp-storage-for-perf-tests/loadtests/staging/
Dec 17, 2020 6:33:47 PM org.apache.beam.runners.dataflow.util.PackageUtil tryStagePackage
INFO: Uploading <109326 bytes, hash 87822f4c8d88e0dcbee62329fcaab803c6fcd02f1ae8274782c9449b528aaa27> to gs://temp-storage-for-perf-tests/loadtests/staging/pipeline-h4IvTI2I4Ny-5iMp_Kq4A8b80C8a6CdHgslEm1KKqic.pb
Dec 17, 2020 6:33:47 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Dataflow SDK version: 2.28.0-SNAPSHOT
Dec 17, 2020 6:33:48 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To access the Dataflow monitoring console, please navigate to https://console.cloud.google.com/dataflow/jobs/us-central1/2020-12-17_10_33_47-11371539742512684120?project=apache-beam-testing
Dec 17, 2020 6:33:48 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Submitted job: 2020-12-17_10_33_47-11371539742512684120
Dec 17, 2020 6:33:48 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To cancel the job using the 'gcloud' tool, run:
> gcloud dataflow jobs --project=apache-beam-testing cancel --region=us-central1 2020-12-17_10_33_47-11371539742512684120
Dec 17, 2020 6:33:49 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
WARNING: 2020-12-17T18:33:47.342Z: The requested max number of ****s (5) is ignored as autoscaling is explicitly disabled (autoscalingAlgorithm=NONE).
Dec 17, 2020 6:34:01 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2020-12-17T18:34:01.083Z: Worker configuration: n1-standard-1 in us-central1-f.
Dec 17, 2020 6:34:04 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2020-12-17T18:34:01.818Z: Expanding CoGroupByKey operations into optimizable parts.
Dec 17, 2020 6:34:04 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2020-12-17T18:34:01.994Z: Expanding GroupByKey operations into optimizable parts.
Dec 17, 2020 6:34:04 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2020-12-17T18:34:02.051Z: Lifting ValueCombiningMappingFns into MergeBucketsMappingFns
Dec 17, 2020 6:34:04 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2020-12-17T18:34:02.234Z: Fusing adjacent ParDo, Read, Write, and Flatten operations
Dec 17, 2020 6:34:04 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2020-12-17T18:34:02.286Z: Fusing consumer Gather time into Read from source
Dec 17, 2020 6:34:04 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2020-12-17T18:34:02.324Z: Fusing consumer Map records into Gather time
Dec 17, 2020 6:34:04 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2020-12-17T18:34:02.360Z: Fusing consumer Write to BQ/PrepareWrite/ParDo(Anonymous) into Map records
Dec 17, 2020 6:34:04 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2020-12-17T18:34:02.429Z: Fusing consumer Write to BQ/StreamingInserts/CreateTables/ParDo(CreateTables) into Write to BQ/PrepareWrite/ParDo(Anonymous)
Dec 17, 2020 6:34:04 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2020-12-17T18:34:02.473Z: Fusing consumer Write to BQ/StreamingInserts/StreamingWriteTables/ShardTableWrites into Write to BQ/StreamingInserts/CreateTables/ParDo(CreateTables)
Dec 17, 2020 6:34:04 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2020-12-17T18:34:02.520Z: Fusing consumer Write to BQ/StreamingInserts/StreamingWriteTables/TagWithUniqueIds into Write to BQ/StreamingInserts/StreamingWriteTables/ShardTableWrites
Dec 17, 2020 6:34:04 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2020-12-17T18:34:02.564Z: Fusing consumer Write to BQ/StreamingInserts/StreamingWriteTables/Reshuffle/Window.Into()/Window.Assign into Write to BQ/StreamingInserts/StreamingWriteTables/TagWithUniqueIds
Dec 17, 2020 6:34:04 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2020-12-17T18:34:02.602Z: Fusing consumer Write to BQ/StreamingInserts/StreamingWriteTables/Reshuffle/GroupByKey/Reify into Write to BQ/StreamingInserts/StreamingWriteTables/Reshuffle/Window.Into()/Window.Assign
Dec 17, 2020 6:34:04 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2020-12-17T18:34:02.645Z: Fusing consumer Write to BQ/StreamingInserts/StreamingWriteTables/Reshuffle/GroupByKey/Write into Write to BQ/StreamingInserts/StreamingWriteTables/Reshuffle/GroupByKey/Reify
Dec 17, 2020 6:34:04 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2020-12-17T18:34:02.681Z: Fusing consumer Write to BQ/StreamingInserts/StreamingWriteTables/Reshuffle/GroupByKey/GroupByWindow into Write to BQ/StreamingInserts/StreamingWriteTables/Reshuffle/GroupByKey/Read
Dec 17, 2020 6:34:04 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2020-12-17T18:34:02.712Z: Fusing consumer Write to BQ/StreamingInserts/StreamingWriteTables/Reshuffle/ExpandIterable into Write to BQ/StreamingInserts/StreamingWriteTables/Reshuffle/GroupByKey/GroupByWindow
Dec 17, 2020 6:34:04 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2020-12-17T18:34:02.774Z: Fusing consumer Write to BQ/StreamingInserts/StreamingWriteTables/GlobalWindow/Window.Assign into Write to BQ/StreamingInserts/StreamingWriteTables/Reshuffle/ExpandIterable
Dec 17, 2020 6:34:04 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2020-12-17T18:34:02.841Z: Fusing consumer Write to BQ/StreamingInserts/StreamingWriteTables/StreamingWrite into Write to BQ/StreamingInserts/StreamingWriteTables/GlobalWindow/Window.Assign
Dec 17, 2020 6:34:04 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2020-12-17T18:34:03.239Z: Executing operation Write to BQ/StreamingInserts/StreamingWriteTables/Reshuffle/GroupByKey/Create
Dec 17, 2020 6:34:04 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2020-12-17T18:34:03.304Z: Starting 5 ****s in us-central1-f...
Dec 17, 2020 6:34:04 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2020-12-17T18:34:03.353Z: Finished operation Write to BQ/StreamingInserts/StreamingWriteTables/Reshuffle/GroupByKey/Create
Dec 17, 2020 6:34:04 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2020-12-17T18:34:03.611Z: Executing operation Read from source+Gather time+Map records+Write to BQ/PrepareWrite/ParDo(Anonymous)+Write to BQ/StreamingInserts/CreateTables/ParDo(CreateTables)+Write to BQ/StreamingInserts/StreamingWriteTables/ShardTableWrites+Write to BQ/StreamingInserts/StreamingWriteTables/TagWithUniqueIds+Write to BQ/StreamingInserts/StreamingWriteTables/Reshuffle/Window.Into()/Window.Assign+Write to BQ/StreamingInserts/StreamingWriteTables/Reshuffle/GroupByKey/Reify+Write to BQ/StreamingInserts/StreamingWriteTables/Reshuffle/GroupByKey/Write
Dec 17, 2020 6:34:09 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2020-12-17T18:34:09.208Z: Your project already contains 100 Dataflow-created metric descriptors, so new user metrics of the form custom.googleapis.com/* will not be created. However, all user metrics are also available in the metric dataflow.googleapis.com/job/user_counter. If you rely on the custom metrics, you can delete old / unused metric descriptors. See https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.list and https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.delete
Dec 17, 2020 6:34:32 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2020-12-17T18:34:32.520Z: Autoscaling: Raised the number of ****s to 1 based on the rate of progress in the currently running stage(s).
Dec 17, 2020 6:34:32 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2020-12-17T18:34:32.543Z: Resized **** pool to 1, though goal was 5. This could be a quota issue.
Dec 17, 2020 6:34:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2020-12-17T18:34:42.794Z: Autoscaling: Raised the number of ****s to 5 based on the rate of progress in the currently running stage(s).
Dec 17, 2020 6:34:57 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2020-12-17T18:34:57.098Z: Workers have started successfully.
Dec 17, 2020 6:34:57 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2020-12-17T18:34:57.126Z: Workers have started successfully.
Dec 17, 2020 6:36:57 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2020-12-17T18:36:55.889Z: Finished operation Read from source+Gather time+Map records+Write to BQ/PrepareWrite/ParDo(Anonymous)+Write to BQ/StreamingInserts/CreateTables/ParDo(CreateTables)+Write to BQ/StreamingInserts/StreamingWriteTables/ShardTableWrites+Write to BQ/StreamingInserts/StreamingWriteTables/TagWithUniqueIds+Write to BQ/StreamingInserts/StreamingWriteTables/Reshuffle/Window.Into()/Window.Assign+Write to BQ/StreamingInserts/StreamingWriteTables/Reshuffle/GroupByKey/Reify+Write to BQ/StreamingInserts/StreamingWriteTables/Reshuffle/GroupByKey/Write
Dec 17, 2020 6:36:57 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2020-12-17T18:36:55.977Z: Executing operation Write to BQ/StreamingInserts/StreamingWriteTables/Reshuffle/GroupByKey/Close
Dec 17, 2020 6:36:57 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2020-12-17T18:36:56.068Z: Finished operation Write to BQ/StreamingInserts/StreamingWriteTables/Reshuffle/GroupByKey/Close
Dec 17, 2020 6:36:57 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2020-12-17T18:36:56.241Z: Executing operation Write to BQ/StreamingInserts/StreamingWriteTables/Reshuffle/GroupByKey/Read+Write to BQ/StreamingInserts/StreamingWriteTables/Reshuffle/GroupByKey/GroupByWindow+Write to BQ/StreamingInserts/StreamingWriteTables/Reshuffle/ExpandIterable+Write to BQ/StreamingInserts/StreamingWriteTables/GlobalWindow/Window.Assign+Write to BQ/StreamingInserts/StreamingWriteTables/StreamingWrite
Dec 17, 2020 6:40:27 PM org.apache.beam.runners.dataflow.DataflowPipelineJob lambda$waitUntilFinish$0
org.apache.beam.sdk.bigqueryioperftests.BigQueryIOIT > testWriteThenRead SKIPPED
Watching 1371 directories to track changes
Watching 1373 directories to track changes
Watching 1374 directories to track changes
> Task :sdks:java:io:bigquery-io-perf-tests:integrationTest FAILED
:sdks:java:io:bigquery-io-perf-tests:integrationTest (Thread[Execution **** for ':' Thread 8,5,main]) completed. Took 6 mins 49.78 secs.
FAILURE: Build failed with an exception.
* What went wrong:
Execution failed for task ':sdks:java:io:bigquery-io-perf-tests:integrationTest'.
> Process 'Gradle Test Executor 1' finished with non-zero exit value 143
This problem might be caused by incorrect test process configuration.
Please refer to the test execution section in the User Manual at https://docs.gradle.org/6.7/userguide/java_testing.html#sec:test_execution
* Try:
Run with --stacktrace option to get the stack trace. Run with --debug option to get more log output. Run with --scan to get full insights.
* Get more help at https://help.gradle.org
Deprecated Gradle features were used in this build, making it incompatible with Gradle 7.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/6.7/userguide/command_line_interface.html#sec:command_line_warnings
BUILD FAILED in 7m 10s
109 actionable tasks: 66 executed, 43 from cache
Watching 1374 directories to track changes
Publishing build scan...
The message received from the daemon indicates that the daemon has disappeared.
Build request sent: Build{id=27675e92-189f-4ceb-9cbc-f25d5d0694ae, currentDir=<https://ci-beam.apache.org/job/beam_BiqQueryIO_Streaming_Performance_Test_Java/ws/src}>
Attempting to read last messages from the daemon log...
Daemon pid: 26426
log file: /home/jenkins/.gradle/daemon/6.7/daemon-26426.out.log
----- Last 20 lines from daemon log file - daemon-26426.out.log -----
Execution failed for task ':sdks:java:io:bigquery-io-perf-tests:integrationTest'.
> Process 'Gradle Test Executor 1' finished with non-zero exit value 143
This problem might be caused by incorrect test process configuration.
Please refer to the test execution section in the User Manual at https://docs.gradle.org/6.7/userguide/java_testing.html#sec:test_execution
* Try:
Run with --stacktrace option to get the stack trace. Run with --debug option to get more log output. Run with --scan to get full insights.
* Get more help at https://help.gradle.org
Deprecated Gradle features were used in this build, making it incompatible with Gradle 7.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/6.7/userguide/command_line_interface.html#sec:command_line_warnings
BUILD FAILED in 7m 10s
109 actionable tasks: 66 executed, 43 from cache
Watching 1374 directories to track changes
Publishing build scan...
Daemon vm is shutting down... The daemon has exited normally or was terminated in response to a user interrupt.
----- End of the daemon log -----
FAILURE: Build failed with an exception.
* What went wrong:
Gradle build daemon disappeared unexpectedly (it may have been killed or may have crashed)
* Try:
Run with --stacktrace option to get the stack trace. Run with --debug option to get more log output. Run with --scan to get full insights.
* Get more help at https://help.gradle.org
Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure
---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org