You are viewing a plain text version of this content. The canonical link for it is here.
Posted to builds@beam.apache.org by Apache Jenkins Server <je...@builds.apache.org> on 2021/07/04 19:01:13 UTC
Build failed in Jenkins: beam_PerformanceTests_MongoDBIO_IT #4749
See <https://ci-beam.apache.org/job/beam_PerformanceTests_MongoDBIO_IT/4749/display/redirect>
Changes:
------------------------------------------
[...truncated 275.47 KB...]
INFO: 2021-07-04T18:58:41.667Z: Cleaning up.
Jul 04, 2021 6:58:43 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2021-07-04T18:58:41.725Z: Stopping **** pool...
Jul 04, 2021 6:59:26 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2021-07-04T18:59:25.168Z: Worker pool stopped.
Jul 04, 2021 6:59:30 PM org.apache.beam.runners.dataflow.DataflowPipelineJob logTerminalState
INFO: Job 2021-07-04_11_55_29-9124921444629934573 finished with status DONE.
Jul 04, 2021 6:59:30 PM org.apache.beam.runners.dataflow.DataflowRunner validateSdkContainerImageOptions
WARNING: Prefer --sdkContainerImage over deprecated legacy option --****HarnessContainerImage.
Jul 04, 2021 6:59:30 PM org.apache.beam.sdk.extensions.gcp.options.GcpOptions$GcpTempLocationFactory tryCreateDefaultBucket
INFO: No tempLocation specified, attempting to use default bucket: dataflow-staging-us-central1-844138762903
Jul 04, 2021 6:59:30 PM org.apache.beam.sdk.extensions.gcp.util.RetryHttpRequestInitializer$LoggingHttpBackOffHandler handleResponse
WARNING: Request failed with code 409, performed 0 retries due to IOExceptions, performed 0 retries due to unsuccessful status codes, HTTP framework says request can be retried, (caller responsible for retrying): https://storage.googleapis.com/storage/v1/b?predefinedAcl=projectPrivate&predefinedDefaultObjectAcl=projectPrivate&project=apache-beam-testing.
Jul 04, 2021 6:59:30 PM org.apache.beam.runners.dataflow.options.DataflowPipelineOptions$StagingLocationFactory create
INFO: No stagingLocation provided, falling back to gcpTempLocation
Jul 04, 2021 6:59:31 PM org.apache.beam.runners.dataflow.DataflowRunner fromOptions
INFO: PipelineOptions.filesToStage was not specified. Defaulting to files from the classpath: will stage 222 files. Enable logging at DEBUG level to see which files will be staged.
Jul 04, 2021 6:59:31 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Executing pipeline on the Dataflow Service, which will have billing implications related to Google Compute Engine usage and other Google Cloud Services.
Jul 04, 2021 6:59:32 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Staging portable pipeline proto to gs://dataflow-staging-us-central1-844138762903/temp/staging/
Jul 04, 2021 6:59:32 PM org.apache.beam.runners.dataflow.util.PackageUtil tryStagePackage
INFO: Uploading <166405 bytes, hash 0f8f7f8e2d8461166ff63e207e9f42fe6b6df99d6eeb86432d354f5e256298ab> to gs://dataflow-staging-us-central1-844138762903/temp/staging/pipeline-D49_ji2EYRZv9j4gfp9C_mtt-Z1u64ZDLTVPXiVimKs.pb
Jul 04, 2021 6:59:34 PM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Uploading 223 files from PipelineOptions.filesToStage to staging location to prepare for execution.
Jul 04, 2021 6:59:34 PM org.apache.beam.runners.dataflow.util.PackageUtil$PackageAttributes forFileToStage
INFO: Staging custom dataflow-****.jar as beam-runners-google-cloud-dataflow-java-legacy-****-2.32.0-SNAPSHOT-ni6oCS1kBBL4-VhvhkqB6sruu6ZfCChHGIhR1YLuF6k.jar
Jul 04, 2021 6:59:34 PM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
INFO: Staging files complete: 223 files cached, 0 files newly uploaded in 0 seconds
Jul 04, 2021 6:59:34 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Read all documents/Read(BoundedMongoDbSource) as step s1
Jul 04, 2021 6:59:34 PM com.mongodb.diagnostics.logging.SLF4JLogger info
INFO: Cluster created with settings {hosts=[34.123.35.163:27017], mode=SINGLE, requiredClusterType=UNKNOWN, serverSelectionTimeout='30000 ms', maxWaitQueueSize=500}
Jul 04, 2021 6:59:34 PM com.mongodb.diagnostics.logging.SLF4JLogger info
INFO: Cluster description not yet available. Waiting for 30000 ms before timing out
Jul 04, 2021 6:59:34 PM com.mongodb.diagnostics.logging.SLF4JLogger info
INFO: Opened connection [connectionId{localValue:3, serverValue:13}] to 34.123.35.163:27017
Jul 04, 2021 6:59:34 PM com.mongodb.diagnostics.logging.SLF4JLogger info
INFO: Monitor thread successfully connected to server with description ServerDescription{address=34.123.35.163:27017, type=STANDALONE, state=CONNECTED, ok=true, version=ServerVersion{versionList=[4, 4, 6]}, minWireVersion=0, maxWireVersion=9, maxDocumentSize=16777216, logicalSessionTimeoutMinutes=30, roundTripTimeNanos=1512972}
Jul 04, 2021 6:59:34 PM com.mongodb.diagnostics.logging.SLF4JLogger info
INFO: Opened connection [connectionId{localValue:4, serverValue:14}] to 34.123.35.163:27017
Jul 04, 2021 6:59:34 PM com.mongodb.diagnostics.logging.SLF4JLogger info
INFO: Closed connection [connectionId{localValue:4, serverValue:14}] to 34.123.35.163:27017 because the pool has been closed.
Jul 04, 2021 6:59:34 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Collect read time metrics as step s2
Jul 04, 2021 6:59:34 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Map documents to Strings/Map as step s3
Jul 04, 2021 6:59:34 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Calculate hashcode/WithKeys/AddKeys/Map as step s4
Jul 04, 2021 6:59:34 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Calculate hashcode/Combine.perKey(Hashing)/GroupByKey as step s5
Jul 04, 2021 6:59:34 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Calculate hashcode/Combine.perKey(Hashing)/Combine.GroupedValues as step s6
Jul 04, 2021 6:59:34 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Calculate hashcode/Values/Values/Map as step s7
Jul 04, 2021 6:59:34 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Calculate hashcode/View.AsIterable/ParDo(ToIsmRecordForGlobalWindow) as step s8
Jul 04, 2021 6:59:34 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Calculate hashcode/View.AsIterable/CreateDataflowView as step s9
Jul 04, 2021 6:59:34 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Calculate hashcode/CreateVoid/Read(CreateSource) as step s10
Jul 04, 2021 6:59:34 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Calculate hashcode/ProduceDefault as step s11
Jul 04, 2021 6:59:34 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding Calculate hashcode/Flatten.PCollections as step s12
Jul 04, 2021 6:59:34 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding PAssert$0/GroupGlobally/Reify.Window/ParDo(Anonymous) as step s13
Jul 04, 2021 6:59:34 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding PAssert$0/GroupGlobally/ParDo(ToSingletonIterables) as step s14
Jul 04, 2021 6:59:34 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding PAssert$0/GroupGlobally/Create.Values/Read(CreateSource) as step s15
Jul 04, 2021 6:59:34 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding PAssert$0/GroupGlobally/Flatten.PCollections as step s16
Jul 04, 2021 6:59:34 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding PAssert$0/GroupGlobally/Window.Into()/Flatten.PCollections as step s17
Jul 04, 2021 6:59:34 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding PAssert$0/GroupGlobally/WithKeys/AddKeys/Map as step s18
Jul 04, 2021 6:59:34 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding PAssert$0/GroupGlobally/GroupByKey as step s19
Jul 04, 2021 6:59:34 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding PAssert$0/GroupGlobally/Values/Values/Map as step s20
Jul 04, 2021 6:59:34 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding PAssert$0/GroupGlobally/ParDo(Concat) as step s21
Jul 04, 2021 6:59:34 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding PAssert$0/GetPane/Map as step s22
Jul 04, 2021 6:59:34 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding PAssert$0/RunChecks as step s23
Jul 04, 2021 6:59:34 PM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
INFO: Adding PAssert$0/VerifyAssertions/ParDo(DefaultConclude) as step s24
Jul 04, 2021 6:59:34 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Dataflow SDK version: 2.32.0-SNAPSHOT
Jul 04, 2021 6:59:35 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To access the Dataflow monitoring console, please navigate to https://console.cloud.google.com/dataflow/jobs/us-central1/2021-07-04_11_59_34-11575277864115487837?project=apache-beam-testing
Jul 04, 2021 6:59:35 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: Submitted job: 2021-07-04_11_59_34-11575277864115487837
Jul 04, 2021 6:59:35 PM org.apache.beam.runners.dataflow.DataflowRunner run
INFO: To cancel the job using the 'gcloud' tool, run:
> gcloud dataflow jobs --project=apache-beam-testing cancel --region=us-central1 2021-07-04_11_59_34-11575277864115487837
Jul 04, 2021 6:59:49 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2021-07-04T18:59:48.344Z: Worker configuration: n1-standard-1 in us-central1-c.
Jul 04, 2021 6:59:49 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2021-07-04T18:59:49.116Z: Expanding CoGroupByKey operations into optimizable parts.
Jul 04, 2021 6:59:49 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2021-07-04T18:59:49.261Z: Expanding GroupByKey operations into optimizable parts.
Jul 04, 2021 6:59:49 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2021-07-04T18:59:49.288Z: Lifting ValueCombiningMappingFns into MergeBucketsMappingFns
Jul 04, 2021 6:59:49 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2021-07-04T18:59:49.402Z: Fusing adjacent ParDo, Read, Write, and Flatten operations
Jul 04, 2021 6:59:49 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2021-07-04T18:59:49.421Z: Elided trivial flatten
Jul 04, 2021 6:59:49 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2021-07-04T18:59:49.441Z: Fusing consumer Calculate hashcode/View.AsIterable/ParDo(ToIsmRecordForGlobalWindow) into Calculate hashcode/Values/Values/Map
Jul 04, 2021 6:59:49 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2021-07-04T18:59:49.468Z: Unzipping flatten s16 for input s14.org.apache.beam.sdk.values.PCollection.<init>:408#b70c45c110743c2b
Jul 04, 2021 6:59:49 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2021-07-04T18:59:49.490Z: Fusing unzipped copy of PAssert$0/GroupGlobally/WithKeys/AddKeys/Map, through flatten PAssert$0/GroupGlobally/Flatten.PCollections, into producer PAssert$0/GroupGlobally/ParDo(ToSingletonIterables)
Jul 04, 2021 6:59:49 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2021-07-04T18:59:49.506Z: Unzipping flatten s12 for input s11.org.apache.beam.sdk.values.PCollection.<init>:408#154393cf2f9d04a6
Jul 04, 2021 6:59:49 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2021-07-04T18:59:49.527Z: Fusing unzipped copy of PAssert$0/GroupGlobally/Reify.Window/ParDo(Anonymous), through flatten Calculate hashcode/Flatten.PCollections, into producer Calculate hashcode/ProduceDefault
Jul 04, 2021 6:59:51 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2021-07-04T18:59:49.551Z: Fusing consumer PAssert$0/GroupGlobally/Reify.Window/ParDo(Anonymous) into Calculate hashcode/Values/Values/Map
Jul 04, 2021 6:59:51 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2021-07-04T18:59:49.586Z: Fusing consumer PAssert$0/GroupGlobally/GroupByKey/Write into PAssert$0/GroupGlobally/GroupByKey/Reify
Jul 04, 2021 6:59:51 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2021-07-04T18:59:49.613Z: Fusing consumer PAssert$0/GroupGlobally/GroupByKey/GroupByWindow into PAssert$0/GroupGlobally/GroupByKey/Read
Jul 04, 2021 6:59:51 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2021-07-04T18:59:49.647Z: Fusing consumer PAssert$0/GroupGlobally/Values/Values/Map into PAssert$0/GroupGlobally/GroupByKey/GroupByWindow
Jul 04, 2021 6:59:51 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2021-07-04T18:59:49.675Z: Fusing consumer PAssert$0/GroupGlobally/ParDo(Concat) into PAssert$0/GroupGlobally/Values/Values/Map
Jul 04, 2021 6:59:51 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2021-07-04T18:59:49.703Z: Fusing consumer PAssert$0/GetPane/Map into PAssert$0/GroupGlobally/ParDo(Concat)
Jul 04, 2021 6:59:51 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2021-07-04T18:59:49.733Z: Fusing consumer PAssert$0/RunChecks into PAssert$0/GetPane/Map
Jul 04, 2021 6:59:51 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2021-07-04T18:59:49.768Z: Fusing consumer PAssert$0/VerifyAssertions/ParDo(DefaultConclude) into PAssert$0/RunChecks
Jul 04, 2021 6:59:51 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2021-07-04T18:59:49.807Z: Unzipping flatten s16-u31 for input s18.org.apache.beam.sdk.values.PCollection.<init>:408#56b99bb29b40d50c-c29
Jul 04, 2021 6:59:51 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2021-07-04T18:59:49.861Z: Fusing unzipped copy of PAssert$0/GroupGlobally/GroupByKey/Reify, through flatten PAssert$0/GroupGlobally/Flatten.PCollections/Unzipped-1, into producer PAssert$0/GroupGlobally/WithKeys/AddKeys/Map
Jul 04, 2021 6:59:51 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2021-07-04T18:59:49.895Z: Unzipping flatten s12-u36 for input s13.org.apache.beam.sdk.values.PCollection.<init>:408#b8ec6571abcba715-c34
Jul 04, 2021 6:59:51 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2021-07-04T18:59:49.920Z: Fusing unzipped copy of PAssert$0/GroupGlobally/ParDo(ToSingletonIterables), through flatten Calculate hashcode/Flatten.PCollections/Unzipped-1, into producer PAssert$0/GroupGlobally/Reify.Window/ParDo(Anonymous)
Jul 04, 2021 6:59:51 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2021-07-04T18:59:49.950Z: Fusing consumer PAssert$0/GroupGlobally/ParDo(ToSingletonIterables) into PAssert$0/GroupGlobally/Reify.Window/ParDo(Anonymous)
Jul 04, 2021 6:59:51 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2021-07-04T18:59:49.989Z: Fusing consumer Collect read time metrics into Read all documents/Read(BoundedMongoDbSource)
Jul 04, 2021 6:59:51 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2021-07-04T18:59:50.023Z: Fusing consumer Map documents to Strings/Map into Collect read time metrics
Jul 04, 2021 6:59:51 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2021-07-04T18:59:50.045Z: Fusing consumer Calculate hashcode/WithKeys/AddKeys/Map into Map documents to Strings/Map
Jul 04, 2021 6:59:51 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2021-07-04T18:59:50.068Z: Fusing consumer Calculate hashcode/Combine.perKey(Hashing)/GroupByKey+Calculate hashcode/Combine.perKey(Hashing)/Combine.GroupedValues/Partial into Calculate hashcode/WithKeys/AddKeys/Map
Jul 04, 2021 6:59:51 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2021-07-04T18:59:50.090Z: Fusing consumer Calculate hashcode/Combine.perKey(Hashing)/GroupByKey/Reify into Calculate hashcode/Combine.perKey(Hashing)/GroupByKey+Calculate hashcode/Combine.perKey(Hashing)/Combine.GroupedValues/Partial
Jul 04, 2021 6:59:51 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2021-07-04T18:59:50.126Z: Fusing consumer Calculate hashcode/Combine.perKey(Hashing)/GroupByKey/Write into Calculate hashcode/Combine.perKey(Hashing)/GroupByKey/Reify
Jul 04, 2021 6:59:51 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2021-07-04T18:59:50.161Z: Fusing consumer Calculate hashcode/Combine.perKey(Hashing)/Combine.GroupedValues into Calculate hashcode/Combine.perKey(Hashing)/GroupByKey/Read
Jul 04, 2021 6:59:51 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2021-07-04T18:59:50.193Z: Fusing consumer Calculate hashcode/Combine.perKey(Hashing)/Combine.GroupedValues/Extract into Calculate hashcode/Combine.perKey(Hashing)/Combine.GroupedValues
Jul 04, 2021 6:59:51 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2021-07-04T18:59:50.217Z: Fusing consumer Calculate hashcode/Values/Values/Map into Calculate hashcode/Combine.perKey(Hashing)/Combine.GroupedValues/Extract
Jul 04, 2021 6:59:51 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2021-07-04T18:59:50.247Z: Fusing consumer Calculate hashcode/ProduceDefault into Calculate hashcode/CreateVoid/Read(CreateSource)
Jul 04, 2021 6:59:51 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2021-07-04T18:59:50.278Z: Fusing consumer PAssert$0/GroupGlobally/WithKeys/AddKeys/Map into PAssert$0/GroupGlobally/Create.Values/Read(CreateSource)
Jul 04, 2021 6:59:51 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2021-07-04T18:59:50.303Z: Fusing consumer PAssert$0/GroupGlobally/GroupByKey/Reify into PAssert$0/GroupGlobally/WithKeys/AddKeys/Map
Jul 04, 2021 6:59:51 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2021-07-04T18:59:50.824Z: Executing operation Calculate hashcode/Combine.perKey(Hashing)/GroupByKey/Create
Jul 04, 2021 6:59:51 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2021-07-04T18:59:50.905Z: Starting 5 ****s in us-central1-c...
Jul 04, 2021 6:59:51 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2021-07-04T18:59:51.341Z: Finished operation Calculate hashcode/Combine.perKey(Hashing)/GroupByKey/Create
Jul 04, 2021 6:59:51 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2021-07-04T18:59:51.500Z: Executing operation Read all documents/Read(BoundedMongoDbSource)+Collect read time metrics+Map documents to Strings/Map+Calculate hashcode/WithKeys/AddKeys/Map+Calculate hashcode/Combine.perKey(Hashing)/GroupByKey+Calculate hashcode/Combine.perKey(Hashing)/Combine.GroupedValues/Partial+Calculate hashcode/Combine.perKey(Hashing)/GroupByKey/Reify+Calculate hashcode/Combine.perKey(Hashing)/GroupByKey/Write
Jul 04, 2021 6:59:58 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2021-07-04T18:59:57.556Z: Your project already contains 100 Dataflow-created metric descriptors, so new user metrics of the form custom.googleapis.com/* will not be created. However, all user metrics are also available in the metric dataflow.googleapis.com/job/user_counter. If you rely on the custom metrics, you can delete old / unused metric descriptors. See https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.list and https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.delete
Jul 04, 2021 7:01:01 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2021-07-04T19:01:00.455Z: Workers have started successfully.
Jul 04, 2021 7:01:01 PM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
INFO: 2021-07-04T19:01:00.479Z: Workers have started successfully.
org.apache.beam.sdk.io.mongodb.MongoDBIOIT > testWriteAndRead SKIPPED
> Task :sdks:java:io:mongodb:integrationTest FAILED
:sdks:java:io:mongodb:integrationTest (Thread[Execution **** for ':',5,main]) completed. Took 5 mins 55.092 secs.
FAILURE: Build failed with an exception.
* What went wrong:
Execution failed for task ':sdks:java:io:mongodb:integrationTest'.
> Process 'Gradle Test Executor 2' finished with non-zero exit value 143
This problem might be caused by incorrect test process configuration.
Please refer to the test execution section in the User Manual at https://docs.gradle.org/6.8.3/userguide/java_testing.html#sec:test_execution
* Try:
Run with --stacktrace option to get the stack trace. Run with --debug option to get more log output. Run with --scan to get full insights.
* Get more help at https://help.gradle.org
Deprecated Gradle features were used in this build, making it incompatible with Gradle 7.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/6.8.3/userguide/command_line_interface.html#sec:command_line_warnings
BUILD FAILED in 6m 9s
113 actionable tasks: 68 executed, 45 from cache
The message received from the daemon indicates that the daemon has disappeared.
Build request sent: Build{id=792f8aef-9dd5-4a12-bdc5-e3676a43019c, currentDir=<https://ci-beam.apache.org/job/beam_PerformanceTests_MongoDBIO_IT/ws/src}>
Attempting to read last messages from the daemon log...
Daemon pid: 16235
log file: /home/jenkins/.gradle/daemon/6.8.3/daemon-16235.out.log
----- Last 20 lines from daemon log file - daemon-16235.out.log -----
FAILURE: Build failed with an exception.
* What went wrong:
Execution failed for task ':sdks:java:io:mongodb:integrationTest'.
> Process 'Gradle Test Executor 2' finished with non-zero exit value 143
This problem might be caused by incorrect test process configuration.
Please refer to the test execution section in the User Manual at https://docs.gradle.org/6.8.3/userguide/java_testing.html#sec:test_execution
* Try:
Run with --stacktrace option to get the stack trace. Run with --debug option to get more log output. Run with --scan to get full insights.
* Get more help at https://help.gradle.org
Deprecated Gradle features were used in this build, making it incompatible with Gradle 7.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/6.8.3/userguide/command_line_interface.html#sec:command_line_warnings
BUILD FAILED in 6m 9s
113 actionable tasks: 68 executed, 45 from cache
Daemon vm is shutting down... The daemon has exited normally or was terminated in response to a user interrupt.
----- End of the daemon log -----
FAILURE: Build failed with an exception.
* What went wrong:
Gradle build daemon disappeared unexpectedly (it may have been killed or may have crashed)
* Try:
Run with --stacktrace option to get the stack trace. Run with --debug option to get more log output. Run with --scan to get full insights.
* Get more help at https://help.gradle.org
Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure
---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org
Jenkins build is back to normal :
beam_PerformanceTests_MongoDBIO_IT #4750
Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PerformanceTests_MongoDBIO_IT/4750/display/redirect>
---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org