You are viewing a plain text version of this content. The canonical link for it is here.
Posted to builds@beam.apache.org by Apache Jenkins Server <je...@builds.apache.org> on 2021/05/04 06:56:31 UTC

Build failed in Jenkins: beam_PerformanceTests_MongoDBIO_IT #4533

See <https://ci-beam.apache.org/job/beam_PerformanceTests_MongoDBIO_IT/4533/display/redirect?page=changes>

Changes:

[Alan Myrvold] [BEAM-10912] Increase Jenkins job logs retention from 14 to 30 days

[Pablo Estrada] Revert "Merge pull request #13617 from [BEAM-11289] [Python] Integrate

[Pablo Estrada] Revert type: ignore caused by PR 13617


------------------------------------------
[...truncated 276.02 KB...]
Build cache key for task ':runners:google-cloud-dataflow-java:****:legacy-****:compileJava' is 7b158250161d13e133b536474268aec4
Task ':runners:google-cloud-dataflow-java:****:legacy-****:compileJava' is not up-to-date because:
  No history is available.
Watching 2733 directories to track changes
Watching 2733 directories to track changes
Watching 2733 directories to track changes
Watching 2765 directories to track changes
Watching 2775 directories to track changes
Watching 2777 directories to track changes
Loaded cache entry for task ':runners:google-cloud-dataflow-java:****:legacy-****:compileJava' with cache key 7b158250161d13e133b536474268aec4
:runners:google-cloud-dataflow-java:****:legacy-****:compileJava (Thread[Execution **** for ':' Thread 11,5,main]) completed. Took 0.336 secs.
:runners:google-cloud-dataflow-java:****:legacy-****:classes (Thread[Execution **** for ':' Thread 11,5,main]) started.

> Task :runners:google-cloud-dataflow-java:****:legacy-****:classes UP-TO-DATE
Skipping task ':runners:google-cloud-dataflow-java:****:legacy-****:classes' as it has no actions.
:runners:google-cloud-dataflow-java:****:legacy-****:classes (Thread[Execution **** for ':' Thread 11,5,main]) completed. Took 0.0 secs.
:runners:google-cloud-dataflow-java:****:legacy-****:shadowJar (Thread[Execution **** for ':' Thread 11,5,main]) started.

> Task :sdks:java:io:google-cloud-platform:testJar
Watching 2697 directories to track changes
Watching 2698 directories to track changes
Caching disabled for task ':sdks:java:io:google-cloud-platform:testJar' because:
  Caching has not been enabled for the task
Task ':sdks:java:io:google-cloud-platform:testJar' is not up-to-date because:
  No history is available.
Watching 2698 directories to track changes
Watching 2778 directories to track changes
:sdks:java:io:google-cloud-platform:testJar (Thread[Execution **** for ':' Thread 10,5,main]) completed. Took 0.296 secs.
:runners:google-cloud-dataflow-java:compileTestJava (Thread[Execution **** for ':' Thread 10,5,main]) started.

> Task :runners:google-cloud-dataflow-java:****:legacy-****:shadowJar FROM-CACHE
Watching 2777 directories to track changes
Watching 2777 directories to track changes
Watching 2777 directories to track changes
Watching 2778 directories to track changes
Custom actions are attached to task ':runners:google-cloud-dataflow-java:****:legacy-****:shadowJar'.
Build cache key for task ':runners:google-cloud-dataflow-java:****:legacy-****:shadowJar' is d89ae3d9b22ffc606290d9c39ae4a8f1
Task ':runners:google-cloud-dataflow-java:****:legacy-****:shadowJar' is not up-to-date because:
  No history is available.
Watching 2778 directories to track changes
Watching 2779 directories to track changes
Loaded cache entry for task ':runners:google-cloud-dataflow-java:****:legacy-****:shadowJar' with cache key d89ae3d9b22ffc606290d9c39ae4a8f1
:runners:google-cloud-dataflow-java:****:legacy-****:shadowJar (Thread[Execution **** for ':' Thread 11,5,main]) completed. Took 0.159 secs.

> Task :runners:google-cloud-dataflow-java:compileTestJava FROM-CACHE
Watching 2778 directories to track changes
Watching 2778 directories to track changes
Watching 2778 directories to track changes
Watching 2789 directories to track changes
Custom actions are attached to task ':runners:google-cloud-dataflow-java:compileTestJava'.
Build cache key for task ':runners:google-cloud-dataflow-java:compileTestJava' is 4c02f1a06d93843e3eaa9464c221625e
Task ':runners:google-cloud-dataflow-java:compileTestJava' is not up-to-date because:
  No history is available.
Watching 2789 directories to track changes
Watching 2789 directories to track changes
Watching 2789 directories to track changes
Watching 2800 directories to track changes
Watching 2801 directories to track changes
Watching 2802 directories to track changes
Loaded cache entry for task ':runners:google-cloud-dataflow-java:compileTestJava' with cache key 4c02f1a06d93843e3eaa9464c221625e
:runners:google-cloud-dataflow-java:compileTestJava (Thread[Execution **** for ':' Thread 10,5,main]) completed. Took 0.184 secs.
:runners:google-cloud-dataflow-java:testClasses (Thread[Execution **** for ':' Thread 10,5,main]) started.

> Task :runners:google-cloud-dataflow-java:testClasses UP-TO-DATE
Skipping task ':runners:google-cloud-dataflow-java:testClasses' as it has no actions.
:runners:google-cloud-dataflow-java:testClasses (Thread[Execution **** for ':' Thread 10,5,main]) completed. Took 0.0 secs.
:runners:google-cloud-dataflow-java:testJar (Thread[Execution **** for ':' Thread 5,5,main]) started.

> Task :runners:google-cloud-dataflow-java:testJar
Watching 2802 directories to track changes
Watching 2802 directories to track changes
Watching 2803 directories to track changes
Caching disabled for task ':runners:google-cloud-dataflow-java:testJar' because:
  Caching has not been enabled for the task
Task ':runners:google-cloud-dataflow-java:testJar' is not up-to-date because:
  No history is available.
Watching 2803 directories to track changes
file or directory '<https://ci-beam.apache.org/job/beam_PerformanceTests_MongoDBIO_IT/ws/src/runners/google-cloud-dataflow-java/build/resources/test',> not found
Watching 2803 directories to track changes
:runners:google-cloud-dataflow-java:testJar (Thread[Execution **** for ':' Thread 5,5,main]) completed. Took 0.032 secs.
:sdks:java:io:mongodb:integrationTest (Thread[Execution **** for ':' Thread 5,5,main]) started.
Gradle Test Executor 3 started executing tests.

> Task :sdks:java:io:mongodb:integrationTest
Watching 2803 directories to track changes
Watching 2803 directories to track changes
Watching 2803 directories to track changes
Watching 2803 directories to track changes
Custom actions are attached to task ':sdks:java:io:mongodb:integrationTest'.
Build cache key for task ':sdks:java:io:mongodb:integrationTest' is a9139168b1f9b1d66dd06e53c3431265
Task ':sdks:java:io:mongodb:integrationTest' is not up-to-date because:
  Task.upToDateWhen is false.
Watching 2803 directories to track changes
Watching 2803 directories to track changes
Starting process 'Gradle Test Executor 3'. Working directory: <https://ci-beam.apache.org/job/beam_PerformanceTests_MongoDBIO_IT/ws/src/sdks/java/io/mongodb> Command: /usr/lib/jvm/java-8-openjdk-amd64/bin/java -DbeamTestPipelineOptions=["--tempRoot=gs://temp-storage-for-perf-tests","--project=apache-beam-testing","--numberOfRecords=10000000","--bigQueryDataset=beam_performance","--bigQueryTable=mongodbioit_results","--influxMeasurement=mongodbioit_results","--influxDatabase=beam_test_metrics","--influxHost=http://10.128.0.96:8086","--mongoDBDatabaseName=beam","--mongoDBHostName=107.178.219.49","--mongoDBPort=27017","--runner=DataflowRunner","--autoscalingAlgorithm=NONE","--numWorkers=5","--****HarnessContainerImage=","--dataflowWorkerJar=<https://ci-beam.apache.org/job/beam_PerformanceTests_MongoDBIO_IT/ws/src/runners/google-cloud-dataflow-java/****/legacy-****/build/libs/beam-runners-google-cloud-dataflow-java-legacy-****-2.31.0-SNAPSHOT.jar","--region=us-central1"]> -Djava.security.manager=****.org.gradle.process.internal.****.child.BootstrapSecurityManager -Dorg.gradle.native=false -Xmx2g -Dfile.encoding=UTF-8 -Duser.country=US -Duser.language=en -Duser.variant -ea -cp /home/jenkins/.gradle/caches/6.8.3/****Main/gradle-****.jar ****.org.gradle.process.internal.****.GradleWorkerMain 'Gradle Test Executor 3'
Successfully started process 'Gradle Test Executor 3'

org.apache.beam.sdk.io.mongodb.MongoDBIOIT STANDARD_ERROR
    SLF4J: Class path contains multiple SLF4J bindings.
    SLF4J: Found binding in [jar:file:/home/jenkins/.gradle/caches/modules-2/files-2.1/org.slf4j/slf4j-jdk14/1.7.30/d35953dd2fe54ebe39fdf18cfd82fe6eb35b25ed/slf4j-jdk14-1.7.30.jar!/org/slf4j/impl/StaticLoggerBinder.class]
    SLF4J: Found binding in [jar:<https://ci-beam.apache.org/job/beam_PerformanceTests_MongoDBIO_IT/ws/src/runners/google-cloud-dataflow-java/****/legacy-****/build/libs/beam-runners-google-cloud-dataflow-java-legacy-****-2.31.0-SNAPSHOT.jar!/org/slf4j/impl/StaticLoggerBinder.class]>
    SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation.
    SLF4J: Actual binding is of type [org.slf4j.impl.JDK14LoggerFactory]
    May 04, 2021 6:55:49 AM com.mongodb.diagnostics.logging.SLF4JLogger info
    INFO: Cluster created with settings {hosts=[107.178.219.49:27017], mode=SINGLE, requiredClusterType=UNKNOWN, serverSelectionTimeout='30000 ms', maxWaitQueueSize=500}

org.apache.beam.sdk.io.mongodb.MongoDBIOIT > testWriteAndRead STANDARD_ERROR
    May 04, 2021 6:55:49 AM com.mongodb.diagnostics.logging.SLF4JLogger info
    INFO: Opened connection [connectionId{localValue:1, serverValue:1}] to 107.178.219.49:27017
    May 04, 2021 6:55:49 AM com.mongodb.diagnostics.logging.SLF4JLogger info
    INFO: Monitor thread successfully connected to server with description ServerDescription{address=107.178.219.49:27017, type=STANDALONE, state=CONNECTED, ok=true, version=ServerVersion{versionList=[4, 4, 5]}, minWireVersion=0, maxWireVersion=9, maxDocumentSize=16777216, logicalSessionTimeoutMinutes=30, roundTripTimeNanos=6411343}
    May 04, 2021 6:55:49 AM com.mongodb.diagnostics.logging.SLF4JLogger info
    INFO: Opened connection [connectionId{localValue:2, serverValue:2}] to 107.178.219.49:27017
    May 04, 2021 6:55:50 AM org.apache.beam.runners.dataflow.DataflowRunner validateSdkContainerImageOptions
    WARNING: Prefer --sdkContainerImage over deprecated legacy option --****HarnessContainerImage.
    May 04, 2021 6:55:51 AM org.apache.beam.sdk.extensions.gcp.options.GcpOptions$GcpTempLocationFactory tryCreateDefaultBucket
    INFO: No tempLocation specified, attempting to use default bucket: dataflow-staging-us-central1-844138762903
    May 04, 2021 6:55:51 AM org.apache.beam.sdk.extensions.gcp.util.RetryHttpRequestInitializer$LoggingHttpBackOffHandler handleResponse
    WARNING: Request failed with code 409, performed 0 retries due to IOExceptions, performed 0 retries due to unsuccessful status codes, HTTP framework says request can be retried, (caller responsible for retrying): https://storage.googleapis.com/storage/v1/b?predefinedAcl=projectPrivate&predefinedDefaultObjectAcl=projectPrivate&project=apache-beam-testing. 
    May 04, 2021 6:55:51 AM org.apache.beam.runners.dataflow.options.DataflowPipelineOptions$StagingLocationFactory create
    INFO: No stagingLocation provided, falling back to gcpTempLocation
    May 04, 2021 6:55:52 AM org.apache.beam.runners.dataflow.DataflowRunner fromOptions
    INFO: PipelineOptions.filesToStage was not specified. Defaulting to files from the classpath: will stage 219 files. Enable logging at DEBUG level to see which files will be staged.
    May 04, 2021 6:55:52 AM org.apache.beam.runners.dataflow.DataflowRunner run
    INFO: Executing pipeline on the Dataflow Service, which will have billing implications related to Google Compute Engine usage and other Google Cloud Services.
    May 04, 2021 6:55:54 AM org.apache.beam.runners.dataflow.DataflowRunner run
    INFO: Staging portable pipeline proto to gs://dataflow-staging-us-central1-844138762903/temp/staging/
    May 04, 2021 6:55:55 AM org.apache.beam.runners.dataflow.util.PackageUtil tryStagePackage
    INFO: Uploading <99630 bytes, hash 15533ac027c63506ecfabed7ecffe16c8811f0d66419e18dfdbd3c53d8d2a74b> to gs://dataflow-staging-us-central1-844138762903/temp/staging/pipeline-FVM6wCfGNQbs-r7X7P_hbIgR8NZkGeGN_b08U9jSp0s.pb
    May 04, 2021 6:55:56 AM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
    INFO: Uploading 220 files from PipelineOptions.filesToStage to staging location to prepare for execution.
    May 04, 2021 6:55:56 AM org.apache.beam.runners.dataflow.util.PackageUtil$PackageAttributes forFileToStage
    INFO: Staging custom dataflow-****.jar as beam-runners-google-cloud-dataflow-java-legacy-****-2.31.0-SNAPSHOT-ysjGHVULangrAzZTWjxgBSOSnKjrmbbduKkfNJyYZi8.jar
    May 04, 2021 6:55:56 AM org.apache.beam.runners.dataflow.util.PackageUtil tryStagePackage
    INFO: Uploading /tmp/test2226424904035290725.zip to gs://dataflow-staging-us-central1-844138762903/temp/staging/test-i-e5cdgK9fDElMRHG-A2PQ1_VUKQPXC2rzHxRtl0I8k.jar
    May 04, 2021 6:55:57 AM org.apache.beam.runners.dataflow.util.PackageUtil tryStagePackage
    INFO: Uploading /tmp/main252501345490065764.zip to gs://dataflow-staging-us-central1-844138762903/temp/staging/main-HLgAHVsMLIwL3JJse-7WFd6DgnR5exhGWgYB3bsRvKA.jar
    May 04, 2021 6:55:57 AM org.apache.beam.runners.dataflow.util.PackageUtil stageClasspathElements
    INFO: Staging files complete: 218 files cached, 2 files newly uploaded in 0 seconds
    May 04, 2021 6:55:57 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding Generate sequence/Read(BoundedCountingSource) as step s1
    May 04, 2021 6:55:57 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding Produce documents/Map as step s2
    May 04, 2021 6:55:57 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding Collect write time metric as step s3
    May 04, 2021 6:55:57 AM org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding Write documents to MongoDB/ParDo(Write) as step s4
    May 04, 2021 6:55:57 AM org.apache.beam.runners.dataflow.DataflowRunner run
    INFO: Dataflow SDK version: 2.31.0-SNAPSHOT
    May 04, 2021 6:55:58 AM org.apache.beam.runners.dataflow.DataflowRunner run
    INFO: To access the Dataflow monitoring console, please navigate to https://console.cloud.google.com/dataflow/jobs/us-central1/2021-05-03_23_55_57-10057825538055905430?project=apache-beam-testing
    May 04, 2021 6:55:58 AM org.apache.beam.runners.dataflow.DataflowRunner run
    INFO: Submitted job: 2021-05-03_23_55_57-10057825538055905430
    May 04, 2021 6:55:58 AM org.apache.beam.runners.dataflow.DataflowRunner run
    INFO: To cancel the job using the 'gcloud' tool, run:
    > gcloud dataflow jobs --project=apache-beam-testing cancel --region=us-central1 2021-05-03_23_55_57-10057825538055905430
    May 04, 2021 6:56:11 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2021-05-04T06:56:10.215Z: Worker configuration: n1-standard-1 in us-central1-f.
    May 04, 2021 6:56:11 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2021-05-04T06:56:11.044Z: Expanding CoGroupByKey operations into optimizable parts.
    May 04, 2021 6:56:11 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2021-05-04T06:56:11.163Z: Expanding GroupByKey operations into optimizable parts.
    May 04, 2021 6:56:11 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2021-05-04T06:56:11.195Z: Lifting ValueCombiningMappingFns into MergeBucketsMappingFns
    May 04, 2021 6:56:11 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2021-05-04T06:56:11.355Z: Fusing adjacent ParDo, Read, Write, and Flatten operations
    May 04, 2021 6:56:11 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2021-05-04T06:56:11.425Z: Fusing consumer Produce documents/Map into Generate sequence/Read(BoundedCountingSource)
    May 04, 2021 6:56:13 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2021-05-04T06:56:11.488Z: Fusing consumer Collect write time metric into Produce documents/Map
    May 04, 2021 6:56:13 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2021-05-04T06:56:11.545Z: Fusing consumer Write documents to MongoDB/ParDo(Write) into Collect write time metric
    May 04, 2021 6:56:13 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2021-05-04T06:56:12.186Z: Executing operation Generate sequence/Read(BoundedCountingSource)+Produce documents/Map+Collect write time metric+Write documents to MongoDB/ParDo(Write)
    May 04, 2021 6:56:13 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2021-05-04T06:56:12.357Z: Starting 5 ****s in us-central1-f...
    May 04, 2021 6:56:21 AM org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2021-05-04T06:56:20.668Z: Your project already contains 100 Dataflow-created metric descriptors, so new user metrics of the form custom.googleapis.com/* will not be created. However, all user metrics are also available in the metric dataflow.googleapis.com/job/user_counter. If you rely on the custom metrics, you can delete old / unused metric descriptors. See https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.list and https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.delete

org.apache.beam.sdk.io.mongodb.MongoDBIOIT > testWriteAndRead SKIPPED
Watching 2805 directories to track changes
Watching 2807 directories to track changes
Watching 2808 directories to track changes

> Task :sdks:java:io:mongodb:integrationTest FAILED
:sdks:java:io:mongodb:integrationTest (Thread[Execution **** for ':' Thread 5,5,main]) completed. Took 44.992 secs.

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':sdks:java:io:mongodb:integrationTest'.
> Process 'Gradle Test Executor 3' finished with non-zero exit value 143
  This problem might be caused by incorrect test process configuration.
  Please refer to the test execution section in the User Manual at https://docs.gradle.org/6.8.3/userguide/java_testing.html#sec:test_execution

* Try:
Run with --stacktrace option to get the stack trace. Run with --debug option to get more log output. Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 7.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/6.8.3/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 1m 3s
107 actionable tasks: 65 executed, 42 from cache
Watching 2808 directories to track changes

Publishing build scan...
The message received from the daemon indicates that the daemon has disappeared.
Build request sent: Build{id=c737afd3-d862-4666-98a4-7afbc93e12f6, currentDir=<https://ci-beam.apache.org/job/beam_PerformanceTests_MongoDBIO_IT/ws/src}>
Attempting to read last messages from the daemon log...
Daemon pid: 11737
  log file: /home/jenkins/.gradle/daemon/6.8.3/daemon-11737.out.log
----- Last  20 lines from daemon log file - daemon-11737.out.log -----
Execution failed for task ':sdks:java:io:mongodb:integrationTest'.
> Process 'Gradle Test Executor 3' finished with non-zero exit value 143
  This problem might be caused by incorrect test process configuration.
  Please refer to the test execution section in the User Manual at https://docs.gradle.org/6.8.3/userguide/java_testing.html#sec:test_execution

* Try:
Run with --stacktrace option to get the stack trace. Run with --debug option to get more log output. Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 7.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/6.8.3/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 1m 3s
107 actionable tasks: 65 executed, 42 from cache
Watching 2808 directories to track changes

Publishing build scan...
Daemon vm is shutting down... The daemon has exited normally or was terminated in response to a user interrupt.
----- End of the daemon log -----


FAILURE: Build failed with an exception.

* What went wrong:
Gradle build daemon disappeared unexpectedly (it may have been killed or may have crashed)

* Try:
Run with --stacktrace option to get the stack trace. Run with --debug option to get more log output. Run with --scan to get full insights.

* Get more help at https://help.gradle.org
Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Jenkins build is back to normal : beam_PerformanceTests_MongoDBIO_IT #4534

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PerformanceTests_MongoDBIO_IT/4534/display/redirect?page=changes>


---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org