You are viewing a plain text version of this content. The canonical link for it is here.
Posted to builds@beam.apache.org by Apache Jenkins Server <je...@builds.apache.org> on 2020/05/10 17:50:00 UTC

beam_LoadTests_Python_ParDo_Flink_Batch - Build # 228 - Aborted

The Apache Jenkins build system has built beam_LoadTests_Python_ParDo_Flink_Batch (build #228)

Status: Aborted

Check console output at https://builds.apache.org/job/beam_LoadTests_Python_ParDo_Flink_Batch/228/ to view the results.

beam_LoadTests_Python_ParDo_Flink_Batch - Build # 251 - Aborted

Posted by Apache Jenkins Server <je...@builds.apache.org>.
The Apache Jenkins build system has built beam_LoadTests_Python_ParDo_Flink_Batch (build #251)

Status: Aborted

Check console output at https://builds.apache.org/job/beam_LoadTests_Python_ParDo_Flink_Batch/251/ to view the results.

beam_LoadTests_Python_ParDo_Flink_Batch - Build # 250 - Aborted

Posted by Apache Jenkins Server <je...@builds.apache.org>.
The Apache Jenkins build system has built beam_LoadTests_Python_ParDo_Flink_Batch (build #250)

Status: Aborted

Check console output at https://builds.apache.org/job/beam_LoadTests_Python_ParDo_Flink_Batch/250/ to view the results.

beam_LoadTests_Python_ParDo_Flink_Batch - Build # 249 - Aborted

Posted by Apache Jenkins Server <je...@builds.apache.org>.
The Apache Jenkins build system has built beam_LoadTests_Python_ParDo_Flink_Batch (build #249)

Status: Aborted

Check console output at https://builds.apache.org/job/beam_LoadTests_Python_ParDo_Flink_Batch/249/ to view the results.

Build failed in Jenkins: beam_LoadTests_Python_ParDo_Flink_Batch #248

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/beam_LoadTests_Python_ParDo_Flink_Batch/248/display/redirect?page=changes>

Changes:

[bhulette] remove as_cloud_object override

[sniemitz] [BEAM-9964] Move workerCacheMb to a user-visible place

[heejong] [BEAM-10125] adding cross-language KafkaIO integration test

[github] [BEAM-10078] uniquify Dataflow specific jars when staging (#11814)

[github] [BEAM-10107] Remove outdated instructions for website updates in releā€¦

[github] [BEAM-9679] Add a CoGroupByKey lesson to the Core Transforms section

[github] [BEAM-10110] Propagate ids for custom coders. (#11832)

[rohde.samuel] Moves apply_WriteToBigQuery to override

[github] [BEAM-10121] Python RowCoder doesn't support nested structs (#11841)

[rohde.samuel] Skip test if py=3.8 see BEAM-9754

[github] [BEAM-7746] Fix type errors and enable checks for


------------------------------------------
[...truncated 46.92 KB...]
697765a85531: Layer already exists
8c39f7b1a31a: Layer already exists
88cfc2fcd059: Layer already exists
7cc1c2d7e744: Layer already exists
760e8d95cf58: Layer already exists
8c02234b8605: Layer already exists
ca59b5db91f8: Pushed
4da4180084ab: Pushed
6dca4ec02d11: Pushed
latest: digest: sha256:aca34f6444ac6cc988bf1e895f8aa8fc4d6c87fb40e2d9f7bd6871d04266a5c3 size: 4938
[Gradle] - Launching build.
[src] $ <https://builds.apache.org/job/beam_LoadTests_Python_ParDo_Flink_Batch/ws/src/gradlew> --continue --max-****s=12 -Dorg.gradle.jvmargs=-Xms2g -Dorg.gradle.jvmargs=-Xmx4g -Pdocker-pull-licenses -Pdocker-repository-root=gcr.io/apache-beam-testing/beam_portability -Pdocker-tag=latest :runners:flink:1.10:job-server-container:docker
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy UP-TO-DATE
> Task :buildSrc:pluginDescriptors UP-TO-DATE
> Task :buildSrc:processResources UP-TO-DATE
> Task :buildSrc:classes UP-TO-DATE
> Task :buildSrc:jar UP-TO-DATE
> Task :buildSrc:assemble UP-TO-DATE
> Task :buildSrc:spotlessGroovy UP-TO-DATE
> Task :buildSrc:spotlessGroovyCheck UP-TO-DATE
> Task :buildSrc:spotlessGroovyGradle UP-TO-DATE
> Task :buildSrc:spotlessGroovyGradleCheck UP-TO-DATE
> Task :buildSrc:spotlessCheck UP-TO-DATE
> Task :buildSrc:pluginUnderTestMetadata UP-TO-DATE
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validateTaskProperties UP-TO-DATE
> Task :buildSrc:check UP-TO-DATE
> Task :buildSrc:build UP-TO-DATE
Configuration on demand is an incubating feature.
> Task :sdks:java:core:generateAvroProtocol NO-SOURCE
> Task :vendor:sdks-java-extensions-protobuf:processResources NO-SOURCE
> Task :runners:core-construction-java:processResources NO-SOURCE
> Task :sdks:java:fn-execution:processResources NO-SOURCE
> Task :sdks:java:extensions:google-cloud-platform-core:processResources NO-SOURCE
> Task :runners:core-java:processResources NO-SOURCE
> Task :sdks:java:core:generateAvroJava NO-SOURCE
> Task :sdks:java:expansion-service:processResources NO-SOURCE
> Task :sdks:java:harness:processResources NO-SOURCE
> Task :runners:java-fn-execution:processResources NO-SOURCE
> Task :runners:java-job-service:processResources NO-SOURCE
> Task :runners:flink:1.10:copyResourcesOverrides NO-SOURCE
> Task :runners:flink:1.10:job-server:processResources NO-SOURCE
> Task :sdks:java:io:google-cloud-platform:processResources NO-SOURCE
> Task :sdks:java:extensions:protobuf:extractProto
> Task :model:job-management:extractProto
> Task :sdks:java:io:kafka:processResources NO-SOURCE
> Task :model:fn-execution:extractProto
> Task :sdks:java:extensions:protobuf:processResources NO-SOURCE
> Task :runners:flink:1.10:job-server-container:copyLicenses
> Task :runners:flink:1.10:job-server-container:dockerClean UP-TO-DATE
> Task :model:job-management:processResources
> Task :model:fn-execution:processResources
> Task :runners:flink:1.10:copySourceOverrides
> Task :runners:flink:1.10:copyTestResourcesOverrides NO-SOURCE
> Task :sdks:java:core:generateGrammarSource FROM-CACHE
> Task :runners:flink:1.10:processResources
> Task :sdks:java:core:processResources
> Task :sdks:java:build-tools:compileJava FROM-CACHE
> Task :sdks:java:build-tools:processResources
> Task :sdks:java:build-tools:classes
> Task :sdks:java:build-tools:jar
> Task :model:pipeline:extractIncludeProto
> Task :model:pipeline:extractProto
> Task :model:pipeline:generateProto
> Task :model:pipeline:compileJava FROM-CACHE
> Task :model:pipeline:processResources
> Task :model:pipeline:classes
> Task :model:pipeline:jar
> Task :model:fn-execution:extractIncludeProto
> Task :model:job-management:extractIncludeProto
> Task :model:fn-execution:generateProto
> Task :model:job-management:generateProto
> Task :model:fn-execution:compileJava FROM-CACHE
> Task :model:fn-execution:classes
> Task :model:job-management:compileJava FROM-CACHE
> Task :model:job-management:classes
> Task :model:pipeline:shadowJar
> Task :model:job-management:shadowJar
> Task :model:fn-execution:shadowJar
> Task :sdks:java:core:compileJava FROM-CACHE
> Task :sdks:java:core:classes
> Task :sdks:java:core:shadowJar
> Task :vendor:sdks-java-extensions-protobuf:compileJava FROM-CACHE
> Task :vendor:sdks-java-extensions-protobuf:classes UP-TO-DATE
> Task :sdks:java:extensions:protobuf:extractIncludeProto
> Task :sdks:java:extensions:protobuf:generateProto NO-SOURCE
> Task :sdks:java:extensions:protobuf:compileJava FROM-CACHE
> Task :sdks:java:extensions:protobuf:classes UP-TO-DATE
> Task :vendor:sdks-java-extensions-protobuf:shadowJar
> Task :sdks:java:extensions:google-cloud-platform-core:compileJava FROM-CACHE
> Task :sdks:java:extensions:google-cloud-platform-core:classes UP-TO-DATE
> Task :sdks:java:extensions:protobuf:jar
> Task :sdks:java:extensions:google-cloud-platform-core:jar
> Task :runners:core-construction-java:compileJava FROM-CACHE
> Task :runners:core-construction-java:classes UP-TO-DATE
> Task :sdks:java:fn-execution:compileJava FROM-CACHE
> Task :sdks:java:fn-execution:classes UP-TO-DATE
> Task :sdks:java:fn-execution:jar
> Task :runners:core-construction-java:jar
> Task :runners:core-java:compileJava FROM-CACHE
> Task :runners:core-java:classes UP-TO-DATE
> Task :runners:core-java:jar
> Task :sdks:java:harness:compileJava FROM-CACHE
> Task :sdks:java:harness:classes UP-TO-DATE
> Task :sdks:java:harness:jar
> Task :sdks:java:harness:shadowJar
> Task :runners:java-fn-execution:compileJava FROM-CACHE
> Task :runners:java-fn-execution:classes UP-TO-DATE
> Task :runners:java-fn-execution:jar
> Task :sdks:java:expansion-service:compileJava FROM-CACHE
> Task :sdks:java:expansion-service:classes UP-TO-DATE
> Task :sdks:java:expansion-service:jar
> Task :runners:java-job-service:compileJava FROM-CACHE
> Task :runners:java-job-service:classes UP-TO-DATE
> Task :runners:java-job-service:jar
> Task :sdks:java:io:kafka:compileJava FROM-CACHE
> Task :sdks:java:io:kafka:classes UP-TO-DATE
> Task :sdks:java:io:kafka:jar
> Task :runners:flink:1.10:compileJava FROM-CACHE
> Task :runners:flink:1.10:classes
> Task :sdks:java:io:google-cloud-platform:compileJava FROM-CACHE
> Task :sdks:java:io:google-cloud-platform:classes UP-TO-DATE
> Task :runners:flink:1.10:jar
> Task :runners:flink:1.10:job-server:compileJava NO-SOURCE
> Task :runners:flink:1.10:job-server:classes UP-TO-DATE
> Task :sdks:java:io:google-cloud-platform:jar
> Task :runners:flink:1.10:job-server:shadowJar
> Task :runners:flink:1.10:job-server-container:copyDockerfileDependencies
> Task :runners:flink:1.10:job-server-container:dockerPrepare
> Task :runners:flink:1.10:job-server-container:docker

Deprecated Gradle features were used in this build, making it incompatible with Gradle 6.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/5.2.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD SUCCESSFUL in 1m 16s
63 actionable tasks: 43 executed, 19 from cache, 1 up-to-date

Publishing build scan...
https://gradle.com/s/i3x7qneuyaezk

[beam_LoadTests_Python_ParDo_Flink_Batch] $ /bin/bash -xe /tmp/jenkins8458426125884843490.sh
+ echo 'Tagging image...'
Tagging image...
[beam_LoadTests_Python_ParDo_Flink_Batch] $ /bin/bash -xe /tmp/jenkins133143073454119972.sh
+ docker tag gcr.io/apache-beam-testing/beam_portability/beam_flink1.10_job_server:latest gcr.io/apache-beam-testing/beam_portability/beam_flink1.10_job_server:latest
[beam_LoadTests_Python_ParDo_Flink_Batch] $ /bin/bash -xe /tmp/jenkins5874312775949415953.sh
+ echo 'Pushing image...'
Pushing image...
[beam_LoadTests_Python_ParDo_Flink_Batch] $ /bin/bash -xe /tmp/jenkins8498950912935751658.sh
+ docker push gcr.io/apache-beam-testing/beam_portability/beam_flink1.10_job_server:latest
The push refers to repository [gcr.io/apache-beam-testing/beam_portability/beam_flink1.10_job_server]
30fb3bb1628c: Preparing
5be4850ccb32: Preparing
58222c0f232b: Preparing
1cc156cc4c8e: Preparing
48d7de8569a5: Preparing
7d0784c2c563: Preparing
715d0a3d2cc2: Preparing
155d997ed77c: Preparing
88cfc2fcd059: Preparing
760e8d95cf58: Preparing
7cc1c2d7e744: Preparing
8c02234b8605: Preparing
7cc1c2d7e744: Waiting
760e8d95cf58: Waiting
7d0784c2c563: Waiting
8c02234b8605: Waiting
715d0a3d2cc2: Waiting
155d997ed77c: Waiting
30fb3bb1628c: Pushed
58222c0f232b: Pushed
5be4850ccb32: Pushed
7d0784c2c563: Layer already exists
715d0a3d2cc2: Layer already exists
155d997ed77c: Layer already exists
88cfc2fcd059: Layer already exists
760e8d95cf58: Layer already exists
7cc1c2d7e744: Layer already exists
8c02234b8605: Layer already exists
48d7de8569a5: Pushed
1cc156cc4c8e: Pushed
latest: digest: sha256:9925e97fe5791c54d6d1c876e8fcf88cf844548dde7d6340dddcb121689b64c9 size: 2841
[EnvInject] - Injecting environment variables from a build step.
[EnvInject] - Injecting as environment variables the properties content 
JOB_SERVER_IMAGE=gcr.io/apache-beam-testing/beam_portability/beam_flink1.10_job_server:latest
CLUSTER_NAME=beam-loadtests-python-pardo-flink-batch-248
DETACHED_MODE=true
HARNESS_IMAGES_TO_PULL=gcr.io/apache-beam-testing/beam_portability/beam_python3.7_sdk:latest
FLINK_NUM_WORKERS=5
FLINK_DOWNLOAD_URL=https://archive.apache.org/dist/flink/flink-1.10.1/flink-1.10.1-bin-scala_2.11.tgz
GCS_BUCKET=gs://beam-flink-cluster
HADOOP_DOWNLOAD_URL=https://repo.maven.apache.org/maven2/org/apache/flink/flink-shaded-hadoop-2-uber/2.8.3-9.0/flink-shaded-hadoop-2-uber-2.8.3-9.0.jar
FLINK_TASKMANAGER_SLOTS=1
ARTIFACTS_DIR=gs://beam-flink-cluster/beam-loadtests-python-pardo-flink-batch-248
GCLOUD_ZONE=us-central1-a

[EnvInject] - Variables injected successfully.
[beam_LoadTests_Python_ParDo_Flink_Batch] $ /bin/bash -xe /tmp/jenkins28262681519066336.sh
+ echo Setting up flink cluster
Setting up flink cluster
[beam_LoadTests_Python_ParDo_Flink_Batch] $ /bin/bash -xe /tmp/jenkins5195585064628609996.sh
+ cd <https://builds.apache.org/job/beam_LoadTests_Python_ParDo_Flink_Batch/ws/src/.test-infra/dataproc>
+ ./flink_cluster.sh create
+ GCLOUD_ZONE=us-central1-a
+ DATAPROC_VERSION=1.2
+ MASTER_NAME=beam-loadtests-python-pardo-flink-batch-248-m
+ INIT_ACTIONS_FOLDER_NAME=init-actions
+ FLINK_INIT=gs://beam-flink-cluster/init-actions/flink.sh
+ BEAM_INIT=gs://beam-flink-cluster/init-actions/beam.sh
+ DOCKER_INIT=gs://beam-flink-cluster/init-actions/docker.sh
+ FLINK_LOCAL_PORT=8081
+ FLINK_TASKMANAGER_SLOTS=1
+ YARN_APPLICATION_MASTER=
+ create
+ upload_init_actions
+ echo 'Uploading initialization actions to GCS bucket: gs://beam-flink-cluster'
Uploading initialization actions to GCS bucket: gs://beam-flink-cluster
+ gsutil cp -r init-actions/beam.sh init-actions/docker.sh init-actions/flink.sh gs://beam-flink-cluster/init-actions
Copying file://init-actions/beam.sh [Content-Type=text/x-sh]...
/ [0 files][    0.0 B/  2.3 KiB]                                                / [1 files][  2.3 KiB/  2.3 KiB]                                                Copying file://init-actions/docker.sh [Content-Type=text/x-sh]...
/ [1 files][  2.3 KiB/  6.0 KiB]                                                / [2 files][  6.0 KiB/  6.0 KiB]                                                -Copying file://init-actions/flink.sh [Content-Type=text/x-sh]...
- [2 files][  6.0 KiB/ 13.7 KiB]                                                - [3 files][ 13.7 KiB/ 13.7 KiB]                                                
Operation completed over 3 objects/13.7 KiB.                                     
+ create_cluster
+ local metadata=flink-snapshot-url=https://archive.apache.org/dist/flink/flink-1.10.1/flink-1.10.1-bin-scala_2.11.tgz,
+ metadata+=flink-start-yarn-session=true,
+ metadata+=flink-taskmanager-slots=1,
+ metadata+=hadoop-jar-url=https://repo.maven.apache.org/maven2/org/apache/flink/flink-shaded-hadoop-2-uber/2.8.3-9.0/flink-shaded-hadoop-2-uber-2.8.3-9.0.jar
+ [[ -n gcr.io/apache-beam-testing/beam_portability/beam_python3.7_sdk:latest ]]
+ metadata+=,beam-sdk-harness-images-to-pull=gcr.io/apache-beam-testing/beam_portability/beam_python3.7_sdk:latest
+ [[ -n gcr.io/apache-beam-testing/beam_portability/beam_flink1.10_job_server:latest ]]
+ metadata+=,beam-job-server-image=gcr.io/apache-beam-testing/beam_portability/beam_flink1.10_job_server:latest
+ local image_version=1.2
+ echo 'Starting dataproc cluster. Dataproc version: 1.2'
Starting dataproc cluster. Dataproc version: 1.2
+ local num_dataproc_****s=6
+ gcloud dataproc clusters create beam-loadtests-python-pardo-flink-batch-248 --region=global --num-****s=6 --initialization-actions gs://beam-flink-cluster/init-actions/docker.sh,gs://beam-flink-cluster/init-actions/beam.sh,gs://beam-flink-cluster/init-actions/flink.sh --metadata flink-snapshot-url=https://archive.apache.org/dist/flink/flink-1.10.1/flink-1.10.1-bin-scala_2.11.tgz,flink-start-yarn-session=true,flink-taskmanager-slots=1,hadoop-jar-url=https://repo.maven.apache.org/maven2/org/apache/flink/flink-shaded-hadoop-2-uber/2.8.3-9.0/flink-shaded-hadoop-2-uber-2.8.3-9.0.jar,beam-sdk-harness-images-to-pull=gcr.io/apache-beam-testing/beam_portability/beam_python3.7_sdk:latest,beam-job-server-image=gcr.io/apache-beam-testing/beam_portability/beam_flink1.10_job_server:latest, --image-version=1.2 --zone=us-central1-a --quiet
Waiting on operation [projects/apache-beam-testing/regions/global/operations/e832ebce-e231-37db-8501-1b7cbbec5d5c].
Waiting for cluster creation operation...
WARNING: For PD-Standard without local SSDs, we strongly recommend provisioning 1TB or larger to ensure consistently high I/O performance. See https://cloud.google.com/compute/docs/disks/performance for information on disk I/O performance.
.....................................................................................................................................................................................................................................................................done.
ERROR: (gcloud.dataproc.clusters.create) Operation [projects/apache-beam-testing/regions/global/operations/e832ebce-e231-37db-8501-1b7cbbec5d5c] failed: Initialization action failed. Failed action 'gs://beam-flink-cluster/init-actions/flink.sh', see output in: gs://dataproc-6c5fbcbb-a2de-406e-9cf7-8c1ce0b6a604-us/google-cloud-dataproc-metainfo/726b4079-eedf-42b2-907c-3247e6e9b334/beam-loadtests-python-pardo-flink-batch-248-m/dataproc-initialization-script-2_output.
Build step 'Execute shell' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Build failed in Jenkins: beam_LoadTests_Python_ParDo_Flink_Batch #247

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/beam_LoadTests_Python_ParDo_Flink_Batch/247/display/redirect?page=changes>

Changes:

[kmjung] Reject fractional values outside of (0.0, 1.0)

[aramaswamis] [BEAM-10067] Minify website assets with --minify flag

[altay] Add a bare bones powered by page.

[kcweaver] [BEAM-10106] Script the deployment of artifacts to pypi

[kcweaver] Add pypi script to release guide.

[kcweaver] Improve log messages in release scripts.

[kcweaver] Install twine in virtual env.

[chamikaramj] Enables proto holders for testing runner API. Without this x-lang can be

[aromanenko.dev] [BEAM-9916] Fixed some IO Urls on website

[github] Avoid re-encoding value state cells each commit (#11823)

[github] [BEAM-9679] Flatten Kata for Go (#11806)

[github] [BEAM-9220] Go Dataflow jobs to use runner v2 (#11207)

[robertwb] Update multi-language roadmap status.

[daniel.o.programmer] [BEAM-9935] Respect allowed split points and fraction in Go.

[github] [BEAM-8280] Blog post: Python typing changes (#11070)

[github] Update website/www/site/content/en/roadmap/connectors-multi-sdk.md


------------------------------------------
[...truncated 46.89 KB...]
f61899846314: Layer already exists
8c39f7b1a31a: Layer already exists
88cfc2fcd059: Layer already exists
760e8d95cf58: Layer already exists
7cc1c2d7e744: Layer already exists
65e263aaa210: Pushed
8c02234b8605: Layer already exists
ff4e18b8f76f: Pushed
d465e36ea9ec: Pushed
latest: digest: sha256:a1519c28575f95b312a0795de88f4a2c28c6201c611f8c7bb3c841dccb30d57f size: 4938
[Gradle] - Launching build.
[src] $ <https://builds.apache.org/job/beam_LoadTests_Python_ParDo_Flink_Batch/ws/src/gradlew> --continue --max-****s=12 -Dorg.gradle.jvmargs=-Xms2g -Dorg.gradle.jvmargs=-Xmx4g -Pdocker-pull-licenses -Pdocker-repository-root=gcr.io/apache-beam-testing/beam_portability -Pdocker-tag=latest :runners:flink:1.10:job-server-container:docker
> Task :buildSrc:compileJava NO-SOURCE
> Task :buildSrc:compileGroovy UP-TO-DATE
> Task :buildSrc:pluginDescriptors UP-TO-DATE
> Task :buildSrc:processResources UP-TO-DATE
> Task :buildSrc:classes UP-TO-DATE
> Task :buildSrc:jar UP-TO-DATE
> Task :buildSrc:assemble UP-TO-DATE
> Task :buildSrc:spotlessGroovy UP-TO-DATE
> Task :buildSrc:spotlessGroovyCheck UP-TO-DATE
> Task :buildSrc:spotlessGroovyGradle UP-TO-DATE
> Task :buildSrc:spotlessGroovyGradleCheck UP-TO-DATE
> Task :buildSrc:spotlessCheck UP-TO-DATE
> Task :buildSrc:pluginUnderTestMetadata UP-TO-DATE
> Task :buildSrc:compileTestJava NO-SOURCE
> Task :buildSrc:compileTestGroovy NO-SOURCE
> Task :buildSrc:processTestResources NO-SOURCE
> Task :buildSrc:testClasses UP-TO-DATE
> Task :buildSrc:test NO-SOURCE
> Task :buildSrc:validateTaskProperties UP-TO-DATE
> Task :buildSrc:check UP-TO-DATE
> Task :buildSrc:build UP-TO-DATE
Configuration on demand is an incubating feature.
> Task :sdks:java:core:generateAvroProtocol NO-SOURCE
> Task :runners:core-construction-java:processResources NO-SOURCE
> Task :runners:core-java:processResources NO-SOURCE
> Task :sdks:java:fn-execution:processResources NO-SOURCE
> Task :sdks:java:extensions:google-cloud-platform-core:processResources NO-SOURCE
> Task :vendor:sdks-java-extensions-protobuf:processResources NO-SOURCE
> Task :sdks:java:expansion-service:processResources NO-SOURCE
> Task :runners:flink:1.10:copyResourcesOverrides NO-SOURCE
> Task :runners:java-fn-execution:processResources NO-SOURCE
> Task :sdks:java:harness:processResources NO-SOURCE
> Task :runners:java-job-service:processResources NO-SOURCE
> Task :sdks:java:core:generateAvroJava NO-SOURCE
> Task :runners:flink:1.10:job-server:processResources NO-SOURCE
> Task :sdks:java:io:google-cloud-platform:processResources NO-SOURCE
> Task :sdks:java:io:kafka:processResources NO-SOURCE
> Task :model:job-management:extractProto
> Task :model:fn-execution:extractProto
> Task :sdks:java:extensions:protobuf:extractProto
> Task :runners:flink:1.10:job-server-container:copyLicenses
> Task :runners:flink:1.10:job-server-container:dockerClean UP-TO-DATE
> Task :sdks:java:extensions:protobuf:processResources NO-SOURCE
> Task :model:job-management:processResources
> Task :sdks:java:core:generateGrammarSource FROM-CACHE
> Task :runners:flink:1.10:copySourceOverrides
> Task :model:fn-execution:processResources
> Task :runners:flink:1.10:copyTestResourcesOverrides NO-SOURCE
> Task :runners:flink:1.10:processResources
> Task :sdks:java:core:processResources
> Task :sdks:java:build-tools:compileJava FROM-CACHE
> Task :sdks:java:build-tools:processResources
> Task :sdks:java:build-tools:classes
> Task :sdks:java:build-tools:jar
> Task :model:pipeline:extractIncludeProto
> Task :model:pipeline:extractProto
> Task :model:pipeline:generateProto
> Task :model:pipeline:compileJava FROM-CACHE
> Task :model:pipeline:processResources
> Task :model:pipeline:classes
> Task :model:pipeline:jar
> Task :model:job-management:extractIncludeProto
> Task :model:fn-execution:extractIncludeProto
> Task :model:job-management:generateProto
> Task :model:fn-execution:generateProto
> Task :model:job-management:compileJava FROM-CACHE
> Task :model:job-management:classes
> Task :model:fn-execution:compileJava FROM-CACHE
> Task :model:fn-execution:classes
> Task :model:pipeline:shadowJar
> Task :model:job-management:shadowJar
> Task :model:fn-execution:shadowJar
> Task :sdks:java:core:compileJava FROM-CACHE
> Task :sdks:java:core:classes
> Task :sdks:java:core:shadowJar
> Task :sdks:java:fn-execution:compileJava FROM-CACHE
> Task :sdks:java:fn-execution:classes UP-TO-DATE
> Task :runners:core-construction-java:compileJava FROM-CACHE
> Task :runners:core-construction-java:classes UP-TO-DATE
> Task :sdks:java:fn-execution:jar
> Task :vendor:sdks-java-extensions-protobuf:compileJava FROM-CACHE
> Task :vendor:sdks-java-extensions-protobuf:classes UP-TO-DATE
> Task :sdks:java:extensions:protobuf:extractIncludeProto
> Task :sdks:java:extensions:protobuf:generateProto NO-SOURCE
> Task :sdks:java:extensions:google-cloud-platform-core:compileJava FROM-CACHE
> Task :sdks:java:extensions:google-cloud-platform-core:classes UP-TO-DATE
> Task :runners:core-construction-java:jar
> Task :sdks:java:extensions:protobuf:compileJava FROM-CACHE
> Task :sdks:java:extensions:protobuf:classes UP-TO-DATE
> Task :sdks:java:extensions:google-cloud-platform-core:jar
> Task :sdks:java:extensions:protobuf:jar
> Task :vendor:sdks-java-extensions-protobuf:shadowJar
> Task :runners:core-java:compileJava FROM-CACHE
> Task :runners:core-java:classes UP-TO-DATE
> Task :runners:core-java:jar
> Task :sdks:java:harness:compileJava FROM-CACHE
> Task :sdks:java:harness:classes UP-TO-DATE
> Task :sdks:java:harness:jar
> Task :sdks:java:harness:shadowJar
> Task :runners:java-fn-execution:compileJava FROM-CACHE
> Task :runners:java-fn-execution:classes UP-TO-DATE
> Task :runners:java-fn-execution:jar
> Task :sdks:java:expansion-service:compileJava FROM-CACHE
> Task :sdks:java:expansion-service:classes UP-TO-DATE
> Task :sdks:java:expansion-service:jar
> Task :runners:java-job-service:compileJava FROM-CACHE
> Task :runners:java-job-service:classes UP-TO-DATE
> Task :sdks:java:io:kafka:compileJava FROM-CACHE
> Task :sdks:java:io:kafka:classes UP-TO-DATE
> Task :runners:java-job-service:jar
> Task :sdks:java:io:kafka:jar
> Task :runners:flink:1.10:compileJava FROM-CACHE
> Task :runners:flink:1.10:classes
> Task :sdks:java:io:google-cloud-platform:compileJava FROM-CACHE
> Task :sdks:java:io:google-cloud-platform:classes UP-TO-DATE
> Task :runners:flink:1.10:jar
> Task :runners:flink:1.10:job-server:compileJava NO-SOURCE
> Task :runners:flink:1.10:job-server:classes UP-TO-DATE
> Task :sdks:java:io:google-cloud-platform:jar
> Task :runners:flink:1.10:job-server:shadowJar
> Task :runners:flink:1.10:job-server-container:copyDockerfileDependencies
> Task :runners:flink:1.10:job-server-container:dockerPrepare
> Task :runners:flink:1.10:job-server-container:docker

Deprecated Gradle features were used in this build, making it incompatible with Gradle 6.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/5.2.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD SUCCESSFUL in 1m 17s
63 actionable tasks: 43 executed, 19 from cache, 1 up-to-date

Publishing build scan...
https://gradle.com/s/2c4tgaqhxwfku

[beam_LoadTests_Python_ParDo_Flink_Batch] $ /bin/bash -xe /tmp/jenkins2668297313185000374.sh
+ echo 'Tagging image...'
Tagging image...
[beam_LoadTests_Python_ParDo_Flink_Batch] $ /bin/bash -xe /tmp/jenkins2856194617414358203.sh
+ docker tag gcr.io/apache-beam-testing/beam_portability/beam_flink1.10_job_server:latest gcr.io/apache-beam-testing/beam_portability/beam_flink1.10_job_server:latest
[beam_LoadTests_Python_ParDo_Flink_Batch] $ /bin/bash -xe /tmp/jenkins7618664489040920744.sh
+ echo 'Pushing image...'
Pushing image...
[beam_LoadTests_Python_ParDo_Flink_Batch] $ /bin/bash -xe /tmp/jenkins5152367375918491216.sh
+ docker push gcr.io/apache-beam-testing/beam_portability/beam_flink1.10_job_server:latest
The push refers to repository [gcr.io/apache-beam-testing/beam_portability/beam_flink1.10_job_server]
8614200dea02: Preparing
69af269bfd0e: Preparing
1a9d7607a94c: Preparing
d63917cfef81: Preparing
940fd107263b: Preparing
7d0784c2c563: Preparing
715d0a3d2cc2: Preparing
155d997ed77c: Preparing
88cfc2fcd059: Preparing
760e8d95cf58: Preparing
7cc1c2d7e744: Preparing
8c02234b8605: Preparing
88cfc2fcd059: Waiting
760e8d95cf58: Waiting
7cc1c2d7e744: Waiting
715d0a3d2cc2: Waiting
7d0784c2c563: Waiting
155d997ed77c: Waiting
69af269bfd0e: Pushed
1a9d7607a94c: Pushed
8614200dea02: Pushed
7d0784c2c563: Layer already exists
715d0a3d2cc2: Layer already exists
155d997ed77c: Layer already exists
7cc1c2d7e744: Layer already exists
760e8d95cf58: Layer already exists
88cfc2fcd059: Layer already exists
8c02234b8605: Layer already exists
940fd107263b: Pushed
d63917cfef81: Pushed
latest: digest: sha256:39ee348a632c7ef979332d12cf77230db50981556db6e6d75c4fbd5e07f05410 size: 2841
[EnvInject] - Injecting environment variables from a build step.
[EnvInject] - Injecting as environment variables the properties content 
JOB_SERVER_IMAGE=gcr.io/apache-beam-testing/beam_portability/beam_flink1.10_job_server:latest
CLUSTER_NAME=beam-loadtests-python-pardo-flink-batch-247
DETACHED_MODE=true
HARNESS_IMAGES_TO_PULL=gcr.io/apache-beam-testing/beam_portability/beam_python3.7_sdk:latest
FLINK_NUM_WORKERS=5
FLINK_DOWNLOAD_URL=https://archive.apache.org/dist/flink/flink-1.10.1/flink-1.10.1-bin-scala_2.11.tgz
GCS_BUCKET=gs://beam-flink-cluster
HADOOP_DOWNLOAD_URL=https://repo.maven.apache.org/maven2/org/apache/flink/flink-shaded-hadoop-2-uber/2.8.3-9.0/flink-shaded-hadoop-2-uber-2.8.3-9.0.jar
FLINK_TASKMANAGER_SLOTS=1
ARTIFACTS_DIR=gs://beam-flink-cluster/beam-loadtests-python-pardo-flink-batch-247
GCLOUD_ZONE=us-central1-a

[EnvInject] - Variables injected successfully.
[beam_LoadTests_Python_ParDo_Flink_Batch] $ /bin/bash -xe /tmp/jenkins6840499597482900377.sh
+ echo Setting up flink cluster
Setting up flink cluster
[beam_LoadTests_Python_ParDo_Flink_Batch] $ /bin/bash -xe /tmp/jenkins1960694305711928810.sh
+ cd <https://builds.apache.org/job/beam_LoadTests_Python_ParDo_Flink_Batch/ws/src/.test-infra/dataproc>
+ ./flink_cluster.sh create
+ GCLOUD_ZONE=us-central1-a
+ DATAPROC_VERSION=1.2
+ MASTER_NAME=beam-loadtests-python-pardo-flink-batch-247-m
+ INIT_ACTIONS_FOLDER_NAME=init-actions
+ FLINK_INIT=gs://beam-flink-cluster/init-actions/flink.sh
+ BEAM_INIT=gs://beam-flink-cluster/init-actions/beam.sh
+ DOCKER_INIT=gs://beam-flink-cluster/init-actions/docker.sh
+ FLINK_LOCAL_PORT=8081
+ FLINK_TASKMANAGER_SLOTS=1
+ YARN_APPLICATION_MASTER=
+ create
+ upload_init_actions
+ echo 'Uploading initialization actions to GCS bucket: gs://beam-flink-cluster'
Uploading initialization actions to GCS bucket: gs://beam-flink-cluster
+ gsutil cp -r init-actions/beam.sh init-actions/docker.sh init-actions/flink.sh gs://beam-flink-cluster/init-actions
Copying file://init-actions/beam.sh [Content-Type=text/x-sh]...
/ [0 files][    0.0 B/  2.3 KiB]                                                / [1 files][  2.3 KiB/  2.3 KiB]                                                Copying file://init-actions/docker.sh [Content-Type=text/x-sh]...
/ [1 files][  2.3 KiB/  6.0 KiB]                                                / [2 files][  6.0 KiB/  6.0 KiB]                                                Copying file://init-actions/flink.sh [Content-Type=text/x-sh]...
/ [2 files][  6.0 KiB/ 13.7 KiB]                                                / [3 files][ 13.7 KiB/ 13.7 KiB]                                                
Operation completed over 3 objects/13.7 KiB.                                     
+ create_cluster
+ local metadata=flink-snapshot-url=https://archive.apache.org/dist/flink/flink-1.10.1/flink-1.10.1-bin-scala_2.11.tgz,
+ metadata+=flink-start-yarn-session=true,
+ metadata+=flink-taskmanager-slots=1,
+ metadata+=hadoop-jar-url=https://repo.maven.apache.org/maven2/org/apache/flink/flink-shaded-hadoop-2-uber/2.8.3-9.0/flink-shaded-hadoop-2-uber-2.8.3-9.0.jar
+ [[ -n gcr.io/apache-beam-testing/beam_portability/beam_python3.7_sdk:latest ]]
+ metadata+=,beam-sdk-harness-images-to-pull=gcr.io/apache-beam-testing/beam_portability/beam_python3.7_sdk:latest
+ [[ -n gcr.io/apache-beam-testing/beam_portability/beam_flink1.10_job_server:latest ]]
+ metadata+=,beam-job-server-image=gcr.io/apache-beam-testing/beam_portability/beam_flink1.10_job_server:latest
+ local image_version=1.2
+ echo 'Starting dataproc cluster. Dataproc version: 1.2'
Starting dataproc cluster. Dataproc version: 1.2
+ local num_dataproc_****s=6
+ gcloud dataproc clusters create beam-loadtests-python-pardo-flink-batch-247 --region=global --num-****s=6 --initialization-actions gs://beam-flink-cluster/init-actions/docker.sh,gs://beam-flink-cluster/init-actions/beam.sh,gs://beam-flink-cluster/init-actions/flink.sh --metadata flink-snapshot-url=https://archive.apache.org/dist/flink/flink-1.10.1/flink-1.10.1-bin-scala_2.11.tgz,flink-start-yarn-session=true,flink-taskmanager-slots=1,hadoop-jar-url=https://repo.maven.apache.org/maven2/org/apache/flink/flink-shaded-hadoop-2-uber/2.8.3-9.0/flink-shaded-hadoop-2-uber-2.8.3-9.0.jar,beam-sdk-harness-images-to-pull=gcr.io/apache-beam-testing/beam_portability/beam_python3.7_sdk:latest,beam-job-server-image=gcr.io/apache-beam-testing/beam_portability/beam_flink1.10_job_server:latest, --image-version=1.2 --zone=us-central1-a --quiet
Waiting on operation [projects/apache-beam-testing/regions/global/operations/e50e3e28-33df-3464-9a92-b025c6aed0ef].
Waiting for cluster creation operation...
WARNING: For PD-Standard without local SSDs, we strongly recommend provisioning 1TB or larger to ensure consistently high I/O performance. See https://cloud.google.com/compute/docs/disks/performance for information on disk I/O performance.
...................................................................................................................................................................................................................done.
ERROR: (gcloud.dataproc.clusters.create) Operation [projects/apache-beam-testing/regions/global/operations/e50e3e28-33df-3464-9a92-b025c6aed0ef] failed: Initialization action failed. Failed action 'gs://beam-flink-cluster/init-actions/docker.sh', see output in: gs://dataproc-6c5fbcbb-a2de-406e-9cf7-8c1ce0b6a604-us/google-cloud-dataproc-metainfo/5252392d-47aa-4dc3-bdc6-301969eb7057/beam-loadtests-python-pardo-flink-batch-247-m/dataproc-initialization-script-0_output.
Build step 'Execute shell' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


beam_LoadTests_Python_ParDo_Flink_Batch - Build # 246 - Aborted

Posted by Apache Jenkins Server <je...@builds.apache.org>.
The Apache Jenkins build system has built beam_LoadTests_Python_ParDo_Flink_Batch (build #246)

Status: Aborted

Check console output at https://builds.apache.org/job/beam_LoadTests_Python_ParDo_Flink_Batch/246/ to view the results.

beam_LoadTests_Python_ParDo_Flink_Batch - Build # 245 - Aborted

Posted by Apache Jenkins Server <je...@builds.apache.org>.
The Apache Jenkins build system has built beam_LoadTests_Python_ParDo_Flink_Batch (build #245)

Status: Aborted

Check console output at https://builds.apache.org/job/beam_LoadTests_Python_ParDo_Flink_Batch/245/ to view the results.

beam_LoadTests_Python_ParDo_Flink_Batch - Build # 244 - Aborted

Posted by Apache Jenkins Server <je...@builds.apache.org>.
The Apache Jenkins build system has built beam_LoadTests_Python_ParDo_Flink_Batch (build #244)

Status: Aborted

Check console output at https://builds.apache.org/job/beam_LoadTests_Python_ParDo_Flink_Batch/244/ to view the results.

beam_LoadTests_Python_ParDo_Flink_Batch - Build # 243 - Aborted

Posted by Apache Jenkins Server <je...@builds.apache.org>.
The Apache Jenkins build system has built beam_LoadTests_Python_ParDo_Flink_Batch (build #243)

Status: Aborted

Check console output at https://builds.apache.org/job/beam_LoadTests_Python_ParDo_Flink_Batch/243/ to view the results.

beam_LoadTests_Python_ParDo_Flink_Batch - Build # 242 - Aborted

Posted by Apache Jenkins Server <je...@builds.apache.org>.
The Apache Jenkins build system has built beam_LoadTests_Python_ParDo_Flink_Batch (build #242)

Status: Aborted

Check console output at https://builds.apache.org/job/beam_LoadTests_Python_ParDo_Flink_Batch/242/ to view the results.

beam_LoadTests_Python_ParDo_Flink_Batch - Build # 241 - Aborted

Posted by Apache Jenkins Server <je...@builds.apache.org>.
The Apache Jenkins build system has built beam_LoadTests_Python_ParDo_Flink_Batch (build #241)

Status: Aborted

Check console output at https://builds.apache.org/job/beam_LoadTests_Python_ParDo_Flink_Batch/241/ to view the results.

beam_LoadTests_Python_ParDo_Flink_Batch - Build # 240 - Aborted

Posted by Apache Jenkins Server <je...@builds.apache.org>.
The Apache Jenkins build system has built beam_LoadTests_Python_ParDo_Flink_Batch (build #240)

Status: Aborted

Check console output at https://builds.apache.org/job/beam_LoadTests_Python_ParDo_Flink_Batch/240/ to view the results.

beam_LoadTests_Python_ParDo_Flink_Batch - Build # 239 - Aborted

Posted by Apache Jenkins Server <je...@builds.apache.org>.
The Apache Jenkins build system has built beam_LoadTests_Python_ParDo_Flink_Batch (build #239)

Status: Aborted

Check console output at https://builds.apache.org/job/beam_LoadTests_Python_ParDo_Flink_Batch/239/ to view the results.

beam_LoadTests_Python_ParDo_Flink_Batch - Build # 238 - Aborted

Posted by Apache Jenkins Server <je...@builds.apache.org>.
The Apache Jenkins build system has built beam_LoadTests_Python_ParDo_Flink_Batch (build #238)

Status: Aborted

Check console output at https://builds.apache.org/job/beam_LoadTests_Python_ParDo_Flink_Batch/238/ to view the results.

beam_LoadTests_Python_ParDo_Flink_Batch - Build # 237 - Aborted

Posted by Apache Jenkins Server <je...@builds.apache.org>.
The Apache Jenkins build system has built beam_LoadTests_Python_ParDo_Flink_Batch (build #237)

Status: Aborted

Check console output at https://builds.apache.org/job/beam_LoadTests_Python_ParDo_Flink_Batch/237/ to view the results.

beam_LoadTests_Python_ParDo_Flink_Batch - Build # 236 - Aborted

Posted by Apache Jenkins Server <je...@builds.apache.org>.
The Apache Jenkins build system has built beam_LoadTests_Python_ParDo_Flink_Batch (build #236)

Status: Aborted

Check console output at https://builds.apache.org/job/beam_LoadTests_Python_ParDo_Flink_Batch/236/ to view the results.

beam_LoadTests_Python_ParDo_Flink_Batch - Build # 235 - Aborted

Posted by Apache Jenkins Server <je...@builds.apache.org>.
The Apache Jenkins build system has built beam_LoadTests_Python_ParDo_Flink_Batch (build #235)

Status: Aborted

Check console output at https://builds.apache.org/job/beam_LoadTests_Python_ParDo_Flink_Batch/235/ to view the results.

beam_LoadTests_Python_ParDo_Flink_Batch - Build # 234 - Aborted

Posted by Apache Jenkins Server <je...@builds.apache.org>.
The Apache Jenkins build system has built beam_LoadTests_Python_ParDo_Flink_Batch (build #234)

Status: Aborted

Check console output at https://builds.apache.org/job/beam_LoadTests_Python_ParDo_Flink_Batch/234/ to view the results.

beam_LoadTests_Python_ParDo_Flink_Batch - Build # 233 - Aborted

Posted by Apache Jenkins Server <je...@builds.apache.org>.
The Apache Jenkins build system has built beam_LoadTests_Python_ParDo_Flink_Batch (build #233)

Status: Aborted

Check console output at https://builds.apache.org/job/beam_LoadTests_Python_ParDo_Flink_Batch/233/ to view the results.

beam_LoadTests_Python_ParDo_Flink_Batch - Build # 232 - Aborted

Posted by Apache Jenkins Server <je...@builds.apache.org>.
The Apache Jenkins build system has built beam_LoadTests_Python_ParDo_Flink_Batch (build #232)

Status: Aborted

Check console output at https://builds.apache.org/job/beam_LoadTests_Python_ParDo_Flink_Batch/232/ to view the results.

beam_LoadTests_Python_ParDo_Flink_Batch - Build # 231 - Aborted

Posted by Apache Jenkins Server <je...@builds.apache.org>.
The Apache Jenkins build system has built beam_LoadTests_Python_ParDo_Flink_Batch (build #231)

Status: Aborted

Check console output at https://builds.apache.org/job/beam_LoadTests_Python_ParDo_Flink_Batch/231/ to view the results.

Build failed in Jenkins: beam_LoadTests_Python_ParDo_Flink_Batch #230

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/beam_LoadTests_Python_ParDo_Flink_Batch/230/display/redirect?page=changes>

Changes:

[sjvanrossum] [BEAM-9776] Fixes filesystem not found error

[kcweaver] [BEAM-9875] Use FlinkRunner instead of PortableRunner in cross-language

[kcweaver] Remove runner check in xlang wordcount.

[kamil.wasilewski] [BEAM-6710] Add landing page with links to relevant dashboards

[kcweaver] [BEAM-9835] [Portable Spark] Broadcast a PCollection at most once.

[robertwb] [BEAM-9935] Respect allowed split points in Python.

[mxm] [BEAM-9930] Add image to Beam Summit blog post

[github] Add License file for the Go SDK  for pkg.go.dev

[mxm] [BEAM-9947] Store updated ParDoPayload for length-prefixed timer coders

[kamil.wasilewski] Fix InfluxDB credentials binding

[mxm] [BEAM-9947] Add unit test to check length prefixing of timer key

[robertwb] More comments.

[rohde.samuel] Return empty iterator on empty cache

[borzoo.esmailloo] [BEAM-9216] Upgrade katas gradle to version 5.2.1

[github] [BEAM-9887] Throw IllegalArgumentException when building Row with

[klk] [BEAM-9952] Revert "remove PipelineRunner.fromOptions call"

[github] Merge pull request #11637 from Waiting for BQ Query and Export jobs for

[robertwb] Make grpc-requiring class definition optional.

[github] [BEAM-9884] Add options to SqlTransform for specifying planner (#11647)

[daniel.o.programmer] [BEAM-9642] Adding Go SDF fallback for unexpanded SDFs.

[kamil.wasilewski] Fix metrics reporting in HDFS IO IT tests

[mxm] [BEAM-9963] Fix coder type in failing ParDo streaming load test

[github] Merge pull request #11567: [BEAM-8132] Report Python metrics to InfluxDB


------------------------------------------
[...truncated 99.94 KB...]
Collecting filelock<4,>=3.0.0
  Using cached filelock-3.0.12-py3-none-any.whl (7.6 kB)
Collecting pluggy<1,>=0.3.0
  Using cached pluggy-0.13.1-py2.py3-none-any.whl (18 kB)
Collecting six<2,>=1.0.0
  Using cached six-1.14.0-py2.py3-none-any.whl (10 kB)
Collecting virtualenv>=14.0.0
  Using cached virtualenv-20.0.20-py2.py3-none-any.whl (4.7 MB)
Collecting py<2,>=1.4.17
  Using cached py-1.8.1-py2.py3-none-any.whl (83 kB)
Collecting toml>=0.9.4
  Using cached toml-0.10.0-py2.py3-none-any.whl (25 kB)
Collecting protobuf>=3.5.0.post1
  Using cached protobuf-3.11.3-cp37-cp37m-manylinux1_x86_64.whl (1.3 MB)
Collecting grpcio>=1.14.2
  Using cached grpcio-1.28.1-cp37-cp37m-manylinux2010_x86_64.whl (2.8 MB)
Collecting importlib-metadata>=0.12; python_version < "3.8"
  Using cached importlib_metadata-1.6.0-py2.py3-none-any.whl (30 kB)
Collecting appdirs<2,>=1.4.3
  Using cached appdirs-1.4.4-py2.py3-none-any.whl (9.6 kB)
Processing /home/jenkins/.cache/pip/wheels/a2/19/da/a15d4e2bedf3062c739b190d5cb5b7b2ecfbccb6b0d93c861b/distlib-0.3.0-py3-none-any.whl
Collecting zipp>=0.5
  Using cached zipp-3.1.0-py3-none-any.whl (4.9 kB)
Installing collected packages: filelock, zipp, importlib-metadata, pluggy, six, appdirs, distlib, virtualenv, py, toml, tox, protobuf, grpcio, grpcio-tools, future, mypy-protobuf
Successfully installed appdirs-1.4.4 distlib-0.3.0 filelock-3.0.12 future-0.18.2 grpcio-1.28.1 grpcio-tools-1.14.2 importlib-metadata-1.6.0 mypy-protobuf-1.18 pluggy-0.13.1 protobuf-3.11.3 py-1.8.1 six-1.14.0 toml-0.10.0 tox-3.11.1 virtualenv-20.0.20 zipp-3.1.0

> Task :sdks:python:apache_beam:testing:load_tests:installGcpTest
Processing <https://builds.apache.org/job/beam_LoadTests_Python_ParDo_Flink_Batch/ws/src/sdks/python/build/apache-beam.tar.gz>
Processing /home/jenkins/.cache/pip/wheels/dc/9a/e9/49e627353476cec8484343c4ab656f1e0d783ee77b9dde2d1f/crcmod-1.7-cp37-cp37m-linux_x86_64.whl
Processing /home/jenkins/.cache/pip/wheels/a4/61/fd/c57e374e580aa78a45ed78d5859b3a44436af17e22ca53284f/dill-0.3.1.1-py3-none-any.whl
Collecting fastavro<0.24,>=0.21.4
  Using cached fastavro-0.23.3-cp37-cp37m-manylinux2010_x86_64.whl (1.4 MB)
Requirement already satisfied: future<1.0.0,>=0.18.2 in <https://builds.apache.org/job/beam_LoadTests_Python_ParDo_Flink_Batch/ws/src/build/gradleenv/1329484227/lib/python3.7/site-packages> (from apache-beam==2.22.0.dev0) (0.18.2)
Requirement already satisfied: grpcio<2,>=1.12.1 in <https://builds.apache.org/job/beam_LoadTests_Python_ParDo_Flink_Batch/ws/src/build/gradleenv/1329484227/lib/python3.7/site-packages> (from apache-beam==2.22.0.dev0) (1.28.1)
Processing /home/jenkins/.cache/pip/wheels/0a/7d/38/ea4eaf831518e6cd867b515b88919a9785eb66f11def5ab859/hdfs-2.5.8-py3-none-any.whl
Collecting httplib2<0.16.0,>=0.8
  Using cached httplib2-0.15.0-py3-none-any.whl (94 kB)
Collecting mock<3.0.0,>=1.0.1
  Using cached mock-2.0.0-py2.py3-none-any.whl (56 kB)
Collecting numpy<2,>=1.14.3
  Using cached numpy-1.18.4-cp37-cp37m-manylinux1_x86_64.whl (20.2 MB)
Collecting pymongo<4.0.0,>=3.8.0
  Using cached pymongo-3.10.1-cp37-cp37m-manylinux2014_x86_64.whl (462 kB)
Processing /home/jenkins/.cache/pip/wheels/86/73/7a/3b3f76a2142176605ff38fbca574327962c71e25a43197a4c1/oauth2client-3.0.0-py3-none-any.whl
Requirement already satisfied: protobuf<4,>=3.5.0.post1 in <https://builds.apache.org/job/beam_LoadTests_Python_ParDo_Flink_Batch/ws/src/build/gradleenv/1329484227/lib/python3.7/site-packages> (from apache-beam==2.22.0.dev0) (3.11.3)
Collecting pydot<2,>=1.2.0
  Using cached pydot-1.4.1-py2.py3-none-any.whl (19 kB)
Collecting python-dateutil<3,>=2.8.0
  Using cached python_dateutil-2.8.1-py2.py3-none-any.whl (227 kB)
Collecting pytz>=2018.3
  Using cached pytz-2020.1-py2.py3-none-any.whl (510 kB)
Collecting typing-extensions<3.8.0,>=3.7.0
  Using cached typing_extensions-3.7.4.2-py3-none-any.whl (22 kB)
Processing /home/jenkins/.cache/pip/wheels/bc/49/5f/fdb5b9d85055c478213e0158ac122b596816149a02d82e0ab1/avro_python3-1.9.2.1-py3-none-any.whl
Collecting pyarrow<0.17.0,>=0.15.1
  Using cached pyarrow-0.16.0-cp37-cp37m-manylinux2014_x86_64.whl (63.1 MB)
Collecting boto3>=1.9
  Using cached boto3-1.13.7-py2.py3-none-any.whl (128 kB)
Collecting cachetools<4,>=3.1.0
  Using cached cachetools-3.1.1-py2.py3-none-any.whl (11 kB)
Processing /home/jenkins/.cache/pip/wheels/34/3b/69/ecd8e6ae89d9d71102a58962c29faa7a9467ba45f99f205920/google_apitools-0.5.28-py3-none-any.whl
Collecting google-cloud-datastore<1.8.0,>=1.7.1
  Using cached google_cloud_datastore-1.7.4-py2.py3-none-any.whl (82 kB)
Collecting google-cloud-pubsub<1.1.0,>=0.39.0
  Using cached google_cloud_pubsub-1.0.2-py2.py3-none-any.whl (118 kB)
Collecting google-cloud-bigquery<=1.24.0,>=1.6.0
  Using cached google_cloud_bigquery-1.24.0-py2.py3-none-any.whl (165 kB)
Collecting google-cloud-core<2,>=0.28.1
  Using cached google_cloud_core-1.3.0-py2.py3-none-any.whl (26 kB)
Collecting google-cloud-bigtable<1.1.0,>=0.31.1
  Using cached google_cloud_bigtable-1.0.0-py2.py3-none-any.whl (232 kB)
Collecting google-cloud-spanner<1.14.0,>=1.13.0
  Using cached google_cloud_spanner-1.13.0-py2.py3-none-any.whl (212 kB)
Collecting grpcio-gcp<1,>=0.2.2
  Using cached grpcio_gcp-0.2.2-py2.py3-none-any.whl (9.4 kB)
Collecting google-cloud-dlp<=0.13.0,>=0.12.0
  Using cached google_cloud_dlp-0.13.0-py2.py3-none-any.whl (151 kB)
Collecting google-cloud-language<2,>=1.3.0
  Using cached google_cloud_language-1.3.0-py2.py3-none-any.whl (83 kB)
Collecting google-cloud-videointelligence<1.14.0,>=1.8.0
  Using cached google_cloud_videointelligence-1.13.0-py2.py3-none-any.whl (177 kB)
Collecting google-cloud-vision<0.43.0,>=0.38.0
  Using cached google_cloud_vision-0.42.0-py2.py3-none-any.whl (435 kB)
Collecting freezegun>=0.3.12
  Using cached freezegun-0.3.15-py2.py3-none-any.whl (14 kB)
Collecting nose>=1.3.7
  Using cached nose-1.3.7-py3-none-any.whl (154 kB)
Processing /home/jenkins/.cache/pip/wheels/2e/29/a9/431158315f33abeaad2905460f5ffb18fc30f7ed7c66c47dee/nose_xunitmp-0.4.1-py3-none-any.whl
Collecting pandas<0.25,>=0.23.4
  Using cached pandas-0.24.2-cp37-cp37m-manylinux1_x86_64.whl (10.1 MB)
Collecting parameterized<0.8.0,>=0.7.1
  Using cached parameterized-0.7.4-py2.py3-none-any.whl (25 kB)
Collecting pyhamcrest!=1.10.0,<2.0.0,>=1.9
  Using cached PyHamcrest-1.10.1-py3-none-any.whl (48 kB)
Processing /home/jenkins/.cache/pip/wheels/5e/03/1e/e1e954795d6f35dfc7b637fe2277bff021303bd9570ecea653/PyYAML-5.3.1-cp37-cp37m-linux_x86_64.whl
Collecting requests_mock<2.0,>=1.7
  Using cached requests_mock-1.8.0-py2.py3-none-any.whl (23 kB)
Collecting tenacity<6.0,>=5.0.2
  Using cached tenacity-5.1.5-py2.py3-none-any.whl (34 kB)
Collecting pytest<5.0,>=4.4.0
  Using cached pytest-4.6.10-py2.py3-none-any.whl (231 kB)
Collecting pytest-xdist<2,>=1.29.0
  Using cached pytest_xdist-1.32.0-py2.py3-none-any.whl (36 kB)
Collecting pytest-timeout<2,>=1.3.3
  Using cached pytest_timeout-1.3.4-py2.py3-none-any.whl (10 kB)
Requirement already satisfied: six>=1.5.2 in <https://builds.apache.org/job/beam_LoadTests_Python_ParDo_Flink_Batch/ws/src/build/gradleenv/1329484227/lib/python3.7/site-packages> (from grpcio<2,>=1.12.1->apache-beam==2.22.0.dev0) (1.14.0)
Processing /home/jenkins/.cache/pip/wheels/9b/04/dd/7daf4150b6d9b12949298737de9431a324d4b797ffd63f526e/docopt-0.6.2-py2.py3-none-any.whl
Collecting requests>=2.7.0
  Using cached requests-2.23.0-py2.py3-none-any.whl (58 kB)
Collecting pbr>=0.11
  Using cached pbr-5.4.5-py2.py3-none-any.whl (110 kB)
Collecting rsa>=3.1.4
  Using cached rsa-4.0-py2.py3-none-any.whl (38 kB)
Collecting pyasn1>=0.1.7
  Using cached pyasn1-0.4.8-py2.py3-none-any.whl (77 kB)
Collecting pyasn1-modules>=0.0.5
  Using cached pyasn1_modules-0.2.8-py2.py3-none-any.whl (155 kB)
Requirement already satisfied: setuptools in <https://builds.apache.org/job/beam_LoadTests_Python_ParDo_Flink_Batch/ws/src/build/gradleenv/1329484227/lib/python3.7/site-packages> (from protobuf<4,>=3.5.0.post1->apache-beam==2.22.0.dev0) (46.2.0)
Collecting pyparsing>=2.1.4
  Using cached pyparsing-2.4.7-py2.py3-none-any.whl (67 kB)
Collecting jmespath<1.0.0,>=0.7.1
  Using cached jmespath-0.9.5-py2.py3-none-any.whl (24 kB)
Collecting s3transfer<0.4.0,>=0.3.0
  Using cached s3transfer-0.3.3-py2.py3-none-any.whl (69 kB)
Collecting botocore<1.17.0,>=1.16.7
  Using cached botocore-1.16.7-py2.py3-none-any.whl (6.2 MB)
Collecting fasteners>=0.14
  Using cached fasteners-0.15-py2.py3-none-any.whl (23 kB)
Collecting google-api-core[grpc]<2.0.0dev,>=1.6.0
  Using cached google_api_core-1.17.0-py2.py3-none-any.whl (70 kB)
Processing /home/jenkins/.cache/pip/wheels/b9/ee/67/2e444183030cb8d31ce8b34cee34a7afdbd3ba5959ea846380/grpc_google_iam_v1-0.12.3-py3-none-any.whl
Collecting google-resumable-media<0.6dev,>=0.5.0
  Using cached google_resumable_media-0.5.0-py2.py3-none-any.whl (38 kB)
Collecting google-auth<2.0dev,>=1.9.0
  Using cached google_auth-1.14.3-py2.py3-none-any.whl (89 kB)
Requirement already satisfied: importlib-metadata>=0.12; python_version < "3.8" in <https://builds.apache.org/job/beam_LoadTests_Python_ParDo_Flink_Batch/ws/src/build/gradleenv/1329484227/lib/python3.7/site-packages> (from pytest<5.0,>=4.4.0->apache-beam==2.22.0.dev0) (1.6.0)
Collecting packaging
  Using cached packaging-20.3-py2.py3-none-any.whl (37 kB)
Requirement already satisfied: pluggy<1.0,>=0.12 in <https://builds.apache.org/job/beam_LoadTests_Python_ParDo_Flink_Batch/ws/src/build/gradleenv/1329484227/lib/python3.7/site-packages> (from pytest<5.0,>=4.4.0->apache-beam==2.22.0.dev0) (0.13.1)
Collecting wcwidth
  Using cached wcwidth-0.1.9-py2.py3-none-any.whl (19 kB)
Collecting attrs>=17.4.0
  Using cached attrs-19.3.0-py2.py3-none-any.whl (39 kB)
Collecting atomicwrites>=1.0
  Using cached atomicwrites-1.4.0-py2.py3-none-any.whl (6.8 kB)
Collecting more-itertools>=4.0.0; python_version > "2.7"
  Using cached more_itertools-8.2.0-py3-none-any.whl (43 kB)
Requirement already satisfied: py>=1.5.0 in <https://builds.apache.org/job/beam_LoadTests_Python_ParDo_Flink_Batch/ws/src/build/gradleenv/1329484227/lib/python3.7/site-packages> (from pytest<5.0,>=4.4.0->apache-beam==2.22.0.dev0) (1.8.1)
Collecting pytest-forked
  Using cached pytest_forked-1.1.3-py2.py3-none-any.whl (4.5 kB)
Collecting execnet>=1.1
  Using cached execnet-1.7.1-py2.py3-none-any.whl (39 kB)
Collecting certifi>=2017.4.17
  Using cached certifi-2020.4.5.1-py2.py3-none-any.whl (157 kB)
Collecting chardet<4,>=3.0.2
  Using cached chardet-3.0.4-py2.py3-none-any.whl (133 kB)
Collecting urllib3!=1.25.0,!=1.25.1,<1.26,>=1.21.1
  Using cached urllib3-1.25.9-py2.py3-none-any.whl (126 kB)
Collecting idna<3,>=2.5
  Using cached idna-2.9-py2.py3-none-any.whl (58 kB)
Collecting docutils<0.16,>=0.10
  Using cached docutils-0.15.2-py3-none-any.whl (547 kB)
Collecting monotonic>=0.1
  Using cached monotonic-1.5-py2.py3-none-any.whl (5.3 kB)
Processing /home/jenkins/.cache/pip/wheels/4c/a1/71/5e427276ceeff277fd76878d1b19fbf4587a2845015d86864b/googleapis_common_protos-1.51.0-py3-none-any.whl
Requirement already satisfied: zipp>=0.5 in <https://builds.apache.org/job/beam_LoadTests_Python_ParDo_Flink_Batch/ws/src/build/gradleenv/1329484227/lib/python3.7/site-packages> (from importlib-metadata>=0.12; python_version < "3.8"->pytest<5.0,>=4.4.0->apache-beam==2.22.0.dev0) (3.1.0)
Collecting apipkg>=1.4
  Using cached apipkg-1.5-py2.py3-none-any.whl (4.9 kB)
Building wheels for collected packages: apache-beam
  Building wheel for apache-beam (setup.py): started
  Building wheel for apache-beam (setup.py): finished with status 'done'
  Created wheel for apache-beam: filename=apache_beam-2.22.0.dev0-py3-none-any.whl size=2056563 sha256=cf9701cd98778f92a1abaa36910ca72938bd8a54bac0c04efb0ac37cb66a387f
  Stored in directory: /home/jenkins/.cache/pip/wheels/90/d8/b7/51c85dc301f777ed726fe78499a1e5dc02890cd8ebeaac231e
Successfully built apache-beam
Installing collected packages: crcmod, dill, pytz, fastavro, docopt, certifi, chardet, urllib3, idna, requests, hdfs, httplib2, pbr, mock, numpy, pymongo, pyasn1, rsa, pyasn1-modules, oauth2client, pyparsing, pydot, python-dateutil, typing-extensions, avro-python3, pyarrow, jmespath, docutils, botocore, s3transfer, boto3, cachetools, monotonic, fasteners, google-apitools, googleapis-common-protos, google-auth, google-api-core, google-cloud-core, google-cloud-datastore, grpc-google-iam-v1, google-cloud-pubsub, google-resumable-media, google-cloud-bigquery, google-cloud-bigtable, google-cloud-spanner, grpcio-gcp, google-cloud-dlp, google-cloud-language, google-cloud-videointelligence, google-cloud-vision, freezegun, nose, nose-xunitmp, pandas, parameterized, pyhamcrest, pyyaml, requests-mock, tenacity, packaging, wcwidth, attrs, atomicwrites, more-itertools, pytest, pytest-forked, apipkg, execnet, pytest-xdist, pytest-timeout, apache-beam
Successfully installed apache-beam-2.22.0.dev0 apipkg-1.5 atomicwrites-1.4.0 attrs-19.3.0 avro-python3-1.9.2.1 boto3-1.13.7 botocore-1.16.7 cachetools-3.1.1 certifi-2020.4.5.1 chardet-3.0.4 crcmod-1.7 dill-0.3.1.1 docopt-0.6.2 docutils-0.15.2 execnet-1.7.1 fastavro-0.23.3 fasteners-0.15 freezegun-0.3.15 google-api-core-1.17.0 google-apitools-0.5.28 google-auth-1.14.3 google-cloud-bigquery-1.24.0 google-cloud-bigtable-1.0.0 google-cloud-core-1.3.0 google-cloud-datastore-1.7.4 google-cloud-dlp-0.13.0 google-cloud-language-1.3.0 google-cloud-pubsub-1.0.2 google-cloud-spanner-1.13.0 google-cloud-videointelligence-1.13.0 google-cloud-vision-0.42.0 google-resumable-media-0.5.0 googleapis-common-protos-1.51.0 grpc-google-iam-v1-0.12.3 grpcio-gcp-0.2.2 hdfs-2.5.8 httplib2-0.15.0 idna-2.9 jmespath-0.9.5 mock-2.0.0 monotonic-1.5 more-itertools-8.2.0 nose-1.3.7 nose-xunitmp-0.4.1 numpy-1.18.4 oauth2client-3.0.0 packaging-20.3 pandas-0.24.2 parameterized-0.7.4 pbr-5.4.5 pyarrow-0.16.0 pyasn1-0.4.8 pyasn1-modules-0.2.8 pydot-1.4.1 pyhamcrest-1.10.1 pymongo-3.10.1 pyparsing-2.4.7 pytest-4.6.10 pytest-forked-1.1.3 pytest-timeout-1.3.4 pytest-xdist-1.32.0 python-dateutil-2.8.1 pytz-2020.1 pyyaml-5.3.1 requests-2.23.0 requests-mock-1.8.0 rsa-4.0 s3transfer-0.3.3 tenacity-5.1.5 typing-extensions-3.7.4.2 urllib3-1.25.9 wcwidth-0.1.9

> Task :sdks:python:apache_beam:testing:load_tests:run
WARNING:root:Make sure that locally built Python SDK docker image has Python 3.7 interpreter.
INFO:root:Using Python SDK docker image: apache/beam_python3.7_sdk:2.22.0.dev. If the image is not available at local, we will try to pull from hub.docker.com
INFO:apache_beam.runners.portability.fn_api_runner.translations:==================== <function lift_combiners at 0x7f7f29f06d08> ====================
WARNING:apache_beam.options.pipeline_options:Discarding unparseable args: ['--iterations=10', '--number_of_counter_operations=0', '--number_of_counters=0']
ERROR:grpc._channel:Exception iterating requests!
Traceback (most recent call last):
  File "<https://builds.apache.org/job/beam_LoadTests_Python_ParDo_Flink_Batch/ws/src/build/gradleenv/1329484227/lib/python3.7/site-packages/grpc/_channel.py",> line 195, in consume_request_iterator
    request = next(request_iterator)
  File "<https://builds.apache.org/job/beam_LoadTests_Python_ParDo_Flink_Batch/ws/src/sdks/python/apache_beam/runners/portability/artifact_service.py",> line 317, in __next__
    raise self._queue.get()
  File "<https://builds.apache.org/job/beam_LoadTests_Python_ParDo_Flink_Batch/ws/src/sdks/python/apache_beam/runners/portability/portable_runner.py",> line 194, in stage
    return self._stage_via_portable_service(channel, staging_session_token)
  File "<https://builds.apache.org/job/beam_LoadTests_Python_ParDo_Flink_Batch/ws/src/sdks/python/apache_beam/runners/portability/portable_runner.py",> line 212, in _stage_via_portable_service
    staging_session_token)
  File "<https://builds.apache.org/job/beam_LoadTests_Python_ParDo_Flink_Batch/ws/src/sdks/python/apache_beam/runners/portability/artifact_service.py",> line 507, in offer_artifacts
    for request in requests:
  File "<https://builds.apache.org/job/beam_LoadTests_Python_ParDo_Flink_Batch/ws/src/build/gradleenv/1329484227/lib/python3.7/site-packages/grpc/_channel.py",> line 416, in __next__
    return self._next()
  File "<https://builds.apache.org/job/beam_LoadTests_Python_ParDo_Flink_Batch/ws/src/build/gradleenv/1329484227/lib/python3.7/site-packages/grpc/_channel.py",> line 706, in _next
    raise self
grpc._channel._MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
	status = StatusCode.UNIMPLEMENTED
	details = "Method not found: org.apache.beam.model.job_management.v1.ArtifactStagingService/ReverseArtifactRetrievalService"
	debug_error_string = "{"created":"@1589292227.924233274","description":"Error received from peer ipv4:127.0.0.1:8098","file":"src/core/lib/surface/call.cc","file_line":1056,"grpc_message":"Method not found: org.apache.beam.model.job_management.v1.ArtifactStagingService/ReverseArtifactRetrievalService","grpc_status":12}"
>
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STOPPED
INFO:apache_beam.runners.portability.portable_runner:Job state changed to STARTING
INFO:apache_beam.runners.portability.portable_runner:Job state changed to RUNNING
ERROR:root:java.util.concurrent.TimeoutException
INFO:apache_beam.runners.portability.portable_runner:Job state changed to FAILED
Traceback (most recent call last):
  File "/usr/lib/python3.7/runpy.py", line 193, in _run_module_as_main
    "__main__", mod_spec)
  File "/usr/lib/python3.7/runpy.py", line 85, in _run_code
    exec(code, run_globals)
  File "<https://builds.apache.org/job/beam_LoadTests_Python_ParDo_Flink_Batch/ws/src/sdks/python/apache_beam/testing/load_tests/pardo_test.py",> line 225, in <module>
    ParDoTest().run()
  File "<https://builds.apache.org/job/beam_LoadTests_Python_ParDo_Flink_Batch/ws/src/sdks/python/apache_beam/testing/load_tests/load_test.py",> line 145, in run
    self.result.wait_until_finish(duration=self.timeout_ms)
  File "<https://builds.apache.org/job/beam_LoadTests_Python_ParDo_Flink_Batch/ws/src/sdks/python/apache_beam/runners/portability/portable_runner.py",> line 583, in wait_until_finish
    raise self._runtime_exception
RuntimeError: Pipeline load-tests-python-flink-batch-pardo-1-0512131546_d87e610a-4534-44ce-96bc-7f9f663f3fef failed in state FAILED: java.util.concurrent.TimeoutException

> Task :sdks:python:apache_beam:testing:load_tests:run FAILED

FAILURE: Build failed with an exception.

* Where:
Build file '<https://builds.apache.org/job/beam_LoadTests_Python_ParDo_Flink_Batch/ws/src/sdks/python/apache_beam/testing/load_tests/build.gradle'> line: 51

* What went wrong:
Execution failed for task ':sdks:python:apache_beam:testing:load_tests:run'.
> error occurred

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 6.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/5.2.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 5m 55s
5 actionable tasks: 4 executed, 1 up-to-date

Publishing build scan...
https://gradle.com/s/iw7gxuefqo7j6

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


beam_LoadTests_Python_ParDo_Flink_Batch - Build # 229 - Aborted

Posted by Apache Jenkins Server <je...@builds.apache.org>.
The Apache Jenkins build system has built beam_LoadTests_Python_ParDo_Flink_Batch (build #229)

Status: Aborted

Check console output at https://builds.apache.org/job/beam_LoadTests_Python_ParDo_Flink_Batch/229/ to view the results.