You are viewing a plain text version of this content. The canonical link for it is here.
Posted to builds@beam.apache.org by Apache Jenkins Server <je...@builds.apache.org> on 2019/10/14 21:38:29 UTC

Build failed in Jenkins: beam_PostCommit_Py_VR_Dataflow #4801

See <https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/4801/display/redirect?page=changes>

Changes:

[ryan] [BEAM-7073]: Add unit test for Avro logical type datum.

[thw] [BEAM-8387] Remove sdk-worker-parallelism option from JobServerDriver


------------------------------------------
[...truncated 331.56 KB...]
 name: 'beamapp-jenkins-1014201420-509623'
 projectId: 'apache-beam-testing'
 stageStates: []
 startTime: '2019-10-14T20:14:41.299958Z'
 steps: []
 tempFiles: []
 type: TypeValueValuesEnum(JOB_TYPE_BATCH, 1)>
root: INFO: Created job with id: [2019-10-14_13_14_39-5771395795173614073]
root: INFO: To access the Dataflow monitoring console, please navigate to https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-14_13_14_39-5771395795173614073?project=apache-beam-testing
root: INFO: Job 2019-10-14_13_14_39-5771395795173614073 is in state JOB_STATE_RUNNING
root: INFO: 2019-10-14T20:14:39.858Z: JOB_MESSAGE_DETAILED: Autoscaling is enabled for job 2019-10-14_13_14_39-5771395795173614073. The number of workers will be between 1 and 1000.
root: INFO: 2019-10-14T20:14:39.858Z: JOB_MESSAGE_DETAILED: Autoscaling was automatically enabled for job 2019-10-14_13_14_39-5771395795173614073.
root: INFO: 2019-10-14T20:14:44.120Z: JOB_MESSAGE_DETAILED: Checking permissions granted to controller Service Account.
root: INFO: 2019-10-14T20:14:45.090Z: JOB_MESSAGE_BASIC: Worker configuration: n1-standard-1 in us-central1-a.
root: INFO: 2019-10-14T20:14:45.670Z: JOB_MESSAGE_DETAILED: Expanding CoGroupByKey operations into optimizable parts.
root: INFO: 2019-10-14T20:14:45.712Z: JOB_MESSAGE_DEBUG: Combiner lifting skipped for step assert:even/Group/GroupByKey: GroupByKey not followed by a combiner.
root: INFO: 2019-10-14T20:14:45.742Z: JOB_MESSAGE_DEBUG: Combiner lifting skipped for step assert:odd/Group/GroupByKey: GroupByKey not followed by a combiner.
root: INFO: 2019-10-14T20:14:45.780Z: JOB_MESSAGE_DEBUG: Combiner lifting skipped for step assert_that/Group/GroupByKey: GroupByKey not followed by a combiner.
root: INFO: 2019-10-14T20:14:45.835Z: JOB_MESSAGE_DETAILED: Expanding GroupByKey operations into optimizable parts.
root: INFO: 2019-10-14T20:14:45.866Z: JOB_MESSAGE_DETAILED: Lifting ValueCombiningMappingFns into MergeBucketsMappingFns
root: INFO: 2019-10-14T20:14:45.997Z: JOB_MESSAGE_DEBUG: Annotating graph with Autotuner information.
root: INFO: 2019-10-14T20:14:46.057Z: JOB_MESSAGE_DETAILED: Fusing adjacent ParDo, Read, Write, and Flatten operations
root: INFO: 2019-10-14T20:14:46.095Z: JOB_MESSAGE_DETAILED: Unzipping flatten s8 for input s6.out
root: INFO: 2019-10-14T20:14:46.140Z: JOB_MESSAGE_DETAILED: Fusing unzipped copy of assert_that/Group/GroupByKey/Reify, through flatten assert_that/Group/Flatten, into producer assert_that/Group/pair_with_0
root: INFO: 2019-10-14T20:14:46.175Z: JOB_MESSAGE_DETAILED: Fusing consumer assert_that/Group/GroupByKey/GroupByWindow into assert_that/Group/GroupByKey/Read
root: INFO: 2019-10-14T20:14:46.210Z: JOB_MESSAGE_DETAILED: Fusing consumer assert_that/Group/Map(_merge_tagged_vals_under_key) into assert_that/Group/GroupByKey/GroupByWindow
root: INFO: 2019-10-14T20:14:46.237Z: JOB_MESSAGE_DETAILED: Fusing consumer assert_that/Unkey into assert_that/Group/Map(_merge_tagged_vals_under_key)
root: INFO: 2019-10-14T20:14:46.263Z: JOB_MESSAGE_DETAILED: Fusing consumer assert_that/Match into assert_that/Unkey
root: INFO: 2019-10-14T20:14:46.294Z: JOB_MESSAGE_DETAILED: Unzipping flatten s18 for input s16.out
root: INFO: 2019-10-14T20:14:46.329Z: JOB_MESSAGE_DETAILED: Fusing unzipped copy of assert:odd/Group/GroupByKey/Reify, through flatten assert:odd/Group/Flatten, into producer assert:odd/Group/pair_with_0
root: INFO: 2019-10-14T20:14:46.366Z: JOB_MESSAGE_DETAILED: Fusing consumer assert:odd/Group/GroupByKey/GroupByWindow into assert:odd/Group/GroupByKey/Read
root: INFO: 2019-10-14T20:14:46.407Z: JOB_MESSAGE_DETAILED: Fusing consumer assert:odd/Group/Map(_merge_tagged_vals_under_key) into assert:odd/Group/GroupByKey/GroupByWindow
root: INFO: 2019-10-14T20:14:46.441Z: JOB_MESSAGE_DETAILED: Fusing consumer assert:odd/Unkey into assert:odd/Group/Map(_merge_tagged_vals_under_key)
root: INFO: 2019-10-14T20:14:46.477Z: JOB_MESSAGE_DETAILED: Fusing consumer assert:odd/Match into assert:odd/Unkey
root: INFO: 2019-10-14T20:14:46.514Z: JOB_MESSAGE_DETAILED: Unzipping flatten s28 for input s26.out
root: INFO: 2019-10-14T20:14:46.551Z: JOB_MESSAGE_DETAILED: Fusing unzipped copy of assert:even/Group/GroupByKey/Reify, through flatten assert:even/Group/Flatten, into producer assert:even/Group/pair_with_0
root: INFO: 2019-10-14T20:14:46.585Z: JOB_MESSAGE_DETAILED: Fusing consumer assert:even/Group/GroupByKey/GroupByWindow into assert:even/Group/GroupByKey/Read
root: INFO: 2019-10-14T20:14:46.621Z: JOB_MESSAGE_DETAILED: Fusing consumer assert:even/Group/Map(_merge_tagged_vals_under_key) into assert:even/Group/GroupByKey/GroupByWindow
root: INFO: 2019-10-14T20:14:46.660Z: JOB_MESSAGE_DETAILED: Fusing consumer assert:even/Unkey into assert:even/Group/Map(_merge_tagged_vals_under_key)
root: INFO: 2019-10-14T20:14:46.693Z: JOB_MESSAGE_DETAILED: Fusing consumer assert:even/Match into assert:even/Unkey
root: INFO: 2019-10-14T20:14:46.728Z: JOB_MESSAGE_DETAILED: Unzipping flatten s8-u31 for input s9-reify-value18-c29
root: INFO: 2019-10-14T20:14:46.766Z: JOB_MESSAGE_DETAILED: Fusing unzipped copy of assert_that/Group/GroupByKey/Write, through flatten assert_that/Group/Flatten/Unzipped-1, into producer assert_that/Group/GroupByKey/Reify
root: INFO: 2019-10-14T20:14:46.803Z: JOB_MESSAGE_DETAILED: Unzipping flatten s18-u36 for input s19-reify-value0-c34
root: INFO: 2019-10-14T20:14:46.841Z: JOB_MESSAGE_DETAILED: Fusing unzipped copy of assert:odd/Group/GroupByKey/Write, through flatten assert:odd/Group/Flatten/Unzipped-1, into producer assert:odd/Group/GroupByKey/Reify
root: INFO: 2019-10-14T20:14:46.875Z: JOB_MESSAGE_DETAILED: Unzipping flatten s28-u41 for input s29-reify-value9-c39
root: INFO: 2019-10-14T20:14:46.910Z: JOB_MESSAGE_DETAILED: Fusing unzipped copy of assert:even/Group/GroupByKey/Write, through flatten assert:even/Group/Flatten/Unzipped-1, into producer assert:even/Group/GroupByKey/Reify
root: INFO: 2019-10-14T20:14:46.944Z: JOB_MESSAGE_DETAILED: Fusing consumer assert_that/Group/GroupByKey/Reify into assert_that/Group/pair_with_1
root: INFO: 2019-10-14T20:14:46.982Z: JOB_MESSAGE_DETAILED: Fusing consumer assert:odd/Group/GroupByKey/Reify into assert:odd/Group/pair_with_1
root: INFO: 2019-10-14T20:14:47.016Z: JOB_MESSAGE_DETAILED: Fusing consumer assert:even/Group/GroupByKey/Reify into assert:even/Group/pair_with_1
root: INFO: 2019-10-14T20:14:47.052Z: JOB_MESSAGE_DETAILED: Fusing consumer assert_that/Group/GroupByKey/Write into assert_that/Group/GroupByKey/Reify
root: INFO: 2019-10-14T20:14:47.089Z: JOB_MESSAGE_DETAILED: Fusing consumer assert:odd/Group/GroupByKey/Write into assert:odd/Group/GroupByKey/Reify
root: INFO: 2019-10-14T20:14:47.132Z: JOB_MESSAGE_DETAILED: Fusing consumer assert:even/Group/GroupByKey/Write into assert:even/Group/GroupByKey/Reify
root: INFO: 2019-10-14T20:14:47.166Z: JOB_MESSAGE_DETAILED: Fusing consumer ClassifyNumbers/FlatMap(<lambda at ptransform_test.py:279>) into Some Numbers/Read
root: INFO: 2019-10-14T20:14:47.202Z: JOB_MESSAGE_DETAILED: Fusing consumer assert_that/WindowInto(WindowIntoFn) into ClassifyNumbers/FlatMap(<lambda at ptransform_test.py:279>)
root: INFO: 2019-10-14T20:14:47.240Z: JOB_MESSAGE_DETAILED: Fusing consumer assert_that/ToVoidKey into assert_that/WindowInto(WindowIntoFn)
root: INFO: 2019-10-14T20:14:47.278Z: JOB_MESSAGE_DETAILED: Fusing consumer assert_that/Group/pair_with_1 into assert_that/ToVoidKey
root: INFO: 2019-10-14T20:14:47.310Z: JOB_MESSAGE_DETAILED: Fusing consumer assert:odd/WindowInto(WindowIntoFn) into ClassifyNumbers/FlatMap(<lambda at ptransform_test.py:279>)
root: INFO: 2019-10-14T20:14:47.346Z: JOB_MESSAGE_DETAILED: Fusing consumer assert:odd/ToVoidKey into assert:odd/WindowInto(WindowIntoFn)
root: INFO: 2019-10-14T20:14:47.383Z: JOB_MESSAGE_DETAILED: Fusing consumer assert:odd/Group/pair_with_1 into assert:odd/ToVoidKey
root: INFO: 2019-10-14T20:14:47.415Z: JOB_MESSAGE_DETAILED: Fusing consumer assert:even/WindowInto(WindowIntoFn) into ClassifyNumbers/FlatMap(<lambda at ptransform_test.py:279>)
root: INFO: 2019-10-14T20:14:47.453Z: JOB_MESSAGE_DETAILED: Fusing consumer assert:even/ToVoidKey into assert:even/WindowInto(WindowIntoFn)
root: INFO: 2019-10-14T20:14:47.489Z: JOB_MESSAGE_DETAILED: Fusing consumer assert:even/Group/pair_with_1 into assert:even/ToVoidKey
root: INFO: 2019-10-14T20:14:47.525Z: JOB_MESSAGE_DETAILED: Fusing consumer assert_that/Group/pair_with_0 into assert_that/Create/Read
root: INFO: 2019-10-14T20:14:47.561Z: JOB_MESSAGE_DETAILED: Fusing consumer assert:odd/Group/pair_with_0 into assert:odd/Create/Read
root: INFO: 2019-10-14T20:14:47.599Z: JOB_MESSAGE_DETAILED: Fusing consumer assert:even/Group/pair_with_0 into assert:even/Create/Read
root: INFO: 2019-10-14T20:14:47.639Z: JOB_MESSAGE_DEBUG: Workflow config is missing a default resource spec.
root: INFO: 2019-10-14T20:14:47.673Z: JOB_MESSAGE_DEBUG: Adding StepResource setup and teardown to workflow graph.
root: INFO: 2019-10-14T20:14:47.697Z: JOB_MESSAGE_DEBUG: Adding workflow start and stop steps.
root: INFO: 2019-10-14T20:14:47.726Z: JOB_MESSAGE_DEBUG: Assigning stage ids.
root: INFO: 2019-10-14T20:14:47.901Z: JOB_MESSAGE_DEBUG: Executing wait step start57
root: INFO: 2019-10-14T20:14:47.984Z: JOB_MESSAGE_BASIC: Executing operation assert:odd/Group/GroupByKey/Create
root: INFO: 2019-10-14T20:14:48.007Z: JOB_MESSAGE_BASIC: Executing operation assert:even/Group/GroupByKey/Create
root: INFO: 2019-10-14T20:14:48.028Z: JOB_MESSAGE_BASIC: Executing operation assert_that/Group/GroupByKey/Create
root: INFO: 2019-10-14T20:14:48.036Z: JOB_MESSAGE_DEBUG: Starting worker pool setup.
root: INFO: 2019-10-14T20:14:48.067Z: JOB_MESSAGE_BASIC: Starting 1 workers in us-central1-a...
root: INFO: 2019-10-14T20:14:48.119Z: JOB_MESSAGE_BASIC: Finished operation assert:even/Group/GroupByKey/Create
root: INFO: 2019-10-14T20:14:48.119Z: JOB_MESSAGE_BASIC: Finished operation assert:odd/Group/GroupByKey/Create
root: INFO: 2019-10-14T20:14:48.139Z: JOB_MESSAGE_BASIC: Finished operation assert_that/Group/GroupByKey/Create
root: INFO: 2019-10-14T20:14:48.204Z: JOB_MESSAGE_DEBUG: Value "assert:even/Group/GroupByKey/Session" materialized.
root: INFO: 2019-10-14T20:14:48.237Z: JOB_MESSAGE_DEBUG: Value "assert:odd/Group/GroupByKey/Session" materialized.
root: INFO: 2019-10-14T20:14:48.270Z: JOB_MESSAGE_DEBUG: Value "assert_that/Group/GroupByKey/Session" materialized.
root: INFO: 2019-10-14T20:14:48.308Z: JOB_MESSAGE_BASIC: Executing operation assert:even/Create/Read+assert:even/Group/pair_with_0+assert:even/Group/GroupByKey/Reify+assert:even/Group/GroupByKey/Write
root: INFO: 2019-10-14T20:14:48.341Z: JOB_MESSAGE_BASIC: Executing operation assert:odd/Create/Read+assert:odd/Group/pair_with_0+assert:odd/Group/GroupByKey/Reify+assert:odd/Group/GroupByKey/Write
root: INFO: 2019-10-14T20:14:48.383Z: JOB_MESSAGE_BASIC: Executing operation Some Numbers/Read+ClassifyNumbers/FlatMap(<lambda at ptransform_test.py:279>)+assert_that/WindowInto(WindowIntoFn)+assert_that/ToVoidKey+assert_that/Group/pair_with_1+assert_that/Group/GroupByKey/Reify+assert_that/Group/GroupByKey/Write+assert:odd/WindowInto(WindowIntoFn)+assert:odd/ToVoidKey+assert:odd/Group/pair_with_1+assert:odd/Group/GroupByKey/Reify+assert:odd/Group/GroupByKey/Write+assert:even/WindowInto(WindowIntoFn)+assert:even/ToVoidKey+assert:even/Group/pair_with_1+assert:even/Group/GroupByKey/Reify+assert:even/Group/GroupByKey/Write
root: INFO: 2019-10-14T20:14:48.421Z: JOB_MESSAGE_BASIC: Executing operation assert_that/Create/Read+assert_that/Group/pair_with_0+assert_that/Group/GroupByKey/Reify+assert_that/Group/GroupByKey/Write
root: INFO: 2019-10-14T20:15:43.897Z: JOB_MESSAGE_DETAILED: Autoscaling: Raised the number of workers to 1 based on the rate of progress in the currently running step(s).
root: INFO: 2019-10-14T20:20:47.870Z: JOB_MESSAGE_DETAILED: Checking permissions granted to controller Service Account.
root: INFO: 2019-10-14T20:26:47.871Z: JOB_MESSAGE_DETAILED: Checking permissions granted to controller Service Account.
root: INFO: 2019-10-14T20:32:47.870Z: JOB_MESSAGE_DETAILED: Checking permissions granted to controller Service Account.
root: INFO: 2019-10-14T20:38:47.869Z: JOB_MESSAGE_DETAILED: Checking permissions granted to controller Service Account.
root: INFO: 2019-10-14T20:44:47.871Z: JOB_MESSAGE_DETAILED: Checking permissions granted to controller Service Account.
root: INFO: 2019-10-14T20:50:47.871Z: JOB_MESSAGE_DETAILED: Checking permissions granted to controller Service Account.
root: INFO: 2019-10-14T20:56:47.871Z: JOB_MESSAGE_DETAILED: Checking permissions granted to controller Service Account.
root: INFO: 2019-10-14T21:02:47.871Z: JOB_MESSAGE_DETAILED: Checking permissions granted to controller Service Account.
oauth2client.transport: INFO: Refreshing due to a 401 (attempt 1/2)
root: INFO: 2019-10-14T21:08:47.871Z: JOB_MESSAGE_DETAILED: Checking permissions granted to controller Service Account.
root: INFO: 2019-10-14T21:14:47.872Z: JOB_MESSAGE_DETAILED: Checking permissions granted to controller Service Account.
root: INFO: 2019-10-14T21:14:48.315Z: JOB_MESSAGE_ERROR: Workflow failed. Causes: The Dataflow job appears to be stuck because no worker activity has been seen in the last 1h. You can get help with Cloud Dataflow at https://cloud.google.com/dataflow/support.
root: INFO: 2019-10-14T21:14:48.391Z: JOB_MESSAGE_BASIC: Finished operation assert:odd/Create/Read+assert:odd/Group/pair_with_0+assert:odd/Group/GroupByKey/Reify+assert:odd/Group/GroupByKey/Write
root: INFO: 2019-10-14T21:14:48.391Z: JOB_MESSAGE_BASIC: Finished operation Some Numbers/Read+ClassifyNumbers/FlatMap(<lambda at ptransform_test.py:279>)+assert_that/WindowInto(WindowIntoFn)+assert_that/ToVoidKey+assert_that/Group/pair_with_1+assert_that/Group/GroupByKey/Reify+assert_that/Group/GroupByKey/Write+assert:odd/WindowInto(WindowIntoFn)+assert:odd/ToVoidKey+assert:odd/Group/pair_with_1+assert:odd/Group/GroupByKey/Reify+assert:odd/Group/GroupByKey/Write+assert:even/WindowInto(WindowIntoFn)+assert:even/ToVoidKey+assert:even/Group/pair_with_1+assert:even/Group/GroupByKey/Reify+assert:even/Group/GroupByKey/Write
root: INFO: 2019-10-14T21:14:48.391Z: JOB_MESSAGE_BASIC: Finished operation assert_that/Create/Read+assert_that/Group/pair_with_0+assert_that/Group/GroupByKey/Reify+assert_that/Group/GroupByKey/Write
root: INFO: 2019-10-14T21:14:48.391Z: JOB_MESSAGE_BASIC: Finished operation assert:even/Create/Read+assert:even/Group/pair_with_0+assert:even/Group/GroupByKey/Reify+assert:even/Group/GroupByKey/Write
root: INFO: 2019-10-14T21:14:49.229Z: JOB_MESSAGE_BASIC: Cancel request is committed for workflow job: 2019-10-14_13_14_39-5771395795173614073.
root: INFO: 2019-10-14T21:14:49.297Z: JOB_MESSAGE_DETAILED: Cleaning up.
root: INFO: 2019-10-14T21:14:49.372Z: JOB_MESSAGE_DEBUG: Starting worker pool teardown.
root: INFO: 2019-10-14T21:14:49.411Z: JOB_MESSAGE_BASIC: Stopping worker pool...
root: INFO: 2019-10-14T21:17:22.258Z: JOB_MESSAGE_DETAILED: Autoscaling: Reduced the number of workers to 0 based on the rate of progress in the currently running step(s).
root: INFO: 2019-10-14T21:17:22.316Z: JOB_MESSAGE_BASIC: Worker pool stopped.
root: INFO: 2019-10-14T21:17:22.360Z: JOB_MESSAGE_DEBUG: Tearing down pending resources...
root: INFO: Job 2019-10-14_13_14_39-5771395795173614073 is in state JOB_STATE_FAILED
--------------------- >> end captured logging << ---------------------
Worker logs: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-14_13_14_44-12661526682762187097?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-14_13_24_54-12598938204034769073?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-14_13_34_55-17298426923065809685?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-14_13_14_35-6478929536224418886?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-14_13_25_47-6910725676609943852?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-14_13_14_40-12009602037238183477?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-14_13_26_47-4806846231304398713?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-14_13_14_39-5771395795173614073?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-14_13_14_39-7980207058636388723?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-14_13_25_26-1524212388157053701?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-14_13_14_39-658617651123779132?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-14_13_25_57-2488446788883629011?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-14_13_14_36-14361867570950140937?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-14_13_25_14-17322079386108356391?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-14_13_34_54-9724721480441205428?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-14_13_14_40-1819225869438994303?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-14_13_25_49-17139839072750714370?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-14_13_35_02-15038640266593975827?project=apache-beam-testing

----------------------------------------------------------------------
XML: nosetests-validatesRunnerBatchTests-df-py37.xml
----------------------------------------------------------------------
XML: <https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/sdks/python/nosetests.xml>
----------------------------------------------------------------------
Ran 18 tests in 3794.255s

FAILED (errors=1)

> Task :sdks:python:test-suites:dataflow:py37:validatesRunnerBatchTests FAILED

> Task :sdks:python:test-suites:dataflow:py37:validatesRunnerStreamingTests
>>> RUNNING integration tests with pipeline options: --runner=TestDataflowRunner --project=apache-beam-testing --staging_location=gs://temp-storage-for-end-to-end-tests/staging-it --temp_location=gs://temp-storage-for-end-to-end-tests/temp-it --output=gs://temp-storage-for-end-to-end-tests/py-it-cloud/output --sdk_location=<https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/sdks/python/build/apache-beam.tar.gz> --requirements_file=postcommit_requirements.txt --num_workers=1 --sleep_secs=20 --streaming --dataflow_worker_jar=<https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/runners/google-cloud-dataflow-java/worker/build/libs/beam-runners-google-cloud-dataflow-java-fn-api-worker-2.17.0-SNAPSHOT.jar> --kms_key_name=projects/apache-beam-testing/locations/global/keyRings/beam-it/cryptoKeys/test --dataflow_kms_key=projects/apache-beam-testing/locations/global/keyRings/beam-it/cryptoKeys/test
>>>   test options: --nocapture --processes=8 --process-timeout=4500 --attr=ValidatesRunner,!sickbay-streaming
running nosetests
running egg_info
writing apache_beam.egg-info/PKG-INFO
writing dependency_links to apache_beam.egg-info/dependency_links.txt
writing entry points to apache_beam.egg-info/entry_points.txt
writing requirements to apache_beam.egg-info/requires.txt
writing top-level names to apache_beam.egg-info/top_level.txt
reading manifest file 'apache_beam.egg-info/SOURCES.txt'
reading manifest template 'MANIFEST.in'
<https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/build/gradleenv/-1734967052/lib/python3.7/site-packages/setuptools/dist.py>:475: UserWarning: Normalizing '2.17.0.dev' to '2.17.0.dev0'
  normalized_version,
warning: no files found matching 'README.md'
warning: no files found matching 'NOTICE'
warning: no files found matching 'LICENSE'
writing manifest file 'apache_beam.egg-info/SOURCES.txt'
<https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/sdks/python/apache_beam/io/gcp/datastore/v1/datastoreio.py>:59: UserWarning: Datastore IO will support Python 3 after replacing googledatastore by google-cloud-datastore, see: BEAM-4543.
  warnings.warn('Datastore IO will support Python 3 after replacing '
<https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/sdks/python/apache_beam/io/vcfio.py>:47: UserWarning: VCF IO will support Python 3 after migration to Nucleus, see: BEAM-5628.
  warnings.warn("VCF IO will support Python 3 after migration to Nucleus, "
<https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/sdks/python/apache_beam/transforms/trigger_test.py>:507: YAMLLoadWarning: calling yaml.load_all() without Loader=... is deprecated, as the default Loader is unsafe. Please read https://msg.pyyaml.org/load for full details.
  for spec in yaml.load_all(open(transcript_filename)):
Worker logs: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-14_14_17_52-6593858964136817786?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-14_14_26_23-5635550943099801468?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-14_14_17_44-3499783740621008568?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-14_14_26_52-16968509698413509894?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-14_14_17_45-8453553745200395865?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-14_14_27_13-1391123897141169872?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-14_14_17_45-3420042616741947407?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-14_14_28_19-13327863003940585631?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-14_14_17_46-1637413389164463865?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-14_14_27_34-476145138525196871?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-14_14_17_46-11272495258406733432?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-14_14_27_35-3175558006625033298?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-14_14_17_46-7216147549683633331?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-14_14_27_18-18313416973371238208?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-14_14_17_44-8708131222441908122?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-10-14_14_27_17-18179666474318702473?project=apache-beam-testing
Runs streaming Dataflow job and verifies that user metrics are reported ... ok
test_dofn_lifecycle (apache_beam.transforms.dofn_lifecycle_test.DoFnLifecycleTest) ... ok
test_flatten_multiple_pcollections_having_multiple_consumers (apache_beam.transforms.ptransform_test.PTransformTest) ... ok
test_as_dict_twice (apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok
test_undeclared_outputs (apache_beam.transforms.ptransform_test.PTransformTest) ... ok
test_par_do_with_multiple_outputs_and_using_yield (apache_beam.transforms.ptransform_test.PTransformTest) ... ok
test_par_do_with_multiple_outputs_and_using_return (apache_beam.transforms.ptransform_test.PTransformTest) ... ok
test_multiple_empty_outputs (apache_beam.transforms.ptransform_test.PTransformTest) ... ok
test_as_list_twice (apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok
test_as_list_and_as_dict_side_inputs (apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok
test_default_value_singleton_side_input (apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok
test_as_singleton_without_unique_labels (apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok
test_flattened_side_input (apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok
test_empty_singleton_side_input (apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok
test_as_singleton_with_different_defaults (apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok
test_iterable_side_input (apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok

----------------------------------------------------------------------
XML: nosetests-validatesRunnerStreamingTests-df-py37.xml
----------------------------------------------------------------------
XML: <https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/sdks/python/nosetests.xml>
----------------------------------------------------------------------
Ran 16 tests in 1255.326s

OK

FAILURE: Build completed with 2 failures.

1: Task failed with an exception.
-----------
* Where:
Build file '<https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/sdks/python/test-suites/dataflow/py36/build.gradle'> line: 101

* What went wrong:
Execution failed for task ':sdks:python:test-suites:dataflow:py36:validatesRunnerStreamingTests'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

2: Task failed with an exception.
-----------
* Where:
Build file '<https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/sdks/python/test-suites/dataflow/py37/build.gradle'> line: 111

* What went wrong:
Execution failed for task ':sdks:python:test-suites:dataflow:py37:validatesRunnerBatchTests'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 6.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/5.2.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 1h 25m 27s
74 actionable tasks: 58 executed, 16 from cache

Publishing build scan...
https://gradle.com/s/dqsivek3m77uq

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Jenkins build is back to normal : beam_PostCommit_Py_VR_Dataflow #4802

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/4802/display/redirect?page=changes>


---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org