You are viewing a plain text version of this content. The canonical link for it is here.
Posted to builds@beam.apache.org by Apache Jenkins Server <je...@builds.apache.org> on 2020/12/26 19:12:43 UTC

Build failed in Jenkins: beam_PostCommit_Python38 #694

See <https://ci-beam.apache.org/job/beam_PostCommit_Python38/694/display/redirect>

Changes:


------------------------------------------
[...truncated 38.42 MB...]
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/-1734967051/lib/python3.8/site-packages/apitools/base/py/base_api.py",> line 603, in __ProcessHttpResponse
    raise exceptions.HttpError.FromResponse(
apitools.base.py.exceptions.HttpError: HttpError accessing <https://www.googleapis.com/resumable/upload/storage/v1/b/temp-storage-for-end-to-end-tests/o?alt=json&name=staging-it%2Fbeamapp-jenkins-1226180516-751989.1609005916.752353%2Fdataflow-worker.jar&uploadType=resumable&upload_id=ABg5-UwhWaKIYRDWC7HM1RD0i7gIDVOivPzuIvavU0Unfqo8_aJvy311EYHKAQvUXj6C_t3_CajtYE1wPpf-y3byC1G5RDcH9Q>: response: <{'content-type': 'text/plain; charset=utf-8', 'x-guploader-uploadid': 'ABg5-UwhWaKIYRDWC7HM1RD0i7gIDVOivPzuIvavU0Unfqo8_aJvy311EYHKAQvUXj6C_t3_CajtYE1wPpf-y3byC1G5RDcH9Q', 'content-length': '0', 'date': 'Sat, 26 Dec 2020 18:05:34 GMT', 'server': 'UploadServer', 'status': '503'}>, content <>
-------------------- >> begin captured logging << --------------------
apache_beam.internal.gcp.auth: INFO: Setting socket default timeout to 60 seconds.
apache_beam.internal.gcp.auth: INFO: socket default timeout is 60.0 seconds.
root: DEBUG: Connecting using Google Application Default Credentials.
oauth2client.transport: INFO: Attempting refresh to obtain initial access_token
apache_beam.io.filesystem: DEBUG: translate_pattern: 'gs://dataflow-samples/shakespeare/kinglear.txt' -> 'gs://dataflow\\-samples/shakespeare/kinglear\\.txt'
root: DEBUG: Unhandled type_constraint: Union[]
root: DEBUG: Unhandled type_constraint: Union[]
root: DEBUG: Unhandled type_constraint: Union[]
root: DEBUG: Unhandled type_constraint: Union[]
root: DEBUG: Unhandled type_constraint: Union[]
root: DEBUG: Unhandled type_constraint: Union[]
root: DEBUG: Unhandled type_constraint: Union[]
root: DEBUG: Unhandled type_constraint: Union[]
root: DEBUG: Unhandled type_constraint: Union[]
apache_beam.runners.portability.stager: INFO: Executing command: ['<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/-1734967051/bin/python',> '-m', 'pip', 'download', '--dest', '/tmp/dataflow-requirements-cache', '-r', 'postcommit_requirements.txt', '--exists-action', 'i', '--no-binary', ':all:']
apache_beam.runners.portability.stager: INFO: Copying Beam SDK "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/build/apache-beam.tar.gz"> to staging location.
root: WARNING: Make sure that locally built Python SDK docker image has Python 3.8 interpreter.
root: INFO: Default Python SDK image for environment is apache/beam_python3.8_sdk:2.28.0.dev
root: INFO: Using provided Python SDK container image: gcr.io/cloud-dataflow/v1beta3/python38:beam-master-20201214
root: INFO: Python SDK container image set to "gcr.io/cloud-dataflow/v1beta3/python38:beam-master-20201214" for Docker environment
apache_beam.runners.portability.fn_api_runner.translations: INFO: ==================== <function eliminate_common_key_with_none at 0x7fa07f3d78b0> ====================
apache_beam.runners.portability.fn_api_runner.translations: DEBUG: 28 [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]
apache_beam.runners.portability.fn_api_runner.translations: DEBUG: Stages: ['ref_AppliedPTransform_ReadFromText/Read/Impulse_4\n  ReadFromText/Read/Impulse:beam:transform:impulse:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_ReadFromText/Read/Map(<lambda at iobase.py:899>)_5\n  ReadFromText/Read/Map(<lambda at iobase.py:899>):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_ReadFromText/Read/SDFBoundedSourceReader/ParDo(SDFBoundedSourceDoFn)_7\n  ReadFromText/Read/SDFBoundedSourceReader/ParDo(SDFBoundedSourceDoFn):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_TopPerPrefix/PerElement/PerElement:PairWithVoid_10\n  TopPerPrefix/PerElement/PerElement:PairWithVoid:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_TopPerPrefix/PerElement/CombinePerKey(CountCombineFn)_11\n  TopPerPrefix/PerElement/CombinePerKey(CountCombineFn):beam:transform:combine_per_key:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_TopPerPrefix/FlatMap(extract_prefixes)_14\n  TopPerPrefix/FlatMap(extract_prefixes):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_TopPerPrefix/LargestPerKey(10)/TopPerKey(10)/CombinePerKey(TopCombineFn)_17\n  TopPerPrefix/LargestPerKey(10)/TopPerKey(10)/CombinePerKey(TopCombineFn):beam:transform:combine_per_key:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Map(<lambda at autocomplete_test.py:67>)_20\n  Map(<lambda at autocomplete_test.py:67>):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Map(<lambda at autocomplete_test.py:70>)_21\n  Map(<lambda at autocomplete_test.py:70>):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_CombineGlobally(sum)/CombinePerKey_24\n  CombineGlobally(sum)/CombinePerKey:beam:transform:combine_per_key:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_CombineGlobally(sum)/UnKey_27\n  CombineGlobally(sum)/UnKey:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_CombineGlobally(sum)/DoOnce/Impulse_29\n  CombineGlobally(sum)/DoOnce/Impulse:beam:transform:impulse:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_CombineGlobally(sum)/DoOnce/FlatMap(<lambda at core.py:2957>)_30\n  CombineGlobally(sum)/DoOnce/FlatMap(<lambda at core.py:2957>):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_CombineGlobally(sum)/DoOnce/Map(decode)_32\n  CombineGlobally(sum)/DoOnce/Map(decode):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_CombineGlobally(sum)/InjectDefault_33\n  CombineGlobally(sum)/InjectDefault:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Create/Impulse_36\n  assert_that/Create/Impulse:beam:transform:impulse:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Create/FlatMap(<lambda at core.py:2957>)_37\n  assert_that/Create/FlatMap(<lambda at core.py:2957>):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Create/Map(decode)_39\n  assert_that/Create/Map(decode):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/WindowInto(WindowIntoFn)_40\n  assert_that/WindowInto(WindowIntoFn):beam:transform:window_into:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/ToVoidKey_41\n  assert_that/ToVoidKey:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Group/pair_with_0_43\n  assert_that/Group/pair_with_0:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Group/pair_with_1_44\n  assert_that/Group/pair_with_1:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Group/Flatten_45\n  assert_that/Group/Flatten:beam:transform:flatten:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Group/GroupByKey_46\n  assert_that/Group/GroupByKey:beam:transform:group_by_key:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Group/Map(_merge_tagged_vals_under_key)_47\n  assert_that/Group/Map(_merge_tagged_vals_under_key):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Unkey_48\n  assert_that/Unkey:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Match_49\n  assert_that/Match:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_CombineGlobally(sum)/KeyWithVoid_23\n  CombineGlobally(sum)/KeyWithVoid:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>']
apache_beam.runners.portability.fn_api_runner.translations: INFO: ==================== <function pack_combiners at 0x7fa07f3d7940> ====================
apache_beam.runners.portability.fn_api_runner.translations: DEBUG: 28 [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]
apache_beam.runners.portability.fn_api_runner.translations: DEBUG: Stages: ['ref_AppliedPTransform_ReadFromText/Read/Impulse_4\n  ReadFromText/Read/Impulse:beam:transform:impulse:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_ReadFromText/Read/Map(<lambda at iobase.py:899>)_5\n  ReadFromText/Read/Map(<lambda at iobase.py:899>):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_ReadFromText/Read/SDFBoundedSourceReader/ParDo(SDFBoundedSourceDoFn)_7\n  ReadFromText/Read/SDFBoundedSourceReader/ParDo(SDFBoundedSourceDoFn):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_TopPerPrefix/PerElement/PerElement:PairWithVoid_10\n  TopPerPrefix/PerElement/PerElement:PairWithVoid:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_TopPerPrefix/FlatMap(extract_prefixes)_14\n  TopPerPrefix/FlatMap(extract_prefixes):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Map(<lambda at autocomplete_test.py:67>)_20\n  Map(<lambda at autocomplete_test.py:67>):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Map(<lambda at autocomplete_test.py:70>)_21\n  Map(<lambda at autocomplete_test.py:70>):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_CombineGlobally(sum)/UnKey_27\n  CombineGlobally(sum)/UnKey:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_CombineGlobally(sum)/DoOnce/Impulse_29\n  CombineGlobally(sum)/DoOnce/Impulse:beam:transform:impulse:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_CombineGlobally(sum)/DoOnce/FlatMap(<lambda at core.py:2957>)_30\n  CombineGlobally(sum)/DoOnce/FlatMap(<lambda at core.py:2957>):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_CombineGlobally(sum)/DoOnce/Map(decode)_32\n  CombineGlobally(sum)/DoOnce/Map(decode):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_CombineGlobally(sum)/InjectDefault_33\n  CombineGlobally(sum)/InjectDefault:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Create/Impulse_36\n  assert_that/Create/Impulse:beam:transform:impulse:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Create/FlatMap(<lambda at core.py:2957>)_37\n  assert_that/Create/FlatMap(<lambda at core.py:2957>):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Create/Map(decode)_39\n  assert_that/Create/Map(decode):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/WindowInto(WindowIntoFn)_40\n  assert_that/WindowInto(WindowIntoFn):beam:transform:window_into:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/ToVoidKey_41\n  assert_that/ToVoidKey:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Group/pair_with_0_43\n  assert_that/Group/pair_with_0:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Group/pair_with_1_44\n  assert_that/Group/pair_with_1:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Group/Flatten_45\n  assert_that/Group/Flatten:beam:transform:flatten:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Group/GroupByKey_46\n  assert_that/Group/GroupByKey:beam:transform:group_by_key:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Group/Map(_merge_tagged_vals_under_key)_47\n  assert_that/Group/Map(_merge_tagged_vals_under_key):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Unkey_48\n  assert_that/Unkey:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Match_49\n  assert_that/Match:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_CombineGlobally(sum)/KeyWithVoid_23\n  CombineGlobally(sum)/KeyWithVoid:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_TopPerPrefix/PerElement/CombinePerKey(CountCombineFn)_11\n  TopPerPrefix/PerElement/CombinePerKey(CountCombineFn):beam:transform:combine_per_key:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_TopPerPrefix/LargestPerKey(10)/TopPerKey(10)/CombinePerKey(TopCombineFn)_17\n  TopPerPrefix/LargestPerKey(10)/TopPerKey(10)/CombinePerKey(TopCombineFn):beam:transform:combine_per_key:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_CombineGlobally(sum)/CombinePerKey_24\n  CombineGlobally(sum)/CombinePerKey:beam:transform:combine_per_key:v1\n  must follow: \n  downstream_side_inputs: <unknown>']
apache_beam.runners.portability.fn_api_runner.translations: INFO: ==================== <function sort_stages at 0x7fa07f3d6160> ====================
apache_beam.runners.portability.fn_api_runner.translations: DEBUG: 28 [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]
apache_beam.runners.portability.fn_api_runner.translations: DEBUG: Stages: ['ref_AppliedPTransform_ReadFromText/Read/Impulse_4\n  ReadFromText/Read/Impulse:beam:transform:impulse:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_ReadFromText/Read/Map(<lambda at iobase.py:899>)_5\n  ReadFromText/Read/Map(<lambda at iobase.py:899>):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_ReadFromText/Read/SDFBoundedSourceReader/ParDo(SDFBoundedSourceDoFn)_7\n  ReadFromText/Read/SDFBoundedSourceReader/ParDo(SDFBoundedSourceDoFn):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_TopPerPrefix/PerElement/PerElement:PairWithVoid_10\n  TopPerPrefix/PerElement/PerElement:PairWithVoid:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_TopPerPrefix/PerElement/CombinePerKey(CountCombineFn)_11\n  TopPerPrefix/PerElement/CombinePerKey(CountCombineFn):beam:transform:combine_per_key:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_TopPerPrefix/FlatMap(extract_prefixes)_14\n  TopPerPrefix/FlatMap(extract_prefixes):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_TopPerPrefix/LargestPerKey(10)/TopPerKey(10)/CombinePerKey(TopCombineFn)_17\n  TopPerPrefix/LargestPerKey(10)/TopPerKey(10)/CombinePerKey(TopCombineFn):beam:transform:combine_per_key:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Map(<lambda at autocomplete_test.py:67>)_20\n  Map(<lambda at autocomplete_test.py:67>):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_Map(<lambda at autocomplete_test.py:70>)_21\n  Map(<lambda at autocomplete_test.py:70>):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_CombineGlobally(sum)/KeyWithVoid_23\n  CombineGlobally(sum)/KeyWithVoid:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_CombineGlobally(sum)/CombinePerKey_24\n  CombineGlobally(sum)/CombinePerKey:beam:transform:combine_per_key:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_CombineGlobally(sum)/UnKey_27\n  CombineGlobally(sum)/UnKey:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_CombineGlobally(sum)/DoOnce/Impulse_29\n  CombineGlobally(sum)/DoOnce/Impulse:beam:transform:impulse:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_CombineGlobally(sum)/DoOnce/FlatMap(<lambda at core.py:2957>)_30\n  CombineGlobally(sum)/DoOnce/FlatMap(<lambda at core.py:2957>):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_CombineGlobally(sum)/DoOnce/Map(decode)_32\n  CombineGlobally(sum)/DoOnce/Map(decode):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_CombineGlobally(sum)/InjectDefault_33\n  CombineGlobally(sum)/InjectDefault:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Create/Impulse_36\n  assert_that/Create/Impulse:beam:transform:impulse:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Create/FlatMap(<lambda at core.py:2957>)_37\n  assert_that/Create/FlatMap(<lambda at core.py:2957>):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Create/Map(decode)_39\n  assert_that/Create/Map(decode):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/WindowInto(WindowIntoFn)_40\n  assert_that/WindowInto(WindowIntoFn):beam:transform:window_into:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/ToVoidKey_41\n  assert_that/ToVoidKey:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Group/pair_with_0_43\n  assert_that/Group/pair_with_0:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Group/pair_with_1_44\n  assert_that/Group/pair_with_1:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Group/Flatten_45\n  assert_that/Group/Flatten:beam:transform:flatten:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Group/GroupByKey_46\n  assert_that/Group/GroupByKey:beam:transform:group_by_key:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Group/Map(_merge_tagged_vals_under_key)_47\n  assert_that/Group/Map(_merge_tagged_vals_under_key):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Unkey_48\n  assert_that/Unkey:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_assert_that/Match_49\n  assert_that/Match:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>']
apache_beam.runners.dataflow.dataflow_runner: WARNING: Typical end users should not use this worker jar feature. It can only be used when FnAPI is enabled.
apache_beam.io.filesystem: DEBUG: translate_pattern: 'gs://dataflow-samples/shakespeare/kinglear.txt' -> 'gs://dataflow\\-samples/shakespeare/kinglear\\.txt'
apache_beam.runners.dataflow.internal.apiclient: INFO: Starting GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-1226180516-751989.1609005916.752353/pipeline.pb...
apache_beam.runners.dataflow.internal.apiclient: INFO: Completed GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-1226180516-751989.1609005916.752353/pipeline.pb in 0 seconds.
apache_beam.runners.dataflow.internal.apiclient: INFO: Starting GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-1226180516-751989.1609005916.752353/requirements.txt...
apache_beam.runners.dataflow.internal.apiclient: INFO: Completed GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-1226180516-751989.1609005916.752353/requirements.txt in 0 seconds.
apache_beam.runners.dataflow.internal.apiclient: INFO: Starting GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-1226180516-751989.1609005916.752353/parameterized-0.7.4.tar.gz...
apache_beam.runners.dataflow.internal.apiclient: INFO: Completed GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-1226180516-751989.1609005916.752353/parameterized-0.7.4.tar.gz in 0 seconds.
apache_beam.runners.dataflow.internal.apiclient: INFO: Starting GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-1226180516-751989.1609005916.752353/pbr-5.5.1.tar.gz...
apache_beam.runners.dataflow.internal.apiclient: INFO: Completed GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-1226180516-751989.1609005916.752353/pbr-5.5.1.tar.gz in 0 seconds.
apache_beam.runners.dataflow.internal.apiclient: INFO: Starting GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-1226180516-751989.1609005916.752353/mock-2.0.0.tar.gz...
apache_beam.runners.dataflow.internal.apiclient: INFO: Completed GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-1226180516-751989.1609005916.752353/mock-2.0.0.tar.gz in 0 seconds.
apache_beam.runners.dataflow.internal.apiclient: INFO: Starting GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-1226180516-751989.1609005916.752353/six-1.15.0.tar.gz...
apache_beam.runners.dataflow.internal.apiclient: INFO: Completed GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-1226180516-751989.1609005916.752353/six-1.15.0.tar.gz in 0 seconds.
apache_beam.runners.dataflow.internal.apiclient: INFO: Starting GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-1226180516-751989.1609005916.752353/funcsigs-1.0.2.tar.gz...
apache_beam.runners.dataflow.internal.apiclient: INFO: Completed GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-1226180516-751989.1609005916.752353/funcsigs-1.0.2.tar.gz in 0 seconds.
apache_beam.runners.dataflow.internal.apiclient: INFO: Starting GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-1226180516-751989.1609005916.752353/pbr-5.4.5.tar.gz...
apache_beam.runners.dataflow.internal.apiclient: INFO: Completed GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-1226180516-751989.1609005916.752353/pbr-5.4.5.tar.gz in 0 seconds.
apache_beam.runners.dataflow.internal.apiclient: INFO: Starting GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-1226180516-751989.1609005916.752353/pbr-5.5.0.tar.gz...
apache_beam.runners.dataflow.internal.apiclient: INFO: Completed GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-1226180516-751989.1609005916.752353/pbr-5.5.0.tar.gz in 0 seconds.
apache_beam.runners.dataflow.internal.apiclient: INFO: Starting GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-1226180516-751989.1609005916.752353/PyHamcrest-1.10.1.tar.gz...
apache_beam.runners.dataflow.internal.apiclient: INFO: Completed GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-1226180516-751989.1609005916.752353/PyHamcrest-1.10.1.tar.gz in 0 seconds.
apache_beam.runners.dataflow.internal.apiclient: INFO: Starting GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-1226180516-751989.1609005916.752353/dataflow_python_sdk.tar...
apache_beam.runners.dataflow.internal.apiclient: INFO: Completed GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-1226180516-751989.1609005916.752353/dataflow_python_sdk.tar in 0 seconds.
apache_beam.runners.dataflow.internal.apiclient: INFO: Starting GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-1226180516-751989.1609005916.752353/dataflow-worker.jar...
--------------------- >> end captured logging << ---------------------

======================================================================
ERROR: test_datastore_wordcount_it (apache_beam.examples.cookbook.datastore_wordcount_it_test.DatastoreWordCountIT)
----------------------------------------------------------------------
Traceback (most recent call last):
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/examples/cookbook/datastore_wordcount_it_test.py",> line 73, in test_datastore_wordcount_it
    datastore_wordcount.run(
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/examples/cookbook/datastore_wordcount.py",> line 252, in run
    result = read_from_datastore(project, known_args, pipeline_options)
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/examples/cookbook/datastore_wordcount.py",> line 196, in read_from_datastore
    result = p.run()
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/pipeline.py",> line 529, in run
    return Pipeline.from_runner_api(
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/pipeline.py",> line 561, in run
    return self.runner.run_pipeline(self, self._options)
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/runners/dataflow/test_dataflow_runner.py",> line 56, in run_pipeline
    self.result = super(TestDataflowRunner,
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/runners/dataflow/dataflow_runner.py",> line 621, in run_pipeline
    self.dataflow_client.create_job(self.job), self)
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/utils/retry.py",> line 260, in wrapper
    return fun(*args, **kwargs)
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/runners/dataflow/internal/apiclient.py",> line 655, in create_job
    self.create_job_description(job)
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/runners/dataflow/internal/apiclient.py",> line 711, in create_job_description
    resources = self._stage_resources(job.proto_pipeline, job.options)
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/runners/dataflow/internal/apiclient.py",> line 607, in _stage_resources
    staged_resources = resource_stager.stage_job_resources(
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/runners/portability/stager.py",> line 315, in stage_job_resources
    self.stage_artifact(
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/runners/dataflow/internal/apiclient.py",> line 975, in stage_artifact
    self._dataflow_application_client._gcs_file_copy(
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/utils/retry.py",> line 260, in wrapper
    return fun(*args, **kwargs)
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/runners/dataflow/internal/apiclient.py",> line 572, in _gcs_file_copy
    self.stage_file(to_folder, to_name, f, total_size=total_size)
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/runners/dataflow/internal/apiclient.py",> line 633, in stage_file
    response = self._storage_client.objects.Insert(request, upload=upload)
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/internal/clients/storage/storage_v1_client.py",> line 1154, in Insert
    return self._RunMethod(
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/-1734967051/lib/python3.8/site-packages/apitools/base/py/base_api.py",> line 731, in _RunMethod
    return self.ProcessHttpResponse(method_config, http_response, request)
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/-1734967051/lib/python3.8/site-packages/apitools/base/py/base_api.py",> line 737, in ProcessHttpResponse
    self.__ProcessHttpResponse(method_config, http_response, request))
  File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/-1734967051/lib/python3.8/site-packages/apitools/base/py/base_api.py",> line 603, in __ProcessHttpResponse
    raise exceptions.HttpError.FromResponse(
apitools.base.py.exceptions.HttpError: HttpError accessing <https://www.googleapis.com/resumable/upload/storage/v1/b/temp-storage-for-end-to-end-tests/o?alt=json&name=staging-it%2Fbeamapp-jenkins-1226180516-045765.1609005916.046102%2Fdataflow-worker.jar&uploadType=resumable&upload_id=ABg5-UzxdndqlaTAyp5UDYm_nh04UDuVKI_7QEZaFXgTLOFgsjeETPJPZd-0dn0V0yxeJSNpYJjqZL17pf4F_8y12NQ7VPy6eA>: response: <{'content-type': 'text/plain; charset=utf-8', 'x-guploader-uploadid': 'ABg5-UzxdndqlaTAyp5UDYm_nh04UDuVKI_7QEZaFXgTLOFgsjeETPJPZd-0dn0V0yxeJSNpYJjqZL17pf4F_8y12NQ7VPy6eA', 'content-length': '0', 'date': 'Sat, 26 Dec 2020 18:05:33 GMT', 'server': 'UploadServer', 'status': '503'}>, content <>
-------------------- >> begin captured logging << --------------------
root: DEBUG: Unhandled type_constraint: Union[]
apache_beam.runners.portability.stager: INFO: Executing command: ['<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/build/gradleenv/-1734967051/bin/python',> '-m', 'pip', 'download', '--dest', '/tmp/dataflow-requirements-cache', '-r', 'postcommit_requirements.txt', '--exists-action', 'i', '--no-binary', ':all:']
apache_beam.runners.portability.stager: INFO: Copying Beam SDK "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/build/apache-beam.tar.gz"> to staging location.
root: WARNING: Make sure that locally built Python SDK docker image has Python 3.8 interpreter.
root: INFO: Default Python SDK image for environment is apache/beam_python3.8_sdk:2.28.0.dev
root: INFO: Using provided Python SDK container image: gcr.io/cloud-dataflow/v1beta3/python38:beam-master-20201214
root: INFO: Python SDK container image set to "gcr.io/cloud-dataflow/v1beta3/python38:beam-master-20201214" for Docker environment
apache_beam.runners.portability.fn_api_runner.translations: INFO: ==================== <function eliminate_common_key_with_none at 0x7fa07f3d78b0> ====================
apache_beam.runners.portability.fn_api_runner.translations: DEBUG: 26 [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]
apache_beam.runners.portability.fn_api_runner.translations: DEBUG: Stages: ['ref_AppliedPTransform_read from datastore/UserQuery/Impulse_4\n  read from datastore/UserQuery/Impulse:beam:transform:impulse:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_read from datastore/UserQuery/FlatMap(<lambda at core.py:2957>)_5\n  read from datastore/UserQuery/FlatMap(<lambda at core.py:2957>):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_read from datastore/UserQuery/Map(decode)_7\n  read from datastore/UserQuery/Map(decode):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_read from datastore/SplitQuery_8\n  read from datastore/SplitQuery:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_read from datastore/Reshuffle/AddRandomKeys_10\n  read from datastore/Reshuffle/AddRandomKeys:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_read from datastore/Reshuffle/ReshufflePerKey/Map(reify_timestamps)_12\n  read from datastore/Reshuffle/ReshufflePerKey/Map(reify_timestamps):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_read from datastore/Reshuffle/ReshufflePerKey/GroupByKey_13\n  read from datastore/Reshuffle/ReshufflePerKey/GroupByKey:beam:transform:group_by_key:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_read from datastore/Reshuffle/ReshufflePerKey/FlatMap(restore_timestamps)_14\n  read from datastore/Reshuffle/ReshufflePerKey/FlatMap(restore_timestamps):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_read from datastore/Reshuffle/RemoveRandomKeys_15\n  read from datastore/Reshuffle/RemoveRandomKeys:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_read from datastore/Read_16\n  read from datastore/Read:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_split_17\n  split:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_pair_with_one_18\n  pair_with_one:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_group_19\n  group:beam:transform:group_by_key:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_count_20\n  count:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_format_21\n  format:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_write/Write/WriteImpl/DoOnce/Impulse_26\n  write/Write/WriteImpl/DoOnce/Impulse:beam:transform:impulse:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_write/Write/WriteImpl/DoOnce/FlatMap(<lambda at core.py:2957>)_27\n  write/Write/WriteImpl/DoOnce/FlatMap(<lambda at core.py:2957>):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_write/Write/WriteImpl/DoOnce/Map(decode)_29\n  write/Write/WriteImpl/DoOnce/Map(decode):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_write/Write/WriteImpl/InitializeWrite_30\n  write/Write/WriteImpl/InitializeWrite:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_write/Write/WriteImpl/WindowInto(WindowIntoFn)_31\n  write/Write/WriteImpl/WindowInto(WindowIntoFn):beam:transform:window_into:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_write/Write/WriteImpl/WriteBundles_32\n  write/Write/WriteImpl/WriteBundles:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_write/Write/WriteImpl/Pair_33\n  write/Write/WriteImpl/Pair:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_write/Write/WriteImpl/GroupByKey_34\n  write/Write/WriteImpl/GroupByKey:beam:transform:group_by_key:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_write/Write/WriteImpl/Extract_35\n  write/Write/WriteImpl/Extract:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_write/Write/WriteImpl/PreFinalize_36\n  write/Write/WriteImpl/PreFinalize:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_write/Write/WriteImpl/FinalizeWrite_37\n  write/Write/WriteImpl/FinalizeWrite:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>']
apache_beam.runners.portability.fn_api_runner.translations: INFO: ==================== <function pack_combiners at 0x7fa07f3d7940> ====================
apache_beam.runners.portability.fn_api_runner.translations: DEBUG: 26 [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]
apache_beam.runners.portability.fn_api_runner.translations: DEBUG: Stages: ['ref_AppliedPTransform_read from datastore/UserQuery/Impulse_4\n  read from datastore/UserQuery/Impulse:beam:transform:impulse:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_read from datastore/UserQuery/FlatMap(<lambda at core.py:2957>)_5\n  read from datastore/UserQuery/FlatMap(<lambda at core.py:2957>):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_read from datastore/UserQuery/Map(decode)_7\n  read from datastore/UserQuery/Map(decode):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_read from datastore/SplitQuery_8\n  read from datastore/SplitQuery:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_read from datastore/Reshuffle/AddRandomKeys_10\n  read from datastore/Reshuffle/AddRandomKeys:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_read from datastore/Reshuffle/ReshufflePerKey/Map(reify_timestamps)_12\n  read from datastore/Reshuffle/ReshufflePerKey/Map(reify_timestamps):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_read from datastore/Reshuffle/ReshufflePerKey/GroupByKey_13\n  read from datastore/Reshuffle/ReshufflePerKey/GroupByKey:beam:transform:group_by_key:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_read from datastore/Reshuffle/ReshufflePerKey/FlatMap(restore_timestamps)_14\n  read from datastore/Reshuffle/ReshufflePerKey/FlatMap(restore_timestamps):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_read from datastore/Reshuffle/RemoveRandomKeys_15\n  read from datastore/Reshuffle/RemoveRandomKeys:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_read from datastore/Read_16\n  read from datastore/Read:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_split_17\n  split:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_pair_with_one_18\n  pair_with_one:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_group_19\n  group:beam:transform:group_by_key:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_count_20\n  count:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_format_21\n  format:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_write/Write/WriteImpl/DoOnce/Impulse_26\n  write/Write/WriteImpl/DoOnce/Impulse:beam:transform:impulse:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_write/Write/WriteImpl/DoOnce/FlatMap(<lambda at core.py:2957>)_27\n  write/Write/WriteImpl/DoOnce/FlatMap(<lambda at core.py:2957>):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_write/Write/WriteImpl/DoOnce/Map(decode)_29\n  write/Write/WriteImpl/DoOnce/Map(decode):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_write/Write/WriteImpl/InitializeWrite_30\n  write/Write/WriteImpl/InitializeWrite:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_write/Write/WriteImpl/WindowInto(WindowIntoFn)_31\n  write/Write/WriteImpl/WindowInto(WindowIntoFn):beam:transform:window_into:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_write/Write/WriteImpl/WriteBundles_32\n  write/Write/WriteImpl/WriteBundles:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_write/Write/WriteImpl/Pair_33\n  write/Write/WriteImpl/Pair:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_write/Write/WriteImpl/GroupByKey_34\n  write/Write/WriteImpl/GroupByKey:beam:transform:group_by_key:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_write/Write/WriteImpl/Extract_35\n  write/Write/WriteImpl/Extract:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_write/Write/WriteImpl/PreFinalize_36\n  write/Write/WriteImpl/PreFinalize:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_write/Write/WriteImpl/FinalizeWrite_37\n  write/Write/WriteImpl/FinalizeWrite:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>']
apache_beam.runners.portability.fn_api_runner.translations: INFO: ==================== <function sort_stages at 0x7fa07f3d6160> ====================
apache_beam.runners.portability.fn_api_runner.translations: DEBUG: 26 [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]
apache_beam.runners.portability.fn_api_runner.translations: DEBUG: Stages: ['ref_AppliedPTransform_read from datastore/UserQuery/Impulse_4\n  read from datastore/UserQuery/Impulse:beam:transform:impulse:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_read from datastore/UserQuery/FlatMap(<lambda at core.py:2957>)_5\n  read from datastore/UserQuery/FlatMap(<lambda at core.py:2957>):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_read from datastore/UserQuery/Map(decode)_7\n  read from datastore/UserQuery/Map(decode):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_read from datastore/SplitQuery_8\n  read from datastore/SplitQuery:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_read from datastore/Reshuffle/AddRandomKeys_10\n  read from datastore/Reshuffle/AddRandomKeys:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_read from datastore/Reshuffle/ReshufflePerKey/Map(reify_timestamps)_12\n  read from datastore/Reshuffle/ReshufflePerKey/Map(reify_timestamps):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_read from datastore/Reshuffle/ReshufflePerKey/GroupByKey_13\n  read from datastore/Reshuffle/ReshufflePerKey/GroupByKey:beam:transform:group_by_key:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_read from datastore/Reshuffle/ReshufflePerKey/FlatMap(restore_timestamps)_14\n  read from datastore/Reshuffle/ReshufflePerKey/FlatMap(restore_timestamps):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_read from datastore/Reshuffle/RemoveRandomKeys_15\n  read from datastore/Reshuffle/RemoveRandomKeys:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_read from datastore/Read_16\n  read from datastore/Read:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_split_17\n  split:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_pair_with_one_18\n  pair_with_one:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_group_19\n  group:beam:transform:group_by_key:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_count_20\n  count:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_format_21\n  format:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_write/Write/WriteImpl/DoOnce/Impulse_26\n  write/Write/WriteImpl/DoOnce/Impulse:beam:transform:impulse:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_write/Write/WriteImpl/DoOnce/FlatMap(<lambda at core.py:2957>)_27\n  write/Write/WriteImpl/DoOnce/FlatMap(<lambda at core.py:2957>):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_write/Write/WriteImpl/DoOnce/Map(decode)_29\n  write/Write/WriteImpl/DoOnce/Map(decode):beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_write/Write/WriteImpl/InitializeWrite_30\n  write/Write/WriteImpl/InitializeWrite:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_write/Write/WriteImpl/WindowInto(WindowIntoFn)_31\n  write/Write/WriteImpl/WindowInto(WindowIntoFn):beam:transform:window_into:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_write/Write/WriteImpl/WriteBundles_32\n  write/Write/WriteImpl/WriteBundles:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_write/Write/WriteImpl/Pair_33\n  write/Write/WriteImpl/Pair:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_write/Write/WriteImpl/GroupByKey_34\n  write/Write/WriteImpl/GroupByKey:beam:transform:group_by_key:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_write/Write/WriteImpl/Extract_35\n  write/Write/WriteImpl/Extract:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_write/Write/WriteImpl/PreFinalize_36\n  write/Write/WriteImpl/PreFinalize:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>', 'ref_AppliedPTransform_write/Write/WriteImpl/FinalizeWrite_37\n  write/Write/WriteImpl/FinalizeWrite:beam:transform:pardo:v1\n  must follow: \n  downstream_side_inputs: <unknown>']
apache_beam.runners.dataflow.dataflow_runner: WARNING: Typical end users should not use this worker jar feature. It can only be used when FnAPI is enabled.
apache_beam.internal.gcp.auth: INFO: Setting socket default timeout to 60 seconds.
apache_beam.internal.gcp.auth: INFO: socket default timeout is 60.0 seconds.
root: DEBUG: Connecting using Google Application Default Credentials.
apache_beam.runners.dataflow.internal.apiclient: INFO: Starting GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-1226180516-045765.1609005916.046102/pipeline.pb...
oauth2client.transport: INFO: Attempting refresh to obtain initial access_token
oauth2client.transport: INFO: Attempting refresh to obtain initial access_token
apache_beam.runners.dataflow.internal.apiclient: INFO: Completed GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-1226180516-045765.1609005916.046102/pipeline.pb in 0 seconds.
apache_beam.runners.dataflow.internal.apiclient: INFO: Starting GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-1226180516-045765.1609005916.046102/requirements.txt...
apache_beam.runners.dataflow.internal.apiclient: INFO: Completed GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-1226180516-045765.1609005916.046102/requirements.txt in 0 seconds.
apache_beam.runners.dataflow.internal.apiclient: INFO: Starting GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-1226180516-045765.1609005916.046102/parameterized-0.7.4.tar.gz...
apache_beam.runners.dataflow.internal.apiclient: INFO: Completed GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-1226180516-045765.1609005916.046102/parameterized-0.7.4.tar.gz in 0 seconds.
apache_beam.runners.dataflow.internal.apiclient: INFO: Starting GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-1226180516-045765.1609005916.046102/pbr-5.5.1.tar.gz...
apache_beam.runners.dataflow.internal.apiclient: INFO: Completed GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-1226180516-045765.1609005916.046102/pbr-5.5.1.tar.gz in 0 seconds.
apache_beam.runners.dataflow.internal.apiclient: INFO: Starting GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-1226180516-045765.1609005916.046102/mock-2.0.0.tar.gz...
apache_beam.runners.dataflow.internal.apiclient: INFO: Completed GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-1226180516-045765.1609005916.046102/mock-2.0.0.tar.gz in 0 seconds.
apache_beam.runners.dataflow.internal.apiclient: INFO: Starting GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-1226180516-045765.1609005916.046102/six-1.15.0.tar.gz...
apache_beam.runners.dataflow.internal.apiclient: INFO: Completed GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-1226180516-045765.1609005916.046102/six-1.15.0.tar.gz in 0 seconds.
apache_beam.runners.dataflow.internal.apiclient: INFO: Starting GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-1226180516-045765.1609005916.046102/funcsigs-1.0.2.tar.gz...
apache_beam.runners.dataflow.internal.apiclient: INFO: Completed GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-1226180516-045765.1609005916.046102/funcsigs-1.0.2.tar.gz in 0 seconds.
apache_beam.runners.dataflow.internal.apiclient: INFO: Starting GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-1226180516-045765.1609005916.046102/pbr-5.4.5.tar.gz...
apache_beam.runners.dataflow.internal.apiclient: INFO: Completed GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-1226180516-045765.1609005916.046102/pbr-5.4.5.tar.gz in 0 seconds.
apache_beam.runners.dataflow.internal.apiclient: INFO: Starting GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-1226180516-045765.1609005916.046102/pbr-5.5.0.tar.gz...
apache_beam.runners.dataflow.internal.apiclient: INFO: Completed GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-1226180516-045765.1609005916.046102/pbr-5.5.0.tar.gz in 0 seconds.
apache_beam.runners.dataflow.internal.apiclient: INFO: Starting GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-1226180516-045765.1609005916.046102/PyHamcrest-1.10.1.tar.gz...
apache_beam.runners.dataflow.internal.apiclient: INFO: Completed GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-1226180516-045765.1609005916.046102/PyHamcrest-1.10.1.tar.gz in 0 seconds.
apache_beam.runners.dataflow.internal.apiclient: INFO: Starting GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-1226180516-045765.1609005916.046102/dataflow_python_sdk.tar...
apache_beam.runners.dataflow.internal.apiclient: INFO: Completed GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-1226180516-045765.1609005916.046102/dataflow_python_sdk.tar in 0 seconds.
apache_beam.runners.dataflow.internal.apiclient: INFO: Starting GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-1226180516-045765.1609005916.046102/dataflow-worker.jar...
--------------------- >> end captured logging << ---------------------
Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-12-26_10_05_50-739012942246181865?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-12-26_10_19_16-3852691306074660199?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-12-26_10_26_30-3562563199594336595?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-12-26_10_33_48-12379882399096224908?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-12-26_10_40_40-17521463370846429278?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-12-26_10_56_58-2994150245777310283?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-12-26_10_05_27-12128362785885207070?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-12-26_10_26_39-11033531086744911894?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-12-26_10_34_26-15463769755217711841?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-12-26_10_42_14-991036710938483690?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-12-26_10_50_38-4762833626527244725?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-12-26_10_58_50-932727555904114197?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-12-26_10_05_26-7589064460095569367?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-12-26_10_16_52-5956823027964607933?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-12-26_10_23_55-6532682488627083752?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-12-26_10_31_07-16406138359959330710?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-12-26_10_39_11-9169887216060493993?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-12-26_10_46_54-7471139452081235787?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-12-26_10_53_46-13590499742180109969?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-12-26_11_00_43-11481553051913178534?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-12-26_10_05_23-3340706854697603552?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-12-26_10_23_16-16493535707356006806?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-12-26_10_31_23-15065727829885445259?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-12-26_10_38_54-375814237656103621?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-12-26_10_47_19-11954027734929920745?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-12-26_10_55_03-1611084637585463046?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-12-26_11_02_22-17801338800389144271?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-12-26_10_07_19-5569169617427353186?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-12-26_10_16_20-2828598792889390041?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-12-26_10_23_38-10282329220951519126?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-12-26_10_31_31-7688210861514436085?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-12-26_10_39_41-13447948721567305389?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-12-26_10_46_34-10029628168951535720?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-12-26_10_54_08-13304357221363077793?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-12-26_11_02_02-14134487848757475591?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-12-26_10_05_53-12127552587963646955?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-12-26_10_15_31-14013861959616074238?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-12-26_10_25_37-7960745132172210499?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-12-26_10_34_52-10142599314564414524?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-12-26_10_42_58-1433003847385151875?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-12-26_10_50_24-11040885191592485134?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-12-26_10_57_49-3684115431482916690?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-12-26_11_05_03-10014684115055591028?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-12-26_10_05_27-6656959389095261567?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-12-26_10_13_38-4286592086808049961?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-12-26_10_22_29-11259644467455614848?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-12-26_10_31_34-13189661551298740002?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-12-26_10_39_50-11018627692534114216?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-12-26_10_47_49-13409837995067353389?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-12-26_10_56_14-14233048008491678300?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-12-26_10_05_56-1133205958400930587?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-12-26_10_14_15-4352282118679371544?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-12-26_10_22_23-12685493064214745807?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-12-26_10_30_46-14951762194448378060?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-12-26_10_38_58-10591573997211775893?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-12-26_10_46_49-7483704135910321111?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-12-26_10_54_36-4633280830980941428?project=apache-beam-testing
Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2020-12-26_11_02_53-2063783446857817649?project=apache-beam-testing

----------------------------------------------------------------------
XML: nosetests-postCommitIT-df-py38.xml
----------------------------------------------------------------------
XML: <https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/nosetests.xml>
----------------------------------------------------------------------
Ran 68 tests in 4070.467s

FAILED (SKIP=7, errors=2)

> Task :sdks:python:test-suites:dataflow:py38:postCommitIT FAILED

FAILURE: Build failed with an exception.

* Where:
Script '<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/test-suites/dataflow/common.gradle'> line: 118

* What went wrong:
Execution failed for task ':sdks:python:test-suites:dataflow:py38:postCommitIT'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with Gradle 7.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See https://docs.gradle.org/6.7.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 1h 12m 14s
212 actionable tasks: 152 executed, 56 from cache, 4 up-to-date
Gradle was unable to watch the file system for changes. The inotify watches limit is too low.

Publishing build scan...
https://gradle.com/s/tu7p4cqbtqnrc

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org


Jenkins build is back to normal : beam_PostCommit_Python38 #695

Posted by Apache Jenkins Server <je...@builds.apache.org>.
See <https://ci-beam.apache.org/job/beam_PostCommit_Python38/695/display/redirect>


---------------------------------------------------------------------
To unsubscribe, e-mail: builds-unsubscribe@beam.apache.org
For additional commands, e-mail: builds-help@beam.apache.org